Modeling Probability: Introduction to Markov Chains

-

Introduction 

Markov Chains are just a fancy way of saying what the probability between two or more things would be using current present state data. They are run on thousands of pieces of data to estimate what the next “state” will be. Let’s take below as an example:

Example

Suppose we feed the past 5 years’ data on the weather into this simulation.
The three big circles that we have identified are what we call “states.” The arrows are the possibility for one state to go to another in a certain time interval you choose. We will be doing a one day interval. In our case, we have three states of the weather: rainy, sunny, and cloudy. The chance for the weather going from sunny to cloudy is 40% (as represented by the arrow). The chance for the weather going from sunny to rainy is 10%. It could also be sunny the next day, and that chance is 50%. Notice that the chances of the Sunny state all add up to 100%. This happens for all the other states as well.
This could also be represented in a table:

Why Use Them?

Markov Chains also have an embedded way to account for uncertainty. For example, stocks can be extremely volatile in some cases (take, for example, the 2008 Stock Market Crash). Markov Chains can be good in accounting for this extreme volatility, and they are the basis of stochastic modeling and predicting. They are found in several different places!

What I have described is a simple Markov Model. For sectors like finance or economics, you would need a more specific version of the Markov Chain Model, such as the Markov Chain Monte Carlo (MCMC) Model.

1. Finance/Economics (Mainly used to model macroeconomic events.)
2. Betting (e.g. sports betting)
3. Biology (e.g. making predictive modeling on what the next sequence of DNA could be)
4. What lunch the cafeteria is gong to have (👀)
5. Game Theory
6. Natural Language Processing (e.g. guess what the next word will be in a paragraph)
7. Chemistry (e.g. modeling chemical reactions)

Pros/Cons

Pros

1. Easy to use out of the get-go and simple.
2. Accounts for uncertainty better than most other models.
3. Has flexibility in its modeling of the transitions.
4. Markov chains have a more natural way of expressing probabilities than most other combinatorial methods (representing how objects are counted).

Cons

1. It isn’t good with “understanding” stuff (it’s another reason why you would use an attention mechanism).
2. Can’t know all possible states/predict all real-world scenarios (e.g. in biology perhaps it doesn’t account for a pathogen that is in the bloodstream).
3. Requires a lot of data to properly predict transitions.
4. Creating/Forming data in a manner that accounts for the states and the “lines” (or transition) to the next state. It can be hard to differentiate between them.
5. Markov proprieties are only good for the example you have created. They cannot be used for multiple different types of probabilistic modeling (e.g. attempting to ask how you make a bridge to an AI trained to assist you for low-level tasks such as creating a to-do list).
6. The simplicity part also means that modeling complex processes can be extremely difficult (and are better not modeled by Markov Chains in most cases).

Conclusion

In total, Markov Chains are just another way to model probability between two possible states while also accounting for some amount of uncertainty. They can be used in any field– from Biology to Finance. Although easy to use and better at modeling uncertainty than most models, it still has issues understanding how each state works and moves, and can’t model complex tasks.

Categories

4 COMMENTS

Latest News

Crew Competes at 98th Stotesbury Cup Regatta

`The Stotesbury Cup Regatta has been held anually on the Schuylkill River in Philadelphia, Pennsylvania. It is the largest...

A Season in Review: Jesuit Competitive Shooting

Once again, the Jesuit Dallas Competitive Shooting Team took aim for success. The team has proven that precision, discipline,...

There’s a New Doc on the Block – An Interview with Jesuit’s latest EdD Recipient: Dr. Brian Goll

Recently, Jesuit’s own Mr. Brian Goll became the next doctor to join Jesuit’s growing roster. Dr. Goll has been...

Escaping the Blistering Dallas Heat – Some Amazing Summertime Adventures

The Summer Situation With summer right around the corner and school getting out, we will all have a ton of...

Addressing the Problem of the Dinner Dilemma

The Big Decision If your family is like mine, then the topic of where to eat for dinner is usually...

A Farewell Delight: Apple Pie

Is there anything more American than a fresh slice of apple pie and a cold scoop of vanilla ice...

Read More#AMDG
Related Articles