A stochastic process has the Markov property if the conditional probability distribution of future states of the process depends only upon the present state; that is, given the present, the future does not depend on the past. A process with this property is said to be Markovian or a Markov process. The most famous Markov process is a Markov chain. Brownian motion is another well-known Markov process.
Alternatively, the Markov property can be formulated as follows. for all and bounded and measurable.
Strong Markov property
Suppose that is a stochastic process on a probability space with natural filtration. Then for any stopping time on, we can define Then is said to have the strong Markov property if, for each stopping time, conditioned on the event, we have that for each, is independent of given. The strong Markov property implies the ordinary Markov property since by taking the stopping time, the ordinary Markov property can be deduced.
In forecasting
In the fields of predictive modelling and probabilistic forecasting, the Markov property is considered desirable since it may enable the reasoning and resolution of the problem that otherwise would not be possible to be resolved because of its intractability. Such a model is known as a Markov model.
Examples
Assume that an urn contains two red balls and one green ball. One ball was drawn yesterday, one ball was drawn today, and the final ball will be drawn tomorrow. All of the draws are "without replacement". Suppose you know that today's ball was red, but you have no information about yesterday's ball. The chance that tomorrow's ball will be red is 1/2. That's because the only two remaining outcomes for this random experiment are:
Day
Outcome 1
Outcome 2
Yesterday
Red
Green
Today
Red
Red
Tomorrow
Green
Red
On the other hand, if you know that both today and yesterday's balls were red, then you are guaranteed to get a green ball tomorrow. This discrepancy shows that the probability distribution for tomorrow's color depends not only on the present value, but is also affected by information about the past. This stochastic process of observed colors doesn't have the Markov property. Using the same experiment above, if sampling "without replacement" is changed to sampling "with replacement," the process of observed colors will have the Markov property. An application of the Markov property in a generalized form is in Markov chain Monte Carlo computations in the context of Bayesian statistics.