Dictionary

Markov chain

noun

Definition of MARKOV CHAIN

:  a usually discrete stochastic process (as a random walk) in which the probabilities of occurrence of various future states depend only on the present state of the system or on the immediately preceding state and not on the path by which the present state was achieved —called also Markoff chain

Origin of MARKOV CHAIN

A. A. Markov †1922 Russian mathematician
First Known Use: 1938

Browse

Next Word in the Dictionary: MarkovianPrevious Word in the Dictionary: markkaAll Words Near: Markov chain
July 28, 2015
pachyderm Hear it
an elephant
Take a 3-minute break and test your skills!
How to use a word that (literally) drives some people nuts.
Test your vocab with our fun, fast game
Ailurophobia, and 9 other unusual fears