A Markov chain is a discrete stochastic process with discrete states and discrete transformations between them. At each time instant the system is in one of the possible states, numbered from one to . At regularly spaced discrete times, the system switches its state, possibly back to the same state. The initial state of the chain is denoted and the states after each time of change are . Standard first order Markov chain has the additional property that the probabilities of the future states depend only on the current state and not the ones before it [48]. Formally this means that
Because of the Markov property, the complete probability distribution of the states of a Markov chain is defined by the initial distribution and the state transition probability matrix
This allows the evaluation of the probability of a sequence of states , given the model parameters , as