Below I discuss a curious feature that arises from the Markov property in an asset pricing model.
Asset returns are functionals of a Markov chain
where
,
and
. Dynamics are governed by the Markov transition kernel
, which contains the conditional transition probabilities
for ordered pairs
.
Realizations are normalized to be indicator functions . The state of the system is characterized by the indicator vector
, where the column of indicators is zero everywhere other than the realized coordinate.
Expected returns are . Under the Markov transition decomposition
with
and leading eigenvector
, expected returns can be written
.
The rows of the transition matrix sum to one. Preservation of probability enforces a zero-sum condition . Assuming in addition that
, the rows of
contain both positive and negative values.
The rows of describe the local redistribution of probabilities away from their invariant values. Because they contain negative entries in every nontrivial case, they are generalized probabilities.
The following simulations illustrate the phenomena of probability redistribution and dimension reduction for a Markov chain.

Each enumerated facet corresponds to a state of a 12-state Markov process. Integer powers of a Markov transition matrix capture transition probabilities over integer horizons. The plots show smoothed frequency counts of generalized probability values for successive horizons (black lines are early iterations – blue lines are later). The top panel shows distributions of probabilities relevant for forecasting k as k goes from periods 1-15, the lower shows horizons from 1 to 8.

The probability of each state is approximately bimodal for mid-range forecasts. For limiting long-run forecasts, the distributions are degenerate with all mass assigned to nonzero probabilities at the invariant values, represented by the right modes. Probability re-distributional effects are represented by the left modes.
An initial transition matrix, say , contains probabilities for each state in terms of the conditional probabilities of arriving in that state from each of every other state; it also contains the probability of staying in that state. The collection of these numbers is the initial distribution of probabilities for each state. This probability becomes less dependent on conditioning as the forecasting horizon grows, ultimately approaching its invariant or long-run value (right modes). The corresponding conditional probability re-distributions approach zero, but are non-negligible for finite-horizon forecasts (left modes).
In any finite sample, ergodic time series are generated by transition probabilities with some conditional dependence. Think, say, of the curve corresponding to the horizon h=8, and consider repetitions of a transition with the corresponding probabilities. This new process is also a Markov process, with an approximate 2-factor structure for dynamics. The trailing factor drives fluctuations that die out in large-time forecasts but are predictable in finite samples.