The ability to distinguish between stochastic systems based on their trajectories is crucial in thermodynamics, chemistry, and biophysics. The Kullback–Leibler (KL) divergence, DKLAB(0,τ), quantifies the distinguishability between the two ensembles of length-τ trajectories from Markov processes A and B. However, evaluating DKLAB(0,τ) from histograms of trajectories faces sufficient sampling difficulties, and no theory explicitly reveals what dynamical features contribute to the distinguishability. This work provides a general formula that decomposes DKLAB(0,τ) in space and time for any Markov processes, arbitrarily far from equilibrium or steady state. It circumvents the sampling difficulty of evaluating DKLAB(0,τ). Furthermore, it explicitly connects trajectory KL divergence with individual transition events and their waiting time statistics. The results provide insights into understanding distinguishability between Markov processes, leading to new theoretical frameworks for designing biological sensors and optimizing signal transduction.

Markov models are powerful in describing various phenomena in physics, chemistry, and biology.1–13 These inherently probabilistic models are particularly effective in capturing the dynamic behaviors of complex systems,14–18 and particularly non-equilibrium behaviors, including stochastic thermodynamic processes, chemical reaction kinetics,4,19–22 and the configuration dynamics of biomolecules.23–26 In biology, Markov models have provided invaluable insights into coarse-grained configuration dynamics of complex molecules, elucidating sub-cellular processes, such as gene regulation networks,27–29 molecular motors,30,31 ion channels,32–34 and various cellular processes.7,35,36

The ability to distinguish between two Markov systems plays a crucial role in solving various practical problems. For example, consider identifying mutations in proteins. By examining how a protein’s stochastic trajectories traversing various meta-stable states differ from that of a wild-type protein, one can tell the difference between the mutant and the wild-type. Another example is signal pattern recognition. Consider a sensor as a Markov system influenced by external signals. The sensor’s ability to distinguish between various signal patterns is demonstrated by its ability to generate different stochastic trajectories in response to each pattern.

In this work, we adopt the master equation description of a Markov process on a finite number of states x ∈ {1, 2, …, N},
(1)
where p(xt) is the probability to find the system at state x at time t, Rxx(t) is the probability transition rate from state x′ to state x at time t, and Rxx(t) = −x′≠xRxx(t) preserves the normalization. The master equation can also be written in a matrix form,
(2)
where p(t)={p(x;t)} is a vector and R̂={Rxx} is the rate matrix.

From an observational perspective, two Markov processes can be distinguished by the statistics of their states and the kinetic transitions among states (trajectories). The state transition trajectory contains more information about the Markov process than the probability distribution over the state space.37–41 To illustrate this, consider two Markov processes with the same stationary probability distribution, yet one system satisfies detailed balance conditions, and the other is dissipative. If both systems reach the stationary distribution, the observer cannot distinguish the two processes by merely observing the statistics of their states. However, observing the trajectories of the two processes can easily distinguish the dissipative system (with non-zero net current) from the detailed balance system (without net current). This argument has interesting indications in biological information sensing: It has been shown that the ensemble of configuration trajectories for a ligand–receptor sensor contains more information than the statistical average of its states. Moreover, by observing the trajectory of a binary-state sensor, one can simultaneously infer multiple environmental parameters (multiplexing).40–42 

Let us consider two ensembles of stochastic trajectories Xτ from two Markov processes A and B. Each trajectory is of time length τ. Here, the probability distribution of length-τ stochastic trajectories for Markov process A is denoted by
(3)
where pA(x0; 0) is the initial state distribution of process A and pA(xi+1|xiti) is the transfer probability to state xi+1 conditioned from state xi at time window [ti, ti + δt). Here, in the limit of δt ≪ 1, Xτ is a continuous-time discrete-state trajectory of time length τ, and PA[Xτ] becomes the probability density of a continuous trajectory. The distinguishability between the ensembles of trajectory (from time 0 to τ) observed from two Markov processes A and B can be characterized by the following Kullback–Leibler (KL) divergence:43 
(4)
Here, the value of DKLAB(0,τ) characterizes the distinguishability between two Markov processes from comparing ensembles of trajectories of length τ. To avoid singularity of DKLAB(0,τ), we assume, for all x′ and x, if RxxA(t)0, then RxxB(t)0, and vice versa.

There are two main drawbacks in representing DKLAB(0,τ) merely from the above definition. First, the evaluation of DKLAB(0,τ) between two finite ensembles of trajectories suffers from insufficient sampling and slow convergence.44,45 Second, the definition of DKLAB(0,τ) in terms of trajectory probabilities only implicitly depends on the transition kinetic rates, and thus, it misses an explicit interpretation of the distinguishability in terms of the kinetic properties of the two systems.

The central result of this Communication—Eqs. (5)(10)—provides insights into understanding and evaluating DKLAB(0,τ), the observational distinguishability between trajectory ensembles from two Markov processes. First, DKLAB(0,τ) can be separated into the initial state KL divergence and the accumulated KL divergence,
(5)
where dKLAB(0) denotes the state-ensemble KL divergence of the two processes’ initial state distributions,
(6)
The accumulated KL divergence DaccAB(0,τ) is the additional distinguishability gained by observing the trajectories between 0 and τ. It follows the following decomposition in time and space (or edge):
(7)
Here, the summation edge xx includes all directed edges (connecting from state x to x′) with non-zero rate Rxx(t) at time t. Furthermore, we find that the distinguishability accumulation rate from each directed edge, ḊKL,xxAB(t), can be written as the product of the detailed current of the edge JxxA and a divergence force FxxAB(t) (see Sec. I of the supplementary material),
(8)
The detailed probability current from x to x′ at time t is
(9)
The divergence force itself is a KL divergence between two transition dwell time distributions,
(10)
where pwA,xx,t(s)=RxxA(t)expRxxA(t)s is the exponential distribution of dwell time s in state x before a transition to state x′ occurs, characterized by the transition probability rate RxxA(t) at time t.

The results reveal several critical properties of the trajectory KL divergence, DKLAB(0,τ). It provides three perspectives to understand the distinguishability between two Markov systems.

1. The initial and the accumulated divergence

When the observed trajectory length reduces to τ → 0, DKLAB(0,τ)=dKLAB(0). In this limit, the trajectory KL divergence of the two Markov processes reduces to the KL divergence between the two initial state distributions. That is, by observing infinitely short trajectories, the distinguishability of two Markov processes equals the distinguishability between the systems’ initial state distributions. For any finite-time observation τ > 0, Eq. (8) implies that as trajectory length τ increases, the distinguishability always accumulates with a non-negative rate: ḊKL,xxAB(t)0,{(t,x,x)|xx}. In other words, the longer the observed trajectory, the greater the revealed distinguishability between the two Markov processes. It is noteworthy that the initial state distribution of process B does not impact DaccAB(0,τ), since its effect is fully captured by the initial dKLAB(0).

2. Temporal decomposition of DKLAB(0,τ)

Here, we define the temporal additivity of the trajectory KL divergence by decomposing one time period into two consecutive segments,
(11)
where DKLAB(ta,tb) denotes the trajectory KL divergence for time starting at ta and ending at tb and dKLAB(t*) denotes the state-ensemble KL divergence between two systems evaluated at time t*. This additivity formula indicates that one can obtain the KL divergence of trajectories starting at 0 and ending at τ as the summation of that for two (or more) consecutive trajectories (0, t*) and (t*, τ), after the removal of the overlapping term(s) dKLAB(t*). This decomposition can also be represented in a cleaner form in terms of accumulated KL divergence,
(12)

3. Spatial (edge-wise) decomposition of DKLAB(0,τ)

Equations (7) and (8) reveal the non-negative spatial (edge-wise) additivity of the trajectory KL divergence, where the accumulated KL divergence is the summation of non-negative contributions from each directed transition via connected edges. Furthermore, Eq. (8) shows that the contribution from each directed edge equals the product of the detailed probability flow rate [Eq. (9)] and a force-like weighting factor [Eq. (10)]. The force-like factor characterizes the distinction between the two transition rates of the same edge from two Markov processes. This additivity of non-negative contribution from each edge implies that coarse-graining of states or ignoring any edge of the graph may cause underestimation of the trajectory KL divergence, which agrees with intuition. Importantly, this spatial decomposition in state transitions leads to a practical formula described at the end of this paper, allowing for the development of enhanced transition event sampling methods toward the efficient evaluation of DKLAB(0,τ).

Entropy plays a central role in nonequilibrium thermodynamics and is closely related to the Jarzynski equality,46 Crooks fluctuation theorem,47 and other forms of fluctuation theorems.45,48,49 Apparently, the temporal and spatial decomposition of DKLAB(0,τ) resembles the decomposition of the entropy production in stochastic thermodynamics.48 However, the connection between DKLAB(0,τ) and the entropy production is non-trivially dependent on the construction of Markov process B.

In stochastic thermodynamics, the entropy production of a Markov process can be obtained from the logarithmic ratio of the probability measure of a trajectory to the probability measure of the time reversal trajectory, which is known as the Radon–Nikodym derivative (RND).50 In contrast, DKLAB(0,τ) defined in this work involves the RND of one trajectory’s probability measures from two Markov processes A and B. The RND and generalized fluctuation theorems under various choices of conjugate processes can be found in Refs. 6, 45, 49, and 5158. Notably, the trajectory KL divergence under arbitrary conjugate process was discussed for diffusive and Fokker–Planck (FP) dynamics in Refs. 51 and 52. It is worthy of further study to determine if the results of this paper, for discrete-state Markov processes, can be directly extended to FP dynamics.

One may naturally speculate that DKLAB(0,τ) is related to thermodynamic entropy production, if process B is chosen as the “reversal” of process A. However, we find that process B should be defined without inverting the direction of time or control protocol. Rather, DKLAB(0,τ) becomes thermodynamic entropy when process B is chosen to be the pseudo-transpose of A, denoted by A*,
(13)
where for A*, the transition rate from state x to x′ at time t is chosen to be identical to the rate from x′ to x of process A at time t. The normalization is preserved for process A* due to the choice of its diagonal elements. Then, by assuming that Markov process A is immersed in a heat reservoir, the time-accumulated KL divergence between processes A and A* equals the entropy change of the reservoir caused by process A (see Sec. II of the supplementary material),
(14)
where kB is Boltzmann’s constant, sres[Xτ] is the stochastic entropy production in the reservoir due to trajectory Xτ, and the angular bracket ⟨⋅⟩A is the trajectory ensemble average for process A.48 

To summarize, for a Markov thermodynamic process A and its “conjugate” dynamics A*, when the two systems are prepared at the same initial distribution (dKLAA*(0)=0), their trajectory KL divergence equals the reservoir’s entropy change due to process A.

In the remaining, we discuss the practical implications of the theory. We first examine how the initial state distribution affects the observational distinguishability between two Markov systems (see Secs. III A and III B). Then, we discuss how DKLAB(0,τ) responds to control protocols (input signal) and illustrate its implications for biological or artificial stochastic sensors. Specifically, the theory inspires the design of optimal sensor that can discern input signals (see Sec. III C) and reveals the optimal control protocol to discern two apparently similar systems (see Sec. III D). In the end (see Secs. III E and III F), we show that the theory provides multiple new approaches to efficiently evaluate the trajectory KL divergence. These approaches get around the sufficient sampling difficulty of evaluating DKLAB(0,τ) from a finite number of trajectories.

Consider two stochastic systems that evolve according to identical kinetic rates, R̂A(t)=R̂B(t)t, and they differ only by their initial-state distributions, pA(0)pB(0). Does measuring their trajectories provide more information than the initial states regarding distinguishing the two systems? No, we show that the difference in their initial state distributions fully captures the distinguishability of the two Markov processes: From Eq. (10), the divergence force is zero, FxxAB(t)=0, for all times and edges due to the identical kinetic rates R̂A(t)=R̂B(t). Then, according to Eqs. (7) and (8), the accumulated KL divergence by observing longer trajectory always equals 0. Thus, Eq. (5) reduces to DKLAB(0,τ)=dKLAB(0). In other words, additional observations of the trajectories beyond the initial state cannot improve the observational distinguishability between Markov systems with fully identical kinetic rates. An alternative proof by using the data processing inequality59 is shown in Sec. III of the supplementary material.

Consider two Markov systems prepared at identical initial state distributions pA(x; 0) = pB(x; 0) = p(x; 0), yet they evolve with different kinetic rates, which could be time-dependent or time-independent. How to prepare the initial distribution p(x; 0) of the two systems such that their trajectories most prominently reveal their distinguishability? We find that the optimal initial distribution to maximize the distinguishability is typically a delta-function distribution. In other words, there exists one (or a few degenerate) optimal state(s), from which the observational distinguishability of the two systems is maximized. To show this, recognize that the accumulated KL divergence follows an initial-state decomposition relation,
(15)
where DaccAB(0,τ)xini=x is the accumulated trajectory KL divergence if both systems are initially prepared perfectly on a same state x. This result indicates that the trajectory KL divergence is linearly dependent on the initial probability distribution, as illustrated by a tilted flat plane in Figure 1. See the proof of the decomposition relation in Sec. IV A of the supplementary material. As a result, the highest accumulated KL divergence must be achieved when the initial distribution is a delta function at the optimal state x*, where DaccAB(0,τ)xini=x*DaccAB(0,τ)xini=x for any state x. Very rarely, when there are degenerate optimal states {x*}, any initial distribution purely comprising these degenerate states also gives the optimal distinguishability. Intuitively, from Eqs. (7) and (8), this optimal initial state to distinguish two time-independent Markov processes is the one that evolves in time to exhibit the largest accumulated weighted flow of transitions, with the weight defined by Eq. (10). Notice that the specific choice of the optimal state depends on both the systems’ kinetic property and the length of the observational time τ. For illustrative purposes, consider two time-independent Markov processes with anomalous relaxations60,61 where one finds a large separation of the speed of relaxation (big gap in the eigenvalues of R̂); we find that the optimal initial state x* can be determined by observing the slowest relaxation eigenmode (see this example in Fig. 1 and Sec. IV B of the supplementary material).
FIG. 1.

Initial distribution dependence of the trajectory KL divergence. Trajectories are obtained from two three-state Markov processes with a common initial distribution (p0, p1, p2), where p2 = 1 − p0p1. Each data point represents a distinct initial distribution, and the points live on a flat plane in accordance with Eq. (15). The maximum divergence is achieved at a delta-function initial distribution (p0, p1, p2) = (0, 1, 0). See Sec. IV B of the supplementary material for details.

FIG. 1.

Initial distribution dependence of the trajectory KL divergence. Trajectories are obtained from two three-state Markov processes with a common initial distribution (p0, p1, p2), where p2 = 1 − p0p1. Each data point represents a distinct initial distribution, and the points live on a flat plane in accordance with Eq. (15). The maximum divergence is achieved at a delta-function initial distribution (p0, p1, p2) = (0, 1, 0). See Sec. IV B of the supplementary material for details.

Close modal

Now, we conclude the discussion of the initial state distribution and shift the focus to temporal control protocol(s) governing/applied to two Markov systems. Here, we describe a sensor under the influence of a signal λ(t) by a Markov system R̂(λ(t)). Under this construction, the results of this work can be used to describe temporal pattern recognition by a sensor (see Sec. III C). It can also help the search for the optimal external control protocol to maximally reveal/enhance the distinction between two apparently similar Markov systems with slight differences in their kinetic properties (see Sec. III D).

Various biological sensors can improve their sensing capability by utilizing non-trivial stochastic trajectories of their internal states evolving under complex external input signals.37–41,62,63 Consider a stochastic sensor as a Markov system hopping between different internal states. Its time-dependent rate matrix is R̂(λ), and the protocol λ(t) represents the external signal. The downstream sensory network, which interacts with the sensor but does not directly interact with the input signal, can discern two temporal protocols [λA(t) and λB(t)] only by “observing” the sensor’s state paths (trajectories). In this case, the maximum information the downstream sensory network can acquire is limited by the information contained within the observed ensemble of sensor’s state trajectories. Thus, the trajectory KL divergence defined in this work captures the upper bound of the distinguishability of the two signals transmitted via the sensor, and an optimal sensor is one with the highest upper bound given the design restrictions. As a result, the theory in this paper can be used in various practical applications for improving the sensor’s sensitivity or optimizing the ability of pattern recognition. This paper does NOT aim to provide a universal analytical solution for the optimal sensor’s design, because the solution depends on various specific details of the practical problem, e.g., the types of protocols that need to be discerned, the physical coupling between the sensor and the signal, as well as other limitations that define the available form of the sensor’s rate matrix R̂(λ). Although not discussed in this paper, we have recently applied this theory to designing optimal concentration sensors and sensors capable of recognizing patterns of time-varying signal-molecule concentrations.

A conjugate problem of Sec. III C is to design the optimal control protocol that maximally reveals the difference between two Markov systems. To illustrate the problem with a practical example, consider two proteins with subtle differences: a wild type and a mutant. The two molecules share the same number of metastable configurations (states); however, the transition rates between states are different for the two, and their rates are influenced by an external control parameter λ (e.g., electric field, temperature, or pH). If the two molecules are subtly different in their kinetic rates and/or their rates’ dependence on the external control, merely observing the steady-state distribution of the molecules in a stationary environment may reveal, to some extent, the differences between the two molecules. However, our theory indicates that one can enhance the ability to distinguish the two systems when we drive them out of the steady state by the same control protocol λ(t) and observe the difference in their stochastic trajectory ensembles [see Eq. (4)]. Moreover, the theory allows for the search of the optimal control protocol λ*(t) to maximally reveal the difference between the two systems. Given R̂A(λ) and R̂B(λ), the specific forms of the two molecule’s kinetic rates and their dependence on the external control parameter λ, the search for the optimal protocol λ*(t) becomes a variational optimization problem of DaccAB(0,τ). Intuitively, the decomposition relations imply that the optimal driving protocol maximizes the accumulated number of transitions over the edges with accumulation weight factor FxxAB(t).

In the following, we discuss how the decomposition relation can circumvent the sampling convergence difficulties in evaluating the trajectory KL divergence.

When directly evaluated from the definition, the convergence of DKLAB(0,τ) can face significant sampling difficulty. It requires that each trajectory observed in process A must also be observed sufficiently many times in process B. This convergence difficulty grows rapidly with the trajectory length τ, making it almost impossible to directly calculate DKLAB(0,τ) for long trajectories. However, the temporal decomposition shown in Eq. (11) makes it possible to get around the difficulty by the following: By dividing the trajectory into multiple consecutive segments, it is easier to evaluate the segment trajectory KL divergences and reconstruct the whole-trajectory KL divergence. The desired whole-trajectory KL divergence can be obtained by adding the segment KL divergences together minus the state distribution KL divergences dKLAB(t*)’s evaluated at the intersecting time points. This approach significantly reduces the sufficient number of trajectories needed to evaluate DKLAB(0,τ) without singularity. In comparison with traditional approaches, this new trajectory-slicing approach significantly reduces the number of trajectories required for a converging and accurate estimation of the trajectory KL divergence. (See Fig. 2 and Sec. V of the supplementary material for details.)

FIG. 2.

Trajectory KL divergence between two binary-state systems with rates R21A=1,R12A=2 and R21B=2,R12B=1. Each trajectory is 10-step long with time step equal to unity. The KL divergence is evaluated from trajectory ensembles of sizes 2000, 10 000, 50 000, and 250 000 by different methods. Each result is averaged over ten repetitions of calculations for randomly generated trajectories, with the error bar showing ±1 standard deviation from the ten repetitions. The trajectory slicing method slices trajectory into consecutive segments of length 2. See Sec. V of the supplementary material for details.

FIG. 2.

Trajectory KL divergence between two binary-state systems with rates R21A=1,R12A=2 and R21B=2,R12B=1. Each trajectory is 10-step long with time step equal to unity. The KL divergence is evaluated from trajectory ensembles of sizes 2000, 10 000, 50 000, and 250 000 by different methods. Each result is averaged over ten repetitions of calculations for randomly generated trajectories, with the error bar showing ±1 standard deviation from the ten repetitions. The trajectory slicing method slices trajectory into consecutive segments of length 2. See Sec. V of the supplementary material for details.

Close modal
The method proposed above still relies on the definition of KL divergence, Eq. (4), and thus, it explicitly depends on the probability density functions (histograms) of segmented trajectories. Below, we show that with merely the sets of transition events, one can evaluate the trajectory KL divergence. Specifically, the temporal and spatial decomposition relations, Eqs. (8)(10), allow us to express the trajectory KL divergence as a weighted statistical average of transition events,
(16)
where k is the index of the transitions (from state xk to xk) within a trajectory Xτ and the angular bracket is an average among all trajectories obtained from process A. Practical applications of the above formula may fall into one of the three scenarios below.

Scenario 1: If the transition rate matrices R̂A(t) and R̂B(t) are known, then one can directly evaluate the trajectory KL divergence with the above equation, by plugging in the values of FxxAB(t) (see Fig. 3).

FIG. 3.

Evaluation of ḊKLABt based on event-counting method. Trajectories are generated from two binary-state Markov systems with rates R21A,R12A,R21B,R12B=1,2,2,1. The color curves are accumulated ḊKLABt evaluated from a single trajectory. The black solid curve is the average over the ten color trajectories. The black dashed line is the analytical result for ḊKLABt.

FIG. 3.

Evaluation of ḊKLABt based on event-counting method. Trajectories are generated from two binary-state Markov systems with rates R21A,R12A,R21B,R12B=1,2,2,1. The color curves are accumulated ḊKLABt evaluated from a single trajectory. The black solid curve is the average over the ten color trajectories. The black dashed line is the analytical result for ḊKLABt.

Close modal

Scenario 2: If the transition rate matrices are time-homogeneous yet unknown, one can utilize existing tools64–67 to evaluate the transition rates for processes A and B and then obtain FxxAB to perform the calculation described in scenario 1.

Scenario 3: If the transition rate matrices are unknown and time-dependent, one may need to utilize the waiting time statistics of transition between states x and x′ at each time t, to first evaluate the time-dependent rate matrices, and then obtain FxxAB(t) for calculating the trajectory KL divergence. In this case, utilizing the method proposed in Sec. III E may be more efficient for evaluating the trajectory KL divergence.

In conclusion, this study rigorously delineates the stochastic distinguishability between two arbitrary Markov processes, quantified through the trajectory KL divergence, DKLAB(0,τ). Our development of a general formula to decompose DKLAB(0,τ) in both spatial and temporal dimensions not only unveils the dynamical traits contributing to the distinguishability of stochastic systems but also allows for the efficient evaluation of this divergence.

Given the flexibility in selecting Markov processes A and B, our theory is poised for broad application across various fields. It promises to extend the scope of fluctuation theorems from stochastic thermodynamics to encompass a wide range of non-thermal processes. Moreover, by analyzing trajectories timed by different mechanisms, our theory applies to clock calibration and offers insights into how a deterministic conception of time may arise from purely stochastic dynamics. Finally, by leveraging this theoretical framework to quantify information within stochastic trajectories, we lay the groundwork for innovative design principles for biological and artificial sensors to recognize and interpret temporal patterns in input signals.

Technical derivations that are too lengthy to be included in the manuscript and numerical simulation codes can be found in the supplementary material.

The National Science Foundation financially supports this work under Grant No. DMR-2145256. Z.L. is grateful to Professor Hong Qian at the University of Washington Seattle, Professor Ying Tang at Beijing Normal University, and Mr. Ruicheng Bao at the University of Science and Technology of China for their constructive comments on this work .

The authors have no conflicts to disclose.

A.P, Z.Z., and J.Z. contributed equally to this work and their names are ordered alphabetically.

Asawari Pagare: Conceptualization (equal); Formal analysis (equal); Methodology (equal); Writing – original draft (equal); Writing – review & editing (equal). Zhongmin Zhang: Conceptualization (equal); Formal analysis (equal); Methodology (equal); Writing – original draft (equal); Writing – review & editing (equal). Jiming Zheng: Conceptualization (equal); Formal analysis (equal); Methodology (equal); Writing – original draft (equal); Writing – review & editing (equal). Zhiyue Lu: Conceptualization (lead); Formal analysis (equal); Funding acquisition (lead); Investigation (equal); Writing – original draft (lead); Writing – review & editing (equal).

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

1.
D. T.
Gillespie
,
Markov Processes: An Introduction for Physical Scientists
(
Elsevier
,
1991
).
2.
R. F.
Fox
, “
Gaussian stochastic processes in physics
,”
Phys. Rep.
48
,
179
283
(
1978
).
3.
A.
Tamir
,
Applications of Markov Chains in Chemical Engineering
(
Elsevier
,
1998
).
4.
D. F.
Anderson
and
T. G.
Kurtz
, “
Continuous time Markov chain models for chemical reaction networks
,” in
Design and Analysis of Biomolecular Circuits: Engineering Approaches to Systems and Synthetic Biology
(
Springer
,
2011
), pp.
3
42
.
5.
L. J.
Allen
,
An Introduction to Stochastic Processes with Applications to Biology
(
CRC Press
,
2010
).
6.
H.
Ge
,
S.
Pressé
,
K.
Ghosh
, and
K. A.
Dill
, “
Markov processes follow from the principle of maximum caliber
,”
J. Chem. Phys.
136
,
010901
(
2012
).
7.
C. A.
Yates
,
M. J.
Ford
, and
R. L.
Mort
, “
A multi-stage representation of cell proliferation as a Markov process
,”
Bull. Math. Biol.
79
,
2905
2928
(
2017
).
8.
K. L.
Chung
,
Lectures from Markov Processes to Brownian Motion
(
Springer Science & Business Media
,
2013
), Vol.
249
.
9.
L.
Streit
, “
Quantum theory and stochastic processes—Some contact points
,” in
Stochastic Processes and Their Applications: Proceedings of the International Conference Held in Nagoya, July 2–6, 1985
(
Springer
,
2006
), pp.
197
213
.
10.
M. B.
Propp
, “
The thermodynamic properties of Markov processes
,” Ph.D. thesis,
Massachusetts Institute of Technology
,
1985
.
11.
M. L.
Mugnai
,
C.
Hyeon
,
M.
Hinczewski
, and
D.
Thirumalai
, “
Theoretical perspectives on biological machines
,”
Rev. Mod. Phys.
92
,
025001
(
2020
).
12.
C.
Van den Broeck
et al, “
Stochastic thermodynamics: A brief introduction
,”
Phys. Complex Colloids
184
,
155
193
(
2013
).
13.
A.
Faggionato
,
D.
Gabrielli
, and
M.
Ribezzi Crivellari
, “
Non-equilibrium thermodynamics of piecewise deterministic Markov processes
,”
J. Stat. Phys.
137
,
259
304
(
2009
).
14.
R. T. A.
Leenders
, “
Models for network dynamics: A Markovian framework
,”
J. Math. Sociol.
20
,
1
21
(
1995
).
15.
T. A.
Snijders
, “
The statistical evaluation of social network dynamics
,”
Sociol. Methodol.
31
,
361
395
(
2001
).
16.
R.
Lambiotte
,
J.-C.
Delvenne
, and
M.
Barahona
, “
Random walks, Markov processes and the multiscale modular organization of complex networks
,”
IEEE Trans. Network Sci. Eng.
1
,
76
90
(
2014
).
17.
P. D.
Dixit
,
A.
Jain
,
G.
Stock
, and
K. A.
Dill
, “
Inferring transition rates of networks from populations in continuous-time Markov processes
,”
J. Chem. Theory Comput.
11
,
5464
5472
(
2015
).
18.
E.
Cascetta
, “
A stochastic process approach to the analysis of temporal dynamics in transportation networks
,”
Transp. Res. Part B
23
,
1
17
(
1989
).
19.
P.
Érdi
and
G.
Lente
,
Stochastic Chemical Kinetics
(
Springer
,
2016
), Vol.
1
.
20.
L.
Carrillo
,
J. A.
Escobar
,
J. B.
Clempner
, and
A. S.
Poznyak
, “
Optimization problems in chemical reactions using continuous-time Markov chains
,”
J. Math. Chem.
54
,
1233
1254
(
2016
).
21.
J.-H.
Prinz
,
H.
Wu
,
M.
Sarich
,
B.
Keller
,
M.
Senne
,
M.
Held
,
J. D.
Chodera
,
C.
Schütte
, and
F.
Noé
, “
Markov models of molecular kinetics: Generation and validation
,”
J. Chem. Phys.
134
,
174105
(
2011
).
22.
J.
Goutsias
and
G.
Jenkinson
, “
Markovian dynamics on complex reaction networks
,”
Phys. Rep.
529
,
199
264
(
2013
).
23.
J. D.
Chodera
and
F.
Noé
, “
Markov state models of biomolecular conformational dynamics
,”
Curr. Opin. Struct. Biol.
25
,
135
144
(
2014
).
24.
F.
Knoch
and
T.
Speck
, “
Non-equilibrium Markov state modeling of periodically driven biomolecules
,”
J. Chem. Phys.
150
,
054103
(
2019
).
25.
E.
Meerbach
,
E.
Dittmer
,
I.
Horenko
, and
C.
Schütte
, “
Multiscale modelling in molecular dynamics: Biomolecular conformations as metastable states
,” in
Computer Simulations in Condensed Matter Systems: From Materials to Chemical Biology
(
2006
), Vol.
1
, pp.
495
517
.
26.
C.
Schütte
,
S.
Winkelmann
, and
C.
Hartmann
, “
Optimal control of molecular dynamics using Markov state models
,”
Math. Program.
134
,
259
282
(
2012
).
27.
B. K.
Chu
,
M. J.
Tse
,
R. R.
Sato
, and
E. L.
Read
, “
Markov state models of gene regulatory networks
,”
BMC Syst. Biol.
11
,
14
17
(
2017
).
28.
H. S.
Booth
,
C. J.
Burden
,
M.
Hegland
,
L.
Santoso
et al, “
Markov process modelling of gene regulation
,”
Austr. Math. Soc. Gazette
32
,
31
41
(
2005
).
29.
S.-Q.
Zhang
,
L. Y.
Wu
,
W. K.
Ching
,
Y.
Jiao
, and
R. H.
Chan
, “
Construction and control of genetic regulatory networks: A multivariate Markov chain approach
,”
J. Biomed. Sci. Eng.
01
,
15
21
(
2008
).
30.
D.
Materassi
,
S.
Roychowdhury
, and
M. V.
Salapaka
, “
A stochastic Markov model for coordinated molecular motors
,” arXiv:1012.4191 (
2010
).
31.
A.
Faggionato
,
D.
Gabrielli
, and
M. R.
Crivellari
, “
Averaging and large deviation principles for fully-coupled piecewise deterministic Markov processes and applications to molecular motors
,” arXiv:0808.1910 (
2008
).
32.
A.
Lampert
and
A.
Korngreen
, “
Markov modeling of ion channels: Implications for understanding disease
,”
Prog. Mol. Biol. Transl. Sci.
123
,
1
21
(
2014
).
33.
D.
Linaro
and
M.
Giugliano
, “
Markov models of ion channels
,” in
Encyclopedia of Computational Neuroscience
(
Springer
,
2022
), pp.
1961
1972
.
34.
C.
Slowey
and
Z.
Lu
, “
Sloppy gear mechanism for coupled stochastic transportation: From antiequilibrium flow to kinetic selectivity
,”
Phys. Rev. Res.
4
,
023234
(
2022
).
35.
P. C.
Bressloff
,
Stochastic Processes in Cell Biology
(
Springer
,
2014
), Vol.
41
.
36.
J.
Rengifo
,
R.
Rosales
,
A.
González
,
H.
Cheng
,
M. D.
Stern
, and
E.
Ríos
, “
Intracellular Ca2+ release as irreversible Markov process
,”
Biophys. J.
83
,
2511
2521
(
2002
).
37.
Y.
Tang
,
A.
Adelaja
,
F. X.-F.
Ye
,
E.
Deeds
,
R.
Wollman
, and
A.
Hoffmann
, “
Quantifying information accumulation encoded in the dynamics of biochemical signaling
,”
Nat. Commun.
12
,
1272
(
2021
).
38.
L.
Hahn
,
A. M.
Walczak
, and
T.
Mora
, “
Dynamical information synergy in biochemical signaling networks
,”
Phys. Rev. Lett.
131
,
128401
(
2023
).
39.
M.
Reinhardt
,
G. c. v.
Tkačik
, and
P. R.
ten Wolde
, “
Path weight sampling: Exact Monte Carlo computation of the mutual information between stochastic trajectories
,”
Phys. Rev. X
13
,
041017
(
2023
).
40.
A.
Pagare
,
S. H.
Min
, and
Z.
Lu
, “
Theoretical upper bound of multiplexing in biological sensory receptors
,”
Phys. Rev. Res.
5
,
023032
(
2023
).
41.
G.
Minas
,
D. J.
Woodcock
,
L.
Ashall
,
C. V.
Harper
,
M. R.
White
, and
D. A.
Rand
, “
Multiplexing information flow through dynamic signalling systems
,”
PLoS Comput. Biol.
16
,
e1008076
(
2020
).
42.
V.
Singh
and
I.
Nemenman
, “
Accurate sensing of multiple ligands with a single receptor
,” arXiv:1506.00288 (
2015
).
43.
L.
Peliti
and
S.
Pigolotti
,
Stochastic Thermodynamics: An Introduction
(
Princeton University Press
,
2021
).
44.
É.
Roldán
, “
Estimating the Kullback–Leibler divergence
,” in
Irreversibility and Dissipation in Microscopic Systems
(
Springer
,
2014
), pp.
61
85
.
45.
I. A.
Martínez
,
G.
Bisker
,
J. M.
Horowitz
, and
J. M.
Parrondo
, “
Inferring broken detailed balance in the absence of observable currents
,”
Nat. Commun.
10
,
3542
(
2019
).
46.
C.
Jarzynski
, “
Nonequilibrium equality for free energy differences
,”
Phys. Rev. Lett.
78
,
2690
(
1997
).
47.
G. E.
Crooks
, “
Entropy production fluctuation theorem and the nonequilibrium work relation for free energy differences
,”
Phys. Rev. E
60
,
2721
(
1999
).
48.
U.
Seifert
, “
Stochastic thermodynamics: Principles and perspectives
,”
Eur. Phys. J. B
64
,
423
431
(
2008
).
49.
C.
Maes
, “
Response theory: A trajectory-based approach
,”
Front. Phys.
8
,
229
(
2020
).
50.
M.
Frater
,
R.
Bitmead
,
R.
Kennedy
, and
B.
Anderson
, “
Rare events and reverse-time models
,” in
Proceedings of the 28th IEEE Conference on Decision and Control
(
IEEE
,
1989
), pp.
1180
1183
.
51.
H.
Ge
,
C.
Jia
, and
X.
Jin
, “
Martingale structure for general thermodynamic functionals of diffusion processes under second-order averaging
,”
J. Stat. Phys.
184
,
17
41
(
2021
).
52.
A.
Dechant
and
S.-i.
Sasa
, “
Fluctuation–response inequality out of equilibrium
,”
Proc. Natl. Acad. Sci. U. S. A.
117
,
6430
6436
(
2020
).
53.
J. M.
Horowitz
and
S.
Vaikuntanathan
, “
Nonequilibrium detailed fluctuation theorem for repeated discrete feedback
,”
Phys. Rev. E
82
,
061120
(
2010
).
54.
R.
Chetrite
and
S.
Gupta
, “
Two refreshing views of fluctuation theorems through kinematics elements and exponential martingale
,”
J. Stat. Phys.
143
,
543
584
(
2011
).
55.
R.
Chetrite
and
K.
Gawȩdzki
, “
Fluctuation relations for diffusion processes
,”
Commun. Math. Phys.
282
,
469
518
(
2008
).
56.
G.
Gallavotti
and
E. G. D.
Cohen
, “
Dynamical ensembles in nonequilibrium statistical mechanics
,”
Phys. Rev. Lett.
74
,
2694
(
1995
).
57.
M.
Baiesi
and
G.
Falasco
, “
Inflow rate, a time-symmetric observable obeying fluctuation relations
,”
Phys. Rev. E
92
,
042162
(
2015
).
58.
R.
Chetrite
, “
Fluctuation relations for diffusion that is thermally driven by a nonstationary bath
,”
Phys. Rev. E
80
,
051107
(
2009
).
59.
T. M.
Cover
,
Elements of Information Theory
(
John Wiley & Sons
,
1999
).
60.
Z.
Lu
and
O.
Raz
, “
Nonequilibrium thermodynamics of the Markovian Mpemba effect and its inverse
,”
Proc. Natl. Acad. Sci. U. S. A.
114
,
5083
5088
(
2017
).
61.
I.
Klich
,
O.
Raz
,
O.
Hirschberg
, and
M.
Vucelja
, “
Mpemba index and anomalous relaxation
,”
Phys. Rev. X
9
,
021060
(
2019
).
62.
J. J.
Hopfield
, “
Kinetic proofreading: A new mechanism for reducing errors in biosynthetic processes requiring high specificity
,”
Proc. Natl. Acad. Sci. U. S. A.
71
,
4135
4139
(
1974
).
63.
H.
Qian
, “
Reducing intrinsic biochemical noise in cells and its thermodynamic limit
,”
J. Mol. Biol.
362
,
387
392
(
2006
).
64.
I. M.
Kenney
and
O.
Beckstein
, “
Thermodynamically consistent determination of free energies and rates in kinetic cycle models
,”
Biophys. Rep.
3
,
100120
(
2023
).
65.
P. D.
Dixit
and
K. A.
Dill
, “
Inferring microscopic kinetic rates from stationary state distributions
,”
J. Chem. Theory Comput.
10
,
3002
3005
(
2014
).
66.
M.
Götz
,
A.
Barth
,
S. S.-R.
Bohr
,
R.
Börner
,
J.
Chen
,
T.
Cordes
,
D. A.
Erie
,
C.
Gebhardt
,
M. C.
Hadzic
,
G. L.
Hamilton
et al, “
A blind benchmark of analysis tools to infer kinetic rate constants from single-molecule fret trajectories
,”
Nat. Commun.
13
,
5402
(
2022
).
67.
N. W.
Barendregt
,
E. G.
Webb
, and
Z. P.
Kilpatrick
, “
Adaptive Bayesian inference of Markov transition rates
,”
Proc. R. Soc. A
479
,
20220453
(
2023
).

Supplementary Material