2. Bivariate Markov Chains
Markov chains have been often assumed for the underlying dynamics of the total system in random environments. When the two subsystems together jointly form a Markov chain in continuous or discrete time, the resulting chain is called the
Bivariate Markov Chain (BMC, a special case of the multivariate Markov chain with two stochastic variables). The processes of the two subsystems are correspondingly said to be marginal processes or a marginal chain. The BMC was used to model ion channel currents [
2]. It was also used to model delays and congestion in a computer network [
3]. Recently, different models of BMC appeared in nonequilibrium statistical physics for capturing or implementing Maxwell’s demon [
4,
5,
6], which can be seen as one marginal chain in the BMC playing feedback control to the other marginal chain. Although the BMC has been studied for decades, there are still challenges on quantifying the dynamics of the whole, as well as the two subsystems. This is because neither of them needs to be a Markovian chain in general [
7], and the quantifications of the probabilities (densities) for the trajectories of the two subsystems involve the complicated random matrix multiplications [
8]. This leads to the problem not exactly being analytically solvable. The corresponding numerical solutions often lack direct mathematical and physical interpretations.
The conventional analysis of the BMC focuses on the mutual information [
9] of the two subsystems for quantifying the underlying information correlations. There are three main representations of this. The first one was proposed and emphasized in the works of Sagawa, T. and Ueda, M. [
11] and Parrondo, J. M. R., Horowitz, J. M. and Sagawa, T. [
10], respectively, for explaining the mechanism of Maxwell’s demon in Szilard’s engine. In this representation, the mutual information between the demon and controlled system characterizes the observation and the feedback of the demon. This leads to an elegant approach, which includes the increment of the mutual information into a unified fluctuation relation. The second representation was proposed by the work of Horowitz, J. M. and Esposito, M. [
12] in an attempt to explain the violation of the second law in a specified BMC, the bipartite model, where the mutual information is divided into two parts corresponding to the two subsystems, respectively, which were said to be the information flows. This representation tries to explain the mechanism of the demon because one can see that the information flows do contribute to the entropy production for both the demon and controlled system. The first two representations are based on the ensembles of the subsystem states. This means that the mutual information is defined only on the time-sliced distributions of the system states, which somehow lack the information of subsystem dynamics: the time-correlations of the observation and feedback of the demon. The last representation was seen in the work of Barato, A. C., Hartich, D. and Seifert, U. [
13], where a more general definition of mutual information in information theory was used, which is defined on the trajectories of the two subsystems. More exactly, this is the so-called
Mutual Information Rate (MIR) [
24], which quantifies the correlation between the two subsystem dynamics. However, due to the difficulties from the possible underlying non-Markovian property of the marginal chains, exactly solvable models and comprehensive conclusions are still challenging from this representation.
In this study, we study the discrete-time BMC in both stochastic information dynamics and thermodynamics. To avoid the technical difficulty caused by non-Markovian dynamics, we first assume that the two marginal chains follow the Markovian dynamics. The non-Markovian case will be discussed elsewhere. We explore the time-irreversibility of BMC and marginal processes in the steady state. Then, we decompose the driving force for the underlying dynamics as the information landscape and information flux [
14,
15,
16], which can be used to describe the time-reversible parts and time-irreversible parts, respectively. We also prove that the non-vanishing flux fully describes the time-irreversibility of BMC and marginal processes.
We focus on the mutual information rate between the two marginal chains. Since the two marginal chains are assumed to be Markov chains here, the mutual information rate is exactly analytically solvable, which can be seen as the averaged conditional correlation between the two subsystem states. Here, the conditional correlations reveal the time correlations between the past states and the future states.
Corresponding to the landscape-flux decomposition in stochastic dynamics, we decompose the MIR into two parts: the time-reversible and time-irreversible parts, respectively. The time-reversible part measures the part of the correlations between the two marginal chains in both forward and backward processes of BMC. The time-irreversible part measures the difference between the correlations in forward and backward processes of BMC, respectively. We can see that a non-vanishing time-irreversible part of the MIR must be driven by a non-vanishing flux in the steady state, and it can be seen as the sufficient condition for a BMC to be time-irreversible.
We also reveal the important fact that the time-irreversible parts of MIR contribute to the nonequilibrium
Entropy Production Rate (EPR) of the BMC by the simple equality:
The decomposition of the MIR and the relation between the time-irreversible part of MIR and EPRs can also be found in stationary and ergodic non-Markovian cases, which will be given in the discussions in the Appendix. This may help to develop a general theory of nonequilibrium non-Markovian interacting information systems.
3. Information Landscape and Information Flux for Determining the Information Dynamics, Time-Irreversibility
Consider the case that two interacting information systems form a finite-state, discrete-time, ergodic and irreducible bivariate Markov chain,
We assume that the information state space of
X is given by
and the information state space of
S is given by
. The information state space of
Z is then given by
. The stochastic information dynamics can then be quantitatively described by the time evolution of the probability distribution of information state space
Z, characterized by the following master equation (or the information system dynamics) in discrete time,
where
is the probability of observing state
z (or joint probability of
and
) at time
t;
are the transition probabilities from
to
, respectively, and have
.
We assume that there exists a unique stationary distribution such that . Then, given an arbitrary initial probability distribution, the probability distribution goes to exponentially fast in time. If the initial distribution is , we say that Z is in Steady State (SS), and our discussion is based on this SS.
The marginal chains of
Z, i.e.,
X and
S, do not need to be Markov chains in general. For the simplicity of analysis, we assume that both marginal chains are Markov chains, and the corresponding transition probabilities are given by
and
(for
and
), respectively. Then, we have the following master equations (or the information system dynamics) for
X and
S, respectively,
and,
where
and
are the probabilities of observing
and
at time
t, respectively.
We consider that both Equations (3) and (4) have unique stationary solutions
and
, which satisfy
and
respectively. Furthermore, we assume that when
Z is in SS,
and
are also achieved. The relations between
,
and
read,
In the rest of this paper, we let , , and denote the time sequences of X, S and Z in time T, respectively.
To characterize the time-irreversibility of the Markov chain C in information dynamics in SS, we introduce the concept of probability flux. Here, we let C denote the arbitrary Markov chain in , and let c, , and denote arbitrary state of C, the stationary distribution of C, the transition probabilities of C and a time sequence of C in time T and in SS, respectively.
The averaged number transitions from the state
to state
c, denoted by
, in unit time in SS can be obtained as:
This is also the probability of the time sequence
, (
). Correspondingly, the averaged number of reverse transitions, denoted by
, reads:
This is also the probability of the time-reverse sequence
, (
). The difference between these two transition numbers measures the time-reversibility of the forward sequence
in SS,
Then,
is said to be the probability flux from
to
c in SS. If
for arbitrary
and
c, then
(
) is time-reversible; otherwise, when
,
is time-irreversible. Clearly, we have from Equation (
6) that:
The transition probability determines the evolution dynamics of the information system. We can decompose the transition probabilities
into two parts: the time-reversible part
and time-irreversible part
, which read:
From this decomposition, we can see that the information system dynamics is determined by two driving forces. One of the driving forces is determined by the steady state probability distribution. This part of the driving force is time-reversible. The other driving force for the information dynamics is the steady state probability flux, which breaks the detailed balance and quantifies the time-irreversibility. Since the steady state probability distribution measures the weight of the information state, therefore it can be used to quantify the
information landscape. If we define the potential landscape for the information system as
, then the driving force
is expressed in term of the difference of the potential landscape. This is analogous to the landscape-flux decomposition of Langevin dynamics in [
15]. Notice that the information landscape is directly related to the steady state probability distribution of the information system. In general, the information landscape is at nonequilibrium since the detailed balance is often broken for general cases. Only when the detailed balance is preserved, the nonequilibrium information landscape is reduced to the equilibrium information landscape. Even though the information landscape is not at equilibrium in general, the driving force
is time-reversible due to the decomposition construction. The steady state probability flux measures the information flow in the dynamics and therefore can be termed as the
information flux. In fact, the nonzero information flux explicitly breaks the detailed balance because of the net flow to or from the system. It is therefore a direct measure of the degree of the nonequilibrium or time-irreversibility in terms of the detailed balance breaking.
Note that the decomposition for the discrete Markovian information process can be viewed as the separation of the current corresponding to the
here and the activity corresponding to the
in a previous study [
19]. The landscape and flux decomposition here for the reduced information dynamics are in a similar spirit as the whole state space decomposition with the information system and the associated environments. When the detailed balance is broken, the information landscape (defined as the negative logarithm of the steady state probability
) is not the same as the equilibrium landscape under the detailed balance. There can be uniqueness issue related to the decomposition. To avoid the confusion, we make a physical choice, or in other words, we can fix the gauge so that the information landscape always coincides with the equilibrium landscape when the detailed balance is satisfied. In other words, we want to make sure the Boltzmann law applies at equilibrium with detailed balance. In this way, we can decompose the information landscape and information flux for nonequilibrium information systems without detailed balance. By solving the linear master equation for the steady state, we can quantify the nonequilibrium information landscape, and from that, we can obtain the corresponding steady state probability flux. Some studies discussed various aspects of this issue [
18,
19,
27,
28].
By Equations (7) and (8), we have the following relations:
As we can see in the next section,
and
are useful for us to quantify time-reversible and time-irreversible observables of
C, respectively.
We give the interpretation that the non-vanishing information flux
fully measures the time-irreversibility of the chain
C in time
T for
. Let
be an arbitrary sequence of
C in SS, and without loss of generality, we let
. Similar to Equation (
6), the measure of the time-irreversibility of
can be given by the difference between the probability of
and that of its time-reversal
, such as:
Then, by the relations given in Equation (
9), we have that
holds for arbitrary
if and only if
or equivalently
. This conclusion can be made for arbitrary
. Thus, non-vanishing
can fully describe the time-irreversibility of
C for
or
Z.
We show the relations between the fluxes of the whole system
and of the subsystem
as follows:
Similarly, we have:
These relations indicate that the subsystem fluxes
and
can be seen as the coarse-grained levels of total system flux
by averaging over the other parts of the system
S and
X, respectively. We should emphasize that non-vanishing
does not mean
X or
S is time-irreversible and vice versa.
4. Mutual Information Decomposition to Time-Reversible and Time-Irreversible Parts
According to information theory, the two interacting information systems represented by bivariate Markov chain
Z can be characterized by the
Mutual Information Rate (MIR) between the marginal chains
X and
S in SS. The mutual information rates represent the correlation between two interacting information systems. The MIR is defined on the probabilities of all possible time sequences,
,
and
and is given by [
24],
It measures the correlation between
X and
S in unit time, or say, the efficient bits of information that
X and
S exchange with each other in unit time. The MIR must be non-negative, and a vanishing
indicates that
X and
S are independent of each other. More explicitly, the corresponding probabilities of these sequences can be evaluated by using Equations (2)–(4); we have:
By substituting these probabilities into Equation (
12) (see
Appendix A), we have the exact expression of MIR as:
where
is the conditional (Markovian) correlation between the states
x and
s when the transition
occurs. This indicates that when the two marginal processes are both Markovian, the MIR is the average of the conditional (Markovian) correlations. These correlations are measurable when transitions occur, and they can be seen as the observables of
Z.
By noting the decomposition of transition probabilities in Equation (
8), we have a corresponding decomposition of
such as:
This means that the mutual information representing the correlations between the two interacting systems can be decomposed into the time-reversible equilibrium part and the time-irreversible nonequilibrium part. The origin of this is from the fact that the underlying information system dynamics is determined by both the time-reversible information landscape and time-irreversible information flux. These equations are very important to establish the link to the time-irreversibility. We now give further interpretation for
and
:
Consider a bivariate Markov chain
Z in SS wherein
X and
S are dependent on each other, i.e.,
. By the ergodicity of
Z, we have the MIR, which measures the averaged conditional correlation along the time sequences
,
Then,
measures the change of the averaged conditional correlation between
X and
S when a sequence of
Z turns back in time,
A negative
shows that the correlation between
X and
S becomes strong in the time-reversal process of
Z; A positive
shows that the correlation becomes weak in the time-reversal process of
Z. Both cases show that the
Z is time-irreversible since we have a non-vanishing
. However, the case of
is complicated, since it indicates either a vanishing
or a non-vanishing
. Anyway, we see that a non-vanishing
is a sufficient condition for
Z to be time-irreversible. On the other hand,
measures the correlation remaining in the backward process of
Z.
The definition of MIR in Equation (
12) turns out to be appropriate for even more general stationary and ergodic (Markovian or non-Markovian) processes. Consequentially, the decomposition of MIR is useful to quantify the correlation between two stationary and ergodic processes in a wider sense, i.e., to monitor the changes of the correlation in the forward and the backward processes. As a special case, the analytical expressions in Equation (
14) are the reduced results, which are valid for Markovian cases. A brief discussion of the decomposition of MIR of more general processes can be found in
Appendix B.
5. Relationship between Mutual Information and Entropy Production
The
Entropy Production Rates (EPR) or energy dissipation (cost) rate at steady state is a quantitative nonequilibrium measure, which characterizes the time-irreversibility of the underlying processes. The EPR of a stationary and ergodic process
C (here
or
S) can be given by the difference between the averaged surprisal (negative logarithmic probability) of the backward sequences
and that of forward sequences
in the long time limit, i.e.,
where
is said to be the EPR of
C [
25];
and
are said to be the surprisal of a forward and a backward sequence of
C, respectively. We see that
C is time-reversible (i.e.,
for arbitrary
for large
T) if and only if
. Additionally, this is due to the form of
, which is exactly a Kullback–Leibler divergence. When
C is Markovian, then
reduces into the following form when
Z,
X or
S is assigned to
C, respectively [
17,
26],
where total and subsystem entropy productions
,
and
correspond to
Z,
X and
S, respectively. Here,
usually contains the detailed interaction information of the system (or subsystems) and environments;
and
provide the coarse-grained information of time-irreversible observables of
X and
Z, respectively. Each non-vanishing EPR indicates that the corresponding Markov chain is time-irreversible. Again, we emphasize that a non-vanishing
does not mean
X or
S is time-irreversible and vice versa.
We are interested in the connection between these EPRs and mutual information. We can associate them with
by noting Equations (10), (11) and (14). We have:
We note that
is intimately related to the EPRs. This builds up a bridge between these EPRs and the irreversible part of the mutual information. Moreover, we also have:
This indicates that the time-irreversible MIR contributes to the detailed EPRs. In other words, the differences of the entropy production rate of the whole system and subsystems provide the origin of the time-irreversible part of the mutual information. This reveals the nonequilibrium thermodynamic origin of the irreversible mutual information or correlations. Of course, since the EPR is related to the flux directly as is seen from the above definitions, the origin of the EPR or nonequilibrium thermodynamics is from the non-vanishing information flux for the nonequilibrium dynamics. On the other hand, the irreversible part of the mutual information measures the correlations, and it contributes to the EPRs of the correlated subsystems.
Furthermore, the last expression in Equation (
17) (also the expressions in Equation (
18)) can be generalized to more general stationary and ergodic processes. A related discussion and demonstration of this can be seen in
Appendix B.
6. A Simple Case: The Blind Demon
As a concrete example, we consider a two-state system coupled to two information baths a and b. The states of the system are denoted by , respectively. Each bath sends an instruction to the system. If the system adopts one of them, it then follows the instruction and makes the change of the state. The instructions generated from one bath are independently and identically distributed (Bernoulli trials). Both the probability distributions of the instructions corresponding to the baths follow Bernoulli distributions and read for bath a and for bath b, respectively. Since the system cannot execute two instructions simultaneously, there exists an information demon that makes choices for the system. The demon is blind to caring about the system, and it makes choices independently and identically distributed. The choices of the demon are denoted by , respectively. The probability distribution of the demon’s choices reads . Still, we use with and to denote the BMC of the system and the demon.
Consequentially, the transition probabilities of the system read:
The transition probabilities of the demon read:
Additionally, the transition probabilities of the joint chain read:
We have the corresponding steady state distributions or the information landscapes as,
We obtain the information fluxes as,
Here, we use the notations
and
(
) to denote the probabilities of the instructions
or
x from bath
a or
b briefly. We obtain the EPRs as:
We evaluate the MIR as:
The time-irreversible part of
reads,