Next Article in Journal
A New Transformation Technique for Reducing Information Entropy: A Case Study on Greyscale Raster Images
Next Article in Special Issue
Planetary Energy Flow and Entropy Production Rate by Earth from 2002 to 2023
Previous Article in Journal
Stability Analysis of a Delayed Rumor Propagation Model with Nonlinear Incidence Incorporating Impulsive Vaccination
Previous Article in Special Issue
Energy Conversion and Entropy Production in Biased Random Walk Processes—From Discrete Modeling to the Continuous Limit
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nonequilibrium Effects on Information Recoverability of the Noisy Channels

1
State Key Laboratory of Electroanalytical Chemistry, Changchun Institute of Applied Chemistry, Changchun 130022, China
2
Center for Theoretical Interdisciplinary Sciences, Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou 325001, China
3
Department of Chemistry, State University of New York, Stony Brook, NY 11794, USA
4
Department of Physics and Astronomy, State University of New York, Stony Brook, NY 11794, USA
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(12), 1589; https://doi.org/10.3390/e25121589
Submission received: 22 April 2023 / Revised: 23 August 2023 / Accepted: 18 November 2023 / Published: 27 November 2023
(This article belongs to the Collection Disorder and Biological Physics)

Abstract

:
We investigated the impact of nonequilibrium conditions on the transmission and recovery of information through noisy channels. By measuring the recoverability of messages from an information source, we demonstrate that the ability to recover information is connected to the nonequilibrium behavior of the information flow, particularly in terms of sequential information transfer. We discovered that the mathematical equivalence of information recoverability and entropy production characterizes the dissipative nature of information transfer. Our findings show that both entropy production (or recoverability) and mutual information increase monotonically with the nonequilibrium strength of information dynamics. These results suggest that the nonequilibrium dissipation cost can enhance the recoverability of noise messages and improve the quality of information transfer. Finally, we propose a simple model to test our conclusions and found that the numerical results support our findings.

1. Introduction

The transfer of information is a crucial topic in physics, psychology, and human society, making it an important area of study for both science and technology [1,2,3,4,5]. Specifically, information theory has proven to be essential in biology, with Shannon’s pioneering work being applied to many biological systems across various scales [6,7,8]. Our central challenge is to understand how organisms can extract and represent useful information optimally from noisy channels, given physical constraints. However, the issues related to information transformation are not limited to biological systems and should be examined across a wide range of systems in physics, chemistry, and engineering. It is, therefore, essential to consider the information transfer in a more-fundamental way.
During the transmission of information, messages are frequently conveyed through noisy channels, where the noise can alter the useful information carried by the messages. Consequently, information recovery becomes a critical issue in practice [9,10,11]. Although there has been considerable research on the performance analysis of specific information recovery techniques in various fields [12,13,14,15,16], there is still a lack of a general theory on the recoverability of useful information. The quantification of recoverability should depend on the information source and the channel’s characteristics, irrespective of the specific information recovery methodologies, such as the maximum likelihood [17,18,19] or maximum a posteriori probability [20,21,22].
In [23], a classical measurement model was developed to investigate nonequilibrium behavior resulting from energy and information exchange between the environment and system. The study revealed that the Markov dynamics for sequential measurements is governed by the information driving force, which can be decomposed into two parts: the equilibrium component, which preserves time reversibility, and the nonequilibrium component, which violates time reversibility. In this work, we considered the information dynamics of noisy channels based on this nonequilibrium setup. From the perspective of nonequilibrium dynamics and thermodynamics, the information source and channel, along with their complexity, are not isolated events, but rather should be regarded as open systems [24,25]. The environment, which can exchange energy and information with these systems, can significantly influence the information transfer and its recoverability. Due to the complexity and environmental impacts, often manifesting as the stochasticity and randomness of noisy received messages, sequential information transfer should be viewed as a nonequilibrium process, leading to nonequilibrium behavior in the corresponding information dynamics. Intuitively, there must be an underlying relationship between the nonequilibrium nature and information recoverability.
This study aimed to quantify the recoverability of messages transmitted through a channel. With appropriate physical settings, we show that the randomness of the information transfer can arise from the interactions between the information receiver and the information source and from the influences of the environmental noise. This randomness can be characterized by the steady state distributions or transition probabilities of the received messages given by the receiver under different potential profiles exerted by the source, which are identified as the information driving force of the measurement dynamics [12,23]. In the sequential information transfer, the source messages randomly appear in time sequences. As a result, the receiver will change the interactions with the source messages in time. This leads to the switchings among the potential profiles and among the corresponding steady state distributions of the received messages. Then, the nonequilibrium can arise when the receiver switches its potential profiles in the sequential information transfer. We identified the nonequilibrium of the information transfer as the nonequilibrium strength, given by the difference between the information driving forces (based on the steady state transition probabilities) of the information transfer dynamics.
The main novel contributions of this work are three-fold. First, we introduce a new metric, called recoverability, to quantify the ability to recover the source messages through noisy channels. Recoverability is measured by the averaged log ratio of channel transition probabilities. Second, we demonstrate the mathematical equivalence between recoverability and entropy production [4,26,27,28,29] in nonequilibrium information dynamics. This reveals the intrinsic connection between information recoverability and thermodynamic dissipation costs. Finally, We prove that both recoverability and mutual information [30] monotonically increase with the nonequilibrium strength. This elucidates how driving a system further from equilibrium can boost information transfer quality. To our knowledge, these quantitative relationships between recoverability, entropy production, and nonequilibrium strengths have not been fully established in prior work. Our framework and analysis provide novel theoretical insights into the nonequilibrium physics of information transmission through noisy channels.
This new approach is not specific to particular models, and our conclusions are generalizable for characterizing information transfer. To further support our findings, we propose a simple information transfer model, which yields numerical results consistent with our conclusions. Lastly, we discuss the potential applications of our work in biophysics.

2. Information Dynamics

2.1. Physical Settings

Information dynamics naturally arises from information transfer, information processing, and sequential physical measurements. As discussed in the Introduction, the Markovian sequential measurement model can be properly applied to the information dynamics. Let us firstly introduce the physical settings of sequential information transfer.
In information transfer, a receiver gains information from a source through physical interaction. The information source randomly selects a message X from a set X of n different messages, according to an a priori distribution P X on X . The interaction with the source causes the receiver to feel a potential U. The profile of U is determined by the message X, which is denoted by U X . The potential profile U X contains n distinguishable wells with different locations, corresponding to the n possible messages of the source. The well with the lowest depth is unique, and its location changes as the potential profile is switched among different messages. The location of the lowest well represents the position of the current message. An illustration for the potential profile is given in Figure 1. The receiver is immersed in an environment with reciprocal temperature β .
To obtain a message X i 1 at time i 1 , the receiver interacts with the source for a period of time τ 1 . The receiver’s position in the potential U changes much more quickly than the source message changes. This allows the receiver to quickly reach equilibrium with the environment while the source message remains fixed during this period. The position of the receiver in the potential U X i 1 at steady state, denoted by Y i , is viewed as the received message. If Y i is found at the lowest well in U X i 1 , the received message is correct ( Y i = X i 1 ). However, environmental noise can drive the receiver out of the lowest well and across the potential barriers. In this case, the receiver is found at another well located at Y i X i 1 , and this is a wrongly received message. The information transition probability, given by the conditional probability of Y when the source’s message is X, can be determined by the following canonical distribution:
q Y | X ( Y i | X i 1 ) = exp β [ F X i 1 U X i 1 ( Y i ) ] .
where F X i 1 = β 1 log { Y i exp [ β U X i 1 ( Y i ) ] } denotes the free energy decided by the profile U X i 1 . The information transition probability distribution q Y | X in Equation (1) quantifies the classical uncertainty of the information transfer channels, and it is influenced by the interaction between the receiver and the source, which is characterized by the potential U X i 1 , and the environmental noise, which is characterized by the reciprocal temperature β .
In practice, the source often transmits messages randomly in time. These messages then come into a time sequence, such as X 1 , X 2 , , X t . The corresponding time sequence of the received messages is obtained as Y 2 , Y 3 , , Y t + 1 , where each message Y i is obtained from X i 1 via the interaction described in the above. For the sake of simplicity, we assumed that the source selects message X i independently of the previously message X i 1 . The received message Y i + 1 is then only determined by the message X i according to the transition probability q Y | X ( Y i + 1 | X i ) , and it is independent of both Y i and X i 1 . This means that the received message is not influenced by the previous received message or the previous source message. If the a priori distribution of the source messages P X is time-invariant, then the probability distribution of the received message is stationary, which is given by the following probability identity:
P Y ( Y i ) = X i 1 X q Y | X ( Y i | X i 1 ) P X ( X i 1 ) .

2.2. Markovian Dynamics of Sequential Information Transfer

Since the time sequences of the source messages are assumed to be stationary and due to the ergodicity, we can formulate the sequential information transfer as a coarse-grained Markov process of the received messages within one receiving period.
When a message Y i has been received corresponding to the source message X i 1 , the source transmits the next message X i , which interacts with the receiver immediately. Consequentially, the potential profile is changed from U X i 1 to U X i promptly. Meanwhile, the receiver position in the potential remains at the previous Y i temporarily, and Y i is recognized as the initial position under the new potential U X i . The conditional joint probability of X i and Y i under the previous source message X i 1 in this potential-switching event is given by P ( Y i , X i | X i 1 ) = q Y | X ( Y i | X i 1 ) P ( X i | X i 1 ) . Here, P ( X i | X i 1 ) represents the transition probability from X i 1 to X i and is given by P ( X i | X i 1 ) = P X ( X i ) , because X i has been assumed to be independent of the previous X i 1 .
Then, the initial probability of the receiver position in the new potential profile U X i , denoted by u i ( Y i | X i 1 , X i ) (it can depend on the previous source message in general cases), is given as follows: u i ( Y i | X i 1 , X i ) = P ( Y i , X i | X i 1 ) / P ( X i | X i 1 ) = q Y | X ( Y i | X i 1 ) . This means that u i is equal to the equilibrium distribution, which is only conditioning on the previous source message exactly.
During the message-receiving period, the source message X i remains unchanged within this period. The transition rates of the receiver between two possible positions s and s in the potential U X i can be represented by r ( s | s , X i ) . The Markov dynamics of the receiver position under X i in continuous time can be given by the following master equation τ u τ = G u τ , where u τ is the distribution of the receiver position at time τ and G is the transition matrix composed of the transition rates r.
Provided the initial distribution u i , the solution of this master equation is given by u τ = exp ( G τ ) u i . Since the receiver is interacting with a single environmental heat bath, then the transition rates r between two positions satisfy the local detailed balance condition, r ( s | s , X i ) / r ( s | s , X i ) = exp [ β ( U X i ( s ) U X i ( s ) ) ] . The final distribution after a long enough message receiving period τ can achieve the equilibrium distribution, i.e., u f = lim τ exp ( G τ ) u i = q Y | X = X i . Here, we take τ = 1 for the long enough time to reach the equilibrium. At the end of the period, we take the position Y i + 1 as the new received message. An illustration of the sequential information transfer is given in Figure 2.
According to the continuous-time dynamics, the matrix K = lim τ exp ( G τ ) is recognized as the transition probability matrix for the information transfer dynamics at a coarse-grained level. The transition probabilities can be given by K ( Y i + 1 , Y i ) k ( Y i + 1 | Y i , X i ) = q Y | X ( Y i + 1 | X i ) , which is recognized as the information driving force [4,29] from the initial position Y i to the final one Y i + 1 within a period.
Due to the description in the above, we obtain the master equation of this coarse-grained Markov process in discrete time as follows:
u f ( Y i + 1 | X i ) = Y i k ( Y i + 1 | Y i , X i ) u i ( Y i | X i ) ,
where
u f ( Y i + 1 | X i ) = q Y | X ( Y i + 1 | X i ) , final equilibrium distribution , k ( Y i + 1 | Y i , X i ) = q Y | X ( Y i + 1 | X i ) , information driving force , u i ( Y i | X i ) = q Y | X ( Y i | X i 1 ) , initial distribution .

3. Information Recoverability

3.1. Decision Rules of Information Recovery

In this section, we will discuss two decision rules used to recover information. Due to the stochasticity of the noisy channel, two different source messages X i 1 and X i 1 can be both transformed into the same received message Y i by the channel, according to the information driving forces q Y | X ( Y i | X i 1 ) and q Y | X ( Y i | X i 1 ) (see Equations (1) and (3)), respectively. If one extracts the original message X i 1 from the received message Y i , a decision rule should be employed to justify that Y i originated from X i 1 rather than another message X i 1 X i 1 .
There are mainly two kinds of decision rules frequently used for recovering the source messages. The first kind is the maximum likelihood rule, that is choosing the message X i 1 such that
l Y i ( X i 1 , X i 1 ) = log q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i 1 ) > 0 , for all X i 1 X i 1 .
The maximum likelihood rule is suitable for the cases where the a priori distribution P X is uniform or P X is not important for the information transfer.
The second kind of decision rule is the so-called maximum a posteriori probability. This means selecting the message X i 1 such that
r Y i ( X i 1 , X i 1 ) = log P X | Y ( X i 1 | Y i ) P X | Y ( X i 1 | Y i ) > 0 , for all X i 1 X i 1 .
where the a posteriori probability follows the Bayes equation as
P X | Y ( X i 1 | Y i ) = P ( X i 1 , Y i ) P Y ( Y i ) = q Y | X ( Y i | X i 1 ) P X ( X i 1 ) X i 1 q Y | X ( Y i | X i 1 ) P X ( X i 1 ) .
If the a priori distribution P X is not uniform and P X carries the significant information of the message, then the maximum a posteriori probability rule is more appropriate than the maximum likelihood rule for the message recovery.
The connection of these two decision rules is that the log ratio r in Equation (5) carries additional information about the a priori distribution P X compared to the log ratio l in Equation (4) as follows:
r Y i ( X i 1 , X i 1 ) = l Y i ( X i 1 , X i 1 ) + γ ,
where γ = log P X ( X i 1 ) P X ( X i 1 ) is the log ratio of the a priori probabilities. When the message X i 1 is transmitted and the log ratio l satisfies Equation (4) or r satisfies Equation (5), then X i 1 can be recovered correctly. Otherwise, if another message X ^ i 1 X i 1 maximizes the likelihood or the a posteriori probability, then X ^ i 1 rather than the original message X i 1 can be chosen by the decision rules incorrectly.

3.2. Information Recoverability of Noisy Channels

The decision rules can be used to justify the transferred information. However, the ability of the noisy channels to recover the transferred information is still unclear. We need a new physical quantity to quantify the recoverability of the source messages. Obviously, this quantity should be independent of the decision rules. In this subsection, we will discuss the definition of information recoverability for the noisy channels.
For illustration, we firstly considered the maximum likelihood decision rule. One should be aware that the information driving force or transition probability q Y | X in Equation (1) works as the key characterization to decide the performances of the decision rules, if the a priori distribution P X is fixed. From the perspective of information theory, the log ratio l in Equation (4) is a fundamental entity to quantify the recoverability of the source messages, because l merely depends on the information driving force q Y | X . Intuitively, while the a priori distribution P X is fixed, the inequality l Y i ( X i 1 , X i 1 ) 0 indicates the situation that X i 1 is completely unrecoverable. On the other hand, as l Y i ( X i 1 , X i 1 ) increases in the positive regime, the a posteriori distribution P X | Y (see Equation (6)) conditioned on Y i tends to be more concentrated on X i 1 than X i 1 X i 1 , and hence, X i 1 becomes more recoverable. On the contrary, as l Y i ( X i 1 , X i 1 ) decreases, then X i 1 becomes less recoverable when Y i is received.
To address our idea more clearly, we rewrite l Y i ( X i 1 , X i 1 ) in the following form:
l Y i ( X i 1 , X i 1 ) = i ( X i 1 , Y i ) + i e ( X i 1 , Y i ) .
with
i ( X i 1 , Y i ) = log q Y | X ( Y i | X i 1 ) P Y ( Y i ) , stochastic mutual information , i e ( X i 1 , Y i ) = log q Y | X ( Y i | X i 1 ) P Y ( Y i ) , stochastic error information , P Y ( Y i ) is given in Equation ( 2 ) .
Here, i ( X i 1 , Y i ) and i e ( X i 1 , Y i ) quantify the stochastic mutual information of the messages X i 1 and X i 1 contained in the received message Y i , respectively. When X i 1 is transmitted and Y i is received, i ( X i 1 , Y i ) is the useful information for recovering X i 1 . On the other hand, i e ( X i 1 , Y i ) quantifies the noise-induced error, which introduces a spurious correlation between Y i and X i 1 . The negative sign in i e represents the part of the useful information from the correct source message X i 1 that is reduced by the error. Then, the log ratio l quantifies the remaining useful information of X i 1 while the error information i ( X i 1 , Y i ) reduces the useful information i ( X i 1 , Y i ) . The larger l becomes, the more useful information of the source message X i 1 is preserved in the received message Y i , and X i 1 becomes more recoverable.
With this consideration, we used the average of the log ratio l to describe the overall recoverability of the source messages, i.e., we can properly define information recoverability R for the maximum likelihood decision rule as follows:
R = l Y i ( X i 1 , X i 1 ) = X i 1 , X i 1 , Y i P X ( X i 1 ) P X ( X i 1 ) q Y | X ( Y i | X i 1 ) log q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i 1 ) ,
where the average is taken over the ensembles of X i 1 , X i 1 , and Y i .
Now, let us discuss the information recoverability for the maximum a posteriori probability decision rule. It can be verified that the average of the log ratio of the a posteriori probabilities r in the maximum a posteriori probability rule (see Equation (5)) is equal to the recoverability R, i.e., R = r Y i ( X i 1 , X i 1 ) , with the log ratio of the a priori probabilities γ (see Equation (7)) vanishing in the average. This implies that the entity R can be used as a new rationale for the characterization of the recoverability, which should not depend on concrete decision rules.

3.3. Information Transfer Rate Enhanced by Recoverability

In this subsection, we derive the novel result that the information transfer rate increases monotonically with recoverability, formally proving that enhancing the ability to recover messages also boosts the rate of reliable information transmission through noisy channels.
The relationship between the information transfer rate and the recoverability can be given by Equations (8) and (9) straightforwardly:
R = l Y i ( X i 1 , X i 1 ) = i ( X i 1 , Y i ) + i e ( X i 1 , Y i ) = I + I e 0
with
I = i ( X i 1 , Y i ) = X i 1 , Y i P X ( X i 1 ) q Y | X ( Y i | X i 1 ) log q Y | X ( Y i | X i 1 ) P Y ( Y i ) 0 I e = i e ( X i 1 , Y i ) = X i 1 , Y i P X ( X i 1 ) P Y ( Y i ) log q Y | X ( Y i | X i 1 ) P Y ( Y i ) 0
Here, I is recognized as the mutual information between the time sequences of the messages X = { X 1 , X 2 , , X t } and Y = { Y 1 , Y 2 , , Y t } ; I e is the averaged error information of the information dynamics with the meaning of I e interpreted in Equations (8) and (9). Since I and I e can be given in terms of the relative entropies as I = D K L { P ( X i 1 , Y i ) | | P Y ( Y i ) P X ( X i 1 ) } and I e = D K L { P ( X i 1 , Y i ) | | P Y ( Y i ) P X ( X i 1 ) } , then I and I e are both positively defined. Thus, the recoverability R is a nonnegative valued, due to Equation (11).
The mutual information I quantifies the useful information of the source messages contained in the received messages and characterizes the rate of the information transfer. It is also called the information transfer rate. Equation (11) implies that the information transfer rate I can be given as a function of the recoverability:
I = R I e .
We can verify that I is a monotonically increasing function with respect to R with a fixed a priori distribution P X . This is because both the mutual information I and recoverability R are a convex function of the information driving force q Y | X when P X is fixed (see [31] and Appendix A of this paper). They achieve the same minimum value of 0 at the points q Y | X ( Y i | X i 1 ) = q Y | X ( Y i | X i 1 ) , where all the source messages cannot be distinguished from each other. Then, I and R are both monotonically increasing functions in the same directions as q Y | X . Therefore, I is a monotonically increasing function of R.
From the convexity of the recoverability, one can easily obtain the following two observations. The first observation is that, if the source and received messages X i 1 and Y i are independent of each other, then all the source messages become unrecoverable. Otherwise, R increases monotonically while each absolute difference between two information driving forces | d Y i ( X i 1 , X i 1 ) | increases (see Appendix A). Here, the difference d is defined as
d Y i ( X i 1 , X i 1 ) = 1 2 [ q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i 1 ) ] .
On the other hand, the information transfer rate I is also a convex function over all the possible information driving forces q Y | X [32]. The proof of the convexity of I can be given by applying the log sum inequality. The unique minimum of I is provided by zero, and this minimum information transfer rate can be achieved if and only if the messages X i 1 and Y i are independent of each other, i.e., d Y i ( X i 1 , X i 1 ) = 0 for all X i 1 X i 1 . Thus, both the information transfer rate I and the recoverability R increase monotonically as each absolute difference | d Y i ( X i 1 , X i 1 ) | increases. This indicates that the information transfer can be enhanced by the recoverability monotonically due to Equation (13). The increase in the information transfer rate also implies the improvement of the information recovery, with smaller upper bounds of the error probabilities of the decision rules in Equations (4) and (5) [33].

4. Nonequilibrium Information Dynamics

In this section, we will develop the model of nonequilibrium information dynamics by introducing the time-reverse sequence of the sequential information transfer process. We will show that the recoverability is closely related to the nonequilibrium behavior of the Markovian information dynamics described in Equation (3). As we can see from the following discussions, the difference d given in Equation (14) works as the nonequilibrium information driving force of the information dynamics. The recoverability R can be shown as the thermodynamic dissipation cost or the entropy production of the information dynamics, driven by the nonequilibrium information driving force d.
The nonequilibrium behavior of the Markovian information dynamics (Equation (3)) can be seen intuitively from the following fact: When the source message is changed from X i 1 to X i , the profile of the potential is changed from U X i 1 to U X i accordingly. Since U X i is different from U X i 1 if X i X i 1 , the corresponding equilibrium distribution of the receiver message q Y | X = X i , which is conditioning on X i , can differ from the previous equilibrium distribution q Y | X = X i 1 , which is conditioning on X i 1 . This indicates that the receiver is driven out of equilibrium in the new potential U X i at the beginning of the message-receiving period. Then, the difference between two information driving forces under different potential profiles at the same received message, where the potential profile is switched, can reflect the degree of this nonequilibrium. With this consideration, we decompose the information driving force in Equation (3) into two parts [4,29]:
q Y | X ( Y i | X i 1 ) = d Y i ( X i 1 , X i ) + m Y i ( X i 1 , X i ) ,
with
d Y i ( X i 1 , X i ) = 1 2 [ q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i ) ] , nonequilibrium strength , m Y i ( X i 1 , X i ) = 1 2 [ q Y | X ( Y i | X i 1 ) + q Y | X ( Y i | X i ) ] , equilibrium strength .
In Equation (15), if the nonequilibrium strength d vanishes at every source and received message, then the information transfer process (Equation (3)) is at the equilibrium state, and the information driving force q Y | X degenerates to its equilibrium strength m. In this situation, the new potential U X i is equal to the previous potential U X i 1 plus a constant a, i.e., U X i = U X i 1 + a , and the receiver stays at the same equilibrium state under X i as that under X i 1 . Then, the nonequilibriumness in the information transfer vanishes, and the vanishing nonequilibrium strength d 0 = 0 at every source and received message is called the equilibrium point of the information transfer. Consequentially, the equilibrium point makes any two different source message X i and X i 1 indistinguishable from each other by the receiver (see the Decision Rules of Information Recovery Section). Clearly, this leads to invalid information transfer. On the other hand, the non-vanishing nonequilibrium strength d 0 indicates that not only the information transfer is a nonequilibrium process, but also some source message X can be distinguished from the others by the receiver.
In addition, the nonequilibrium strength d shows the same form as Equation (14), where we can regard the source message X i 1 in Equation (14) as X i in Equation (16) because the source select the message in the i.i.d. way. This d guarantees the convexity of the information recoverability (Equation (10), Appendix A of this paper). This means that the information recoverability is a non-decreasing function of the absolute nonequilibrium strength. Then, as a consequence, the information system is further driven away from the equilibrium state, which will give rise to the better recoverability. This leads to the connection between the nonequilibrium thermodynamics and the recoverability.

5. Nonequilibrium Information Thermodynamics

The nonequilibrium behavior of a system can give rise to a thermodynamic dissipation cost in the form of energy, matter, or information, which can be quantified by the entropy production [31,34,35]. In this section, we will discuss the entropy production of the nonequilibrium information dynamics. It is shown that the entropy production can be used to characterize the averaged dissipative information during the nonequilibrium process.
The entropy production can be written in terms of the work performed on the receiver. From the perspective of stochastic thermodynamics [18], for receiving an incoming message X i , a stochastic work w = U X i ( Y ( i ) ) U X i 1 ( Y ( i ) ) should be performed on the receiver to change the potential profile from U X i 1 to U X i . Meanwhile, the free energy change Δ F = F X i 1 F X i (the free energy is shown in Equation (1) quantifies the work to change the potential profile within an equilibrium information transfer. The difference between w and Δ F quantifies the energy dissipation within a nonequilibrium information transfer, which is recognized as the stochastic entropy production when receiving one single message. This stochastic entropy production can be shown as the logarithmic ratio between two transition probabilities at the same received message, as follows:
σ = β ( w Δ F ) = log q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i ) ,
where
w = U X i ( Y ( i ) ) U X i 1 ( Y ( i ) ) , stochastic work , Δ F = F X i 1 F X i , free energy difference .
On the other hand, the receiver can be regarded as a finite-state information storage. It stores the information from the source message, which may be corrupted by environmental noise. However, due to its finite memory, the receiver must erase the stored information of the source message X i 1 when a new source message X i is coming. According to the theory of information thermodynamics [1,2,3,4,5], the work w is needed to erase the information of X i 1 and to write the new information of X i in the receiver.
The information of the source message X i 1 stored in the receiver is quantified by the stochastic mutual information between the previous received message Y i and X i 1 , i = log q Y | X ( Y i | X i 1 ) P Y ( Y i ) (see Equation (9)). On the other hand, not all the work is used to erase the information. Due to environmental noise, a part of the work introduces error information, which shows the spurious correlation between Y i and the incoming source message X i , quantified by the negative stochastic mutual information i e = log q Y | X ( Y i | X i ) P Y ( Y i ) (see Equation (9)). Then, it can be shown that the stochastic entropy production in Equation (17) can be rewritten as the sum of the useful information i and the error information i e :
σ = i + i e ,
where
i = log q Y | X ( Y i | X i 1 ) P Y ( Y i ) , stochastic mutual information , i e = log q Y | X ( Y i | X i ) P Y ( Y i ) , stochastic error information , P Y ( Y i ) is given in Equation ( 2 ) .
Equation (18) indicates that the stochastic entropy production σ not only quantifies the work dissipation in one single information transfer, but also justifies the detailed recoverability of each received messages, as shown in Equation (8).
The average of the stochastic entropy production Σ = σ is the key characterization of a nonequilibrium process in thermodynamics, which is given as follows:
Σ = β ( W Δ F ) = I + I e = Y i , X i 1 , X i P X ( X i 1 ) P X ( X i ) q Y | X ( Y i | X i 1 log q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i ) 0 ,
where
W = w = β 1 Σ , averaged work , Δ F = Δ F , averaged free energy difference equals 0 in this case , I = i = X i 1 , Y i P X ( X i 1 ) q Y | X ( Y i | X i 1 ) log q Y | X ( Y i | X i 1 ) P Y ( Y i ) 0 , useful information of X i 1 , I e = i e = X i , Y i P X ( X i ) P Y ( Y i ) log P Y ( Y i ) q Y | X ( Y i | X i ) 0 , error information from X i .
Equation (19) expresses the first equality in terms of the averaged energy or work dissipation (see Equation (17)). Since the averaged free energy Δ F vanishes in this case, then the averaged work W is completely converted into heat and dissipated into the environment. Since the averaged entropy production Σ is shown as the relative entropy D K L ( q Y | X = X i 1 | | q Y | X = X i ) (the third equality in Equation (19)), then Σ is always nonnegative, which is equivalently shown as the work bound W Δ F = 0 . This is the thermodynamic second law for the information transfer. Here, the equal sign in the second law implies that the information transfer process is at the equilibrium point d = 0 (see Equation (16)) and the discussion below). Otherwise, if Σ > 0 , then the measurement is at a nonequilibrium point.
On the other hand, the second equality in Equation (19) establishes the bridge between the nonequilibrium energy dissipation and the information transfer. The entropy production is taken as the sum of the useful information, quantified by the mutual information I between the source and received information, and the error information I e . This expression shows that the entropy production works as the recoverability given in Equation (10):
Σ = R .
This equation shows a novel result that recoverability is equivalent to entropy production. This means that information retrieval is linked to thermodynamic costs. In other words, improving the recoverability means increasing the nonequilibrium dissipation cost.
In addition, since both I and I e are shown as the relative entropies in the forms of D K L ( q Y | X | | P Y ) 0 and D K L ( P Y | | q Y | X ) 0 , then I and I e are both nonnegative. It is seen that the error information I e 0 works as the interference, which reduces the recoverability of information transfer and enlarges the energy dissipation. For this reason, the work bound in the second law can be given more tightly by the acquired information β 1 I , shown as
W β 1 I 0 .
This lower bound indicates that, for valid information transfer I > 0 , we should perform a positive work of β 1 I at least to erase the previous transmitted information stored in the receiver. Here, Equation (21) is the so-called generalized Landauer principle [3,4], which gives the minimum work or the minimum entropy production estimation for valid information erasing and transfer in this model on the average level. Since both the entropy production and mutual information are a monotonically increasing function of the nonequilibrium strength, the entropy production or the work performed on the receiver is a monotonically increasing function of the valid information transfer. This indicates that the more-useful information of one source message is transferred, and more energy or work dissipation is needed.

6. Numerical Results

In this section, we will present a simple example that can test our previous conclusions numerically.
At first, we set the source messages to be X i = 1 , 2 , 3 and the possible received messages Y i = 1 , 2 , 3 . The a priori distribution P X is given by the following column vector:
P X = [ P X ( 1 ) ; P X ( 2 ) ; P X ( 3 ) ] T = [ 0.4610 ; 0.0753 ; 0.4637 ] T .
The information driving force or the transition probabilities are given by the following matrix:
q Y | X = 0.4410 0.5558 0.1562 0.4903 0.3848 0.3067 0.0687 0.0594 0.5371 ,
where q Y | X follows a given potential, which is shown in Equation (1)), and the labels of the columns and rows of q Y | X represent the indices of the source and received messages, respectively. We then evaluated the recoverability R along a stochastic time sequence of the source and received messages, Z = { ( X 1 , Y 2 ) , ( X 2 , Y 3 ) , ( X t , Y t + 1 ) } . This time sequence Z is generated by the given probabilities in Equations (22,23). Following the Markov nature, the probabilities of Z can be given by P ( Z ) = i = 1 t P X ( X i ) q Y | X ( Y i + 1 | X i ) . If the time t is long enough, then the sequence Z will enter the typical set of the joint sequences of the source and received messages [31]. This means that all the time averages of Z will converge to the typical statistics of the joint sequences of the source and received messages. However, if we generate another time sequence Z = { ( X 1 , Y 2 ) , ( X 2 , Y 3 ) , ( X t , Y t + 1 ) } , where the sequence of the received messages is the same as that in the typical sequence Z, but with the original sequence of the source messages being randomly shuffled, we then obtain a non-typical sequence Z with probability P ( Z ) = i = 1 t P X ( X i ) q Y | X ( Y i + 1 | X i ) . The recoverability R in the information transfer can be evaluated by the time average of the log ratio of the probabilities P ( Z ) to P ( Z ) :
R = 1 t log P ( Z ) P ( Z ) , for long enough t .
On the other hand, we can evaluate the average of the stochastic entropy production in Equation (17) along the same time sequence Z as follows:
Σ = 1 t i = 2 t log q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i ) , for long enough t .
The entropy production Σ converges to the same theoretical value with the recoverability R (Equation (10)) in the long time limit. The related results are shown Figure 3, which verify the fact that the thermodynamic dissipation cost works as the recoverability.
We demonstrate that both the mutual information I (Equation (12)) and the recoverability or the entropy production (Equation (19)) are convex functions of the nonequilibrium strength d (Equations (14) and (16), Appendix A of this paper). Due to the i.i.d. assumption in the Physical Settings Section, we can drop the time indices in the information driving force q X | Y . Then, the nonequilibrium decomposition of q X | Y in Equation (15) becomes
q Y | X ( y | x ) = m y ( x , x ) + d y ( x , x ) , for x x
with
m y ( x , x ) = 1 2 [ q Y | X ( y | x ) + q Y | X ( y | x ) ] , d y ( x , x ) = 1 2 [ q Y | X ( y | x ) q Y | X ( y | x ) ] .
where m and d are recognized as the equilibrium and nonequilibrium strengths, respectively.
By substituting Equation (24) into the expressions of I (Equation (12)) and R (Equation (19)), respectively, we have that
I = x , y P X ( x ) [ m y ( x , x ) + d y ( x , x ) ] log m y ( x , x ) + d y ( x , x ) x P X ( x ) [ m y ( x , x ) + d y ( x , x ) ] R = x , x , y P X ( x ) P X ( x ) d y ( x , x ) log m y ( x , x ) + d y ( x , x ) m y ( x , x ) d y ( x , x )
Here, we note that the total number of source messages need not be equal to that of the received messages in practice. Under this consideration, we set the source messages to be x = 1 , 2 (2 messages) and the possible received messages to be y = 1 , 2 , 3 (3 messages). Then, the information driving force q Y | X becomes a 3 × 2 matrix, which can be given by
q Y | X = q Y | X ( 1 | 1 ) q Y | X ( 1 | 2 ) q Y | X ( 2 | 1 ) q Y | X ( 2 | 2 ) q Y | X ( 3 | 1 ) q Y | X ( 3 | 2 ) .
According to the nonequilibrium decomposition given in Equation (24), for this case, we have the equilibrium and nonequilibrium strengths as follows:
m y = m y ( 1 , 2 ) = m y ( 2 , 1 ) = 1 2 [ q Y | X ( y | 1 ) + q Y | X ( y | 2 ) ] , d y = d y ( 1 , 2 ) = d y ( 2 , 1 ) = 1 2 [ q Y | X ( y | 1 ) q Y | X ( y | 2 ) ] .
Due to the nonnegativity and the normalization of the conditional probabilities, i.e., q Y | X 0 and y q Y | X ( y | x ) = 1 , the constraints on m and d can be given as follows:
m y 0 , y m y = 1 , and y d y = 0 .
By combining these constraints, we can obtain the inequality constraints on d:
m y d y m y , if m y < 1 / 2 m y 1 d y 1 m y , if m y 1 / 2 .
According to Equation (26), we have the explicit forms of the mutual information and the recoverability for this case as follows:
I = p y = 1 3 ( m y + d y ) log m y + d y m y + ( 2 p 1 ) d y + ( 1 p ) y = 1 3 ( m y d y ) log m y d y m y + ( 2 p 1 ) d y , R = 2 p ( 1 p ) y = 1 3 d y log m y + d y m y d y ,
where the a priori probabilities P X ( 1 ) = p and P X ( 2 ) = 1 p . For no loss of generality, we set m 1 = m 2 = m 3 = 1 / 3 for the numerical calculations. Then, we can obtain the inequality constraints on d given in Equation (28) more explicitly:
1 / 3 d 1 1 / 3 , 1 / 3 d 2 1 / 3 , and 1 / 3 d 1 + d 2 1 / 3 .
Here, we should note the identity d 3 = d 1 d 2 , which is due to the equality constraint on d in Equation (27). This yields the last inequality for d 1 + d 2 . We then randomly select the a priori probability p, which is shown in Table 1. We plot the mutual information I and recoverability R as functions of the nonequilibrium strengths ( d 1 , d 2 ) , which are shown in Figure 4 and Figure 5, respectively. These results show that I and R are both convex for ( d 1 , d 2 ) . We next plot I as the function of R and R as the function of I, which are shown in Figure 6 and Figure 7, respectively. These results show that both the recoverability and the mutual information are monotonically increasing functions of each other. This indicates that the increasing information recoverability can enhance the information transfer rate, while the increasing information transfer rate enhances the information recoverability.

7. Conclusions

In this study, we investigated the nonequilibrium effects on the information recovery. By considering the information dynamics of the sequential information transfer (Equation (3)), we can quantify the recoverability of the source messages as the averaged log ratio of the transition probabilities of the channel (Equations (10) and (19)). We see that the difference between the transition probabilities works as the nonequilibrium strength behind the information dynamics (Equation (15)). The recoverability can be shown as the thermodynamic dissipation cost or the entropy production of the information dynamics (Equations (19) and (20)), driven by the nonequilibrium information driving force. This shows that the dissipation cost is essential for the information recoverability. The recoverability increases monotonically as the nonequilibrium information driving force or the nonequilibrium strength increases. On the other hand, as a function of the recoverability (Equation (19)), the mutual information also increases monotonically as the nonequilibrium strength increases. This demonstrates that the nonequilibrium cost can boost the information transfer from the thermodynamic perspective. In a similar spirit, increasing the information transfer rate can improve the information recoverability. The numerical results support our conclusions.
Finally, we discuss some examples that may have potential applications of the model and conclusion in the present work. As is well known, information transfer plays an important role in biology. For example, in biological sensory adaptation, which is an important regulatory function possessed by many living systems, organisms continuously monitor the time-varying environments while simultaneously adjusting themselves to maintain their sensitivity and fitness in response [36]. In this process, the information is transferred from the stochastic environments to the sensory neurons in the brain towards the objects that can be treated as the noisy channels. It is obvious that our model provides a simplified version of this process. The recoverability determines the accuracy of the response of the living system to the environment. A similar process also happens at the cellular level, where cells sense the information from the environment and transmit it towards signal transduction cascades to transcription factors in order to survive in a time-varying environment. As a response, a suitable gene expression is then initiated [37]. Similarly, in gene regulation, the time-varying transcription factor profiles are converted into distinct gene expression patterns through specific promoter activation and transcription dynamics [38,39,40,41]. In this situation, X i 1 and Y i in Equation (3) can represent the input and the output states at a fixed time. In fact, because the transcription factor profiles are time-varying, the input signals and the corresponding output signals compose the complete trajectories of the input and output processes on a considered time interval [ 0 , t ] . In [42,43], it was shown that the mutual information can be used to quantify the cumulative amount of information exchanged along these trajectories. In the present work, we introduced the joint time sequence Z to describe the trajectories of the input and output processes. We believe that the model discussed in the present work can be applied to these above-mentioned situations. We will address the issues of the nonequilibrium recoverability in these examples in the near future.

Author Contributions

Design: J.W. and Q.Z. Modeling and Simulations: Q.Z.; Methodology, J.W.; Writing–Original draft, Q.Z.; Writing—Review & Editing, R.L. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

Q. Z. thanks the support from National Natural Science Foundation (12234019).

Data Availability Statement

Data available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Here, we prove the convexity of the recoverability in Equation (10).
Proof. 
Select two different information driving forces q Y | X ( 1 ) and q Y | X ( 2 ) . Then, the convex combination q Y | X ( λ ) = λ q Y | X ( 1 ) + ( 1 λ ) q Y | X ( 2 ) is also an information driving force, which is due to the nonnegativity and the normalization of the probability distributions. Then, we have
λ R 1 + ( 1 λ ) R 2 = λ X i 1 , X i 1 , Y i P X ( X i 1 ) P X ( X i 1 ) q Y | X ( 1 ) ( Y i | X i 1 ) log q Y | X ( 1 ) ( Y i | X i 1 ) q Y | X ( 1 ) ( Y i | X i 1 ) + ( 1 λ ) X i 1 , X i 1 , Y i P X ( X i 1 ) P X ( X i 1 ) q Y | X ( 2 ) ( Y i | X i 1 ) log q Y | X ( 2 ) ( Y i | X i 1 ) q Y | X ( 2 ) ( Y i | X i 1 ) X i 1 , X i 1 , Y i P X ( X i 1 ) P X ( X i 1 ) q Y | X ( λ ) ( Y i | X i 1 ) log q Y | X ( λ ) ( Y i | X i 1 ) q Y | X ( λ ) ( Y i | X i 1 ) = R λ .
The log sum inequality i a i log a i b i ( i a i ) log i a i i b i was applied in Equation (A1) [31]. Thus, R is convex with respect to q Y | X with the unique minimum point R = 0 if and only if q Y | X ( Y i | X i 1 ) = q Y | X ( Y i | X i 1 ) for all Y i and X i 1 X i 1 . Note that, in this paper, we chose the conventional rule that the convex function is a synonym for the concave-up function with a positive second derivative. This completes the proof.
Since both the mutual information I and the recoverability R are convex functions of the transition probabilities q Y | X , then both of them are the convex functions of the affine transformation of q Y | X , i.e.,
d Y i ( X i 1 , X i 1 ) = 1 2 [ q Y | X ( Y i | X i 1 ) q Y | X ( Y i | X i 1 ) ] , m Y i ( X i 1 , X i 1 ) = 1 2 [ q Y | X ( Y i | X i 1 ) + q Y | X ( Y i | X i 1 ) ] .
This means that both I and R are convex functions of the difference d by fixing every m. □

References

  1. Koski, J.V.; Kutvonen, A.; Khaymovich, I.M.; Ala-Nissila, T.; Pekola, J.P. On-Chip Maxwell’s Demon as an Information-Powered Refrigerator. Phys. Rev. Lett. 2015, 115, 260602. [Google Scholar] [CrossRef] [PubMed]
  2. McGrath, T.; Jones, N.S.; Ten Wolde, P.R.; Ouldridge, T.E. Biochemical Machines for the Interconversion of Mutual Information and Work. Phys. Rev. Lett. 2017, 118, 028101. [Google Scholar] [CrossRef]
  3. Sagawa, T.; Ueda, M. Fluctuation theorem with information exchange: Role of correlations in stochastic thermodynamics. Phys. Rev. Lett. 2012, 11, 180602. [Google Scholar] [CrossRef] [PubMed]
  4. Horowitz, J.M.; Esposito, M. Thermodynamics with Continuous Information Flow. Phys. Rev. X 2014, 4, 031015. [Google Scholar] [CrossRef]
  5. Parrondo, J.M.R.; Horowitz, J.M.; Sagawa, T. Thermodynamics of information. Nat. Phys. 2015, 11, 131–139. [Google Scholar] [CrossRef]
  6. Strong, S.P.; Koberle, R.; Van Steveninck, R.R.D.R.; Bialek, W. Entropy and Information in Neural Spike Trains. Phys. Rev. Lett. 1998, 80, 197. [Google Scholar] [CrossRef]
  7. Tkačik, G.; Bialek, W. Information processing in living systems. arXiv 2016, arXiv:1412.8752. [Google Scholar] [CrossRef]
  8. Petkova, M.D.; Tkačik, G.; Bialek, W.; Wieschaus, E.F.; Gregor, T. Optimal decoding of information from a genetic network. arXiv 2016, arXiv:1612.08084. [Google Scholar]
  9. Mark, B.L.; Ephraim, Y. Explicit Causal Recursive Estimators for Continuous-Time Bivariate Markov Chains. IEEE Trans. Signal Process. 2014, 62, 2709–2718. [Google Scholar] [CrossRef]
  10. Ephraim, Y.; Mark, B.L. Bivariate Markov Processes and Their Estimation. Found. Trends Signal Process. 2013, 6, 1–95. [Google Scholar] [CrossRef]
  11. Hartich, D.; Barato, A.C.; Seifert, U. Stochastic thermodynamics of bipartite systems: Transfer entropy inequalities and a Maxwell’s demon interpretation. J. Stat. Stat. Mech. 2014, 2014, P02016. [Google Scholar] [CrossRef]
  12. Zeng, Q.; Wang, J. Information Landscape and Flux, Mutual Information Rate Decomposition and Connections to Entropy Production. Entropy 2017, 19, 678. [Google Scholar] [CrossRef]
  13. Wang, J.; Xu, L.; Wang, E.K. Potential landscape and flux framework of nonequilibrium networks: Robustness, dissipation, and coherence of biochemical oscillations. Proc. Natl. Acad. Sci. USA 2008, 105, 12271–12276. [Google Scholar] [CrossRef]
  14. Wang, J. Landscape and flux theory of non-equilibrium dynamical systems with application to biology. Adv. Phys. 2015, 64, 1–137. [Google Scholar] [CrossRef]
  15. Li, C.; Wang, E.; Wang, J. Potential flux landscapes determine the global stability of a Lorenz chaotic attractor under intrinsic fluctuations. J. Chem. Phys. 2012, 136, 194108. [Google Scholar] [CrossRef] [PubMed]
  16. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  17. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley: New York, NY, USA, 2003. [Google Scholar]
  18. Seifert, U. Stochastic thermodynamics, fluctuation theorems and molecular machines. Rep. Prog. Phys. 2012, 75, 126001. [Google Scholar] [CrossRef]
  19. Jacod, J.; Protter, P. Probability Essentials; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
  20. Wu, S.J.; Chu, M.T. Markov chains with memory, tensor formulation, and the dynamics of power iteration. Appl. Math. Comput. 2017, 303, 226–239. [Google Scholar] [CrossRef]
  21. Gray, R.; Kieffer, J. Mutual information rate, distortion, and quantization in metric spaces. IEEE Trans. Inf. Theory 1980, 26, 412–422. [Google Scholar] [CrossRef]
  22. Maes, C.; Redig, F.; van Moffaert, A. On the definition of entropy production, via examples. J. Math. Phys. 2000, 41, 1528–1554. [Google Scholar] [CrossRef]
  23. Zeng, Q.; Wang, J. Nonequilibrium Enhanced Classical M easurement and Estimation. J. Stat. Phys. 2022, 189, 10. [Google Scholar] [CrossRef]
  24. Yan, L.; Ge, X. Entropy-Based Energy Dissipation Analysis of Mobile Communication Systems. arXiv 2023, arXiv:2304.06988. [Google Scholar] [CrossRef]
  25. Tasnim, F.; Freitas, N.; Wolpert, D.H. The fundamental thermodynamic costs of communication. arXiv 2023, arXiv:2302.04320. [Google Scholar]
  26. Ball, F.; Yeo, G.F. Lumpability and Marginalisability for Continuous-Time Markov Chains. J. Appl. Probab. 1993, 30, 518–528. [Google Scholar] [CrossRef]
  27. Mandal, D.; Jarzynski, C. Work and information processing in a solvable model of Maxwell’s demon. Proc. Natl. Acad. Sci. USA 2012, 109, 11641–11645. [Google Scholar] [CrossRef]
  28. Barato, A.C.; Seifert, U. Unifying three perspectives on information processing in stochastic thermodynamics. Phys. Rev. Lett. 2014, 112, 219–233. [Google Scholar] [CrossRef]
  29. Gallager, R.G. Information Theory and Reliable Communication; Wiley: New York, NY, USA, 1968. [Google Scholar]
  30. Gaspard, P. Time-reversed dynamical entropy and irreversibility in Markovian random processes. J. Stat. Phys. 2004, 117, 599–615. [Google Scholar] [CrossRef]
  31. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley&Sons: Hoboken, NJ, USA, 2006; ISBN 13 978-0-471-24195-9. [Google Scholar]
  32. Verdu, S.; Han, T.S. A general formula for channel capacity. IEEE Trans. Inf. Theory 1994, 40, 1147–1157. [Google Scholar] [CrossRef]
  33. Barato, A.C.; Hartich, D.; Seifert, U. Rate of Mutual Information Between Coarse-Grained Non-Markovian Variables. J. Stat. Phys. 2013, 153, 460–478. [Google Scholar] [CrossRef]
  34. Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, MA, USA, 2004. [Google Scholar]
  35. Holliday, T.; Goldsmith, A.; Glynn, P. Capacity of finite state channels based on Lyapunov exponents of random matrices. IEEE Trans. Inf. Theory 2006, 52, 3509–3532. [Google Scholar] [CrossRef]
  36. Lan, G.; Sartori, P.; Neumann, S.; Sourjik, V.; Tu, Y. The energy–speed–accuracy trade-off in sensory adaptation. Nat. Phys. 2012, 8, 422–428. [Google Scholar] [CrossRef] [PubMed]
  37. Elowitz, M.B.; Levine, A.J.; Siggia, E.D.; Swain, P.S. Stochastic Gene Expression in a Single Cell. Science 2002, 297, 1183–1186. [Google Scholar] [CrossRef]
  38. Detwiler, P.B.; Ramanathan, S.; Sengupta, A.; Shraiman, B.I. Engineering aspects of enzymatic signal transduction: Photoreceptors in the retina. Biophys. J. 2000, 79, 2801–2817. [Google Scholar] [CrossRef] [PubMed]
  39. Tkacik, G.; Callan, C.G., Jr.; Bialek, W. Information flow and optimization in transcriptional control. Proc. Natl. Acad. Sci. USA 2008, 105, 265–270. [Google Scholar] [CrossRef] [PubMed]
  40. Tkacik, G.; Callan, C.G., Jr.; Bialek, W. Information capacity of genetic regulatory elements. Phys. Rev. E 2008, 78, 011910. [Google Scholar] [CrossRef] [PubMed]
  41. Ziv, E.; Nemenman, I.; Wiggins, C.H. Optimal signal processing in small stochastic biochemical networks. PLoS ONE 2007, 2, e1077. [Google Scholar] [CrossRef]
  42. Moor, A.L.; Zechner, C. Dynamic Information Transfer in Stochastic Biochemical Networks. arXiv 2022, arXiv:2208.04162v1. [Google Scholar] [CrossRef]
  43. Tostevin, F.; Ten Wolde, P.R. Mutual Information between Input and Output Trajectories of Biochemi cal Networks. Phys. Rev. Lett. 2009, 102, 218101. [Google Scholar] [CrossRef]
Figure 1. An illustration of the potential profiles U X in an information transfer. There are two “flat” wells in the potential: a higher well with height H > 0 and a lower well with height 0. The location of the lower well depends on the source message X = 0 or X = 1 . The left figure shows that the lower well is located at the left half of the area when X = 0 . The right figure shows that the lower well is located at the right half when X = 1 . When the receiver is found at the left half of the area, the received message is Y = 0 ; the right half corresponds to Y = 1 . The locations of the lower well represent the correctly received message.
Figure 1. An illustration of the potential profiles U X in an information transfer. There are two “flat” wells in the potential: a higher well with height H > 0 and a lower well with height 0. The location of the lower well depends on the source message X = 0 or X = 1 . The left figure shows that the lower well is located at the left half of the area when X = 0 . The right figure shows that the lower well is located at the right half when X = 1 . When the receiver is found at the left half of the area, the received message is Y = 0 ; the right half corresponds to Y = 1 . The locations of the lower well represent the correctly received message.
Entropy 25 01589 g001
Figure 2. An illustration of a sequential information transfer. The time sequence of the source messages is given by X i 1 = 1 , X i = 0 . The sequence of the received message is given by Y i , Y i + 1 . (a) The receiver receives correct messages as Y i = 1 , Y i + 1 = 0 , where all the received messages appear at the correct locations of the lower wells in the potential profiles. (b) The receiver obtains the wrong messages as Y i = 0 , Y i + 1 = 1 , where all the received messages appear at the locations of the higher wells in the potential profiles.
Figure 2. An illustration of a sequential information transfer. The time sequence of the source messages is given by X i 1 = 1 , X i = 0 . The sequence of the received message is given by Y i , Y i + 1 . (a) The receiver receives correct messages as Y i = 1 , Y i + 1 = 0 , where all the received messages appear at the correct locations of the lower wells in the potential profiles. (b) The receiver obtains the wrong messages as Y i = 0 , Y i + 1 = 1 , where all the received messages appear at the locations of the higher wells in the potential profiles.
Entropy 25 01589 g002
Figure 3. Equivalence of the entropy production and the recoverability. Black line, the time average of the stochastic entropy production. Blue line, the time average of the stochastic recoverability. Red line, the theoretical value. Both the entropy production and the recoverability converge to the theoretical value with time.
Figure 3. Equivalence of the entropy production and the recoverability. Black line, the time average of the stochastic entropy production. Blue line, the time average of the stochastic recoverability. Red line, the theoretical value. Both the entropy production and the recoverability converge to the theoretical value with time.
Entropy 25 01589 g003
Figure 4. The mutual information as a function of the nonequilibrium strengths ( d 1 , d 2 ) . The mutual information is calculated corresponding to the a priori probabilities given in Table 1.
Figure 4. The mutual information as a function of the nonequilibrium strengths ( d 1 , d 2 ) . The mutual information is calculated corresponding to the a priori probabilities given in Table 1.
Entropy 25 01589 g004
Figure 5. The recoverability as a function of the nonequilibrium strengths ( d 1 , d 2 ) . Each recoverability is calculated corresponding to the a priori probability shown in Table 1.
Figure 5. The recoverability as a function of the nonequilibrium strengths ( d 1 , d 2 ) . Each recoverability is calculated corresponding to the a priori probability shown in Table 1.
Entropy 25 01589 g005
Figure 6. The mutual information as a function of the recoverability at fixed d 1 . The mutual information and the recoverability are calculated corresponding to the a priori probabilities given in Table 1.
Figure 6. The mutual information as a function of the recoverability at fixed d 1 . The mutual information and the recoverability are calculated corresponding to the a priori probabilities given in Table 1.
Entropy 25 01589 g006
Figure 7. The recoverability as a function of the mutual information at fixed d 1 . The mutual information and the recoverability are calculated corresponding to the a priori probabilities given in Table 1.
Figure 7. The recoverability as a function of the mutual information at fixed d 1 . The mutual information and the recoverability are calculated corresponding to the a priori probabilities given in Table 1.
Entropy 25 01589 g007
Table 1. The a priori probabilities P X ( 1 ) = p used for numerical illustrations.
Table 1. The a priori probabilities P X ( 1 ) = p used for numerical illustrations.
(a)(b)(c)(d)
p 0.8212 0.0154 0.0430 0.1690
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, Q.; Li, R.; Wang, J. Nonequilibrium Effects on Information Recoverability of the Noisy Channels. Entropy 2023, 25, 1589. https://doi.org/10.3390/e25121589

AMA Style

Zeng Q, Li R, Wang J. Nonequilibrium Effects on Information Recoverability of the Noisy Channels. Entropy. 2023; 25(12):1589. https://doi.org/10.3390/e25121589

Chicago/Turabian Style

Zeng, Qian, Ran Li, and Jin Wang. 2023. "Nonequilibrium Effects on Information Recoverability of the Noisy Channels" Entropy 25, no. 12: 1589. https://doi.org/10.3390/e25121589

APA Style

Zeng, Q., Li, R., & Wang, J. (2023). Nonequilibrium Effects on Information Recoverability of the Noisy Channels. Entropy, 25(12), 1589. https://doi.org/10.3390/e25121589

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop