Previous Article in Journal
A Critical Candidate Node-Based Attack Model of Network Controllability
Previous Article in Special Issue
Entropy Production of Run-and-Tumble Particles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Non-Equilibrium Enhancement of Classical Information Transmission

by
Qian Zeng
1 and
Jin Wang
2,*
1
State Key Laboratory of Electroanalytical Chemistry, Changchun Institute of Applied Chemistry, Changchun 130022, China
2
Department of Chemistry and Physics, State University of New York, Stony Brook, NY 11794, USA
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(7), 581; https://doi.org/10.3390/e26070581
Submission received: 30 May 2024 / Revised: 1 July 2024 / Accepted: 3 July 2024 / Published: 8 July 2024
(This article belongs to the Collection Disorder and Biological Physics)

Abstract

:
Information transmission plays a crucial role across various fields, including physics, engineering, biology, and society. The efficiency of this transmission is quantified by mutual information and its associated information capacity. While studies in closed systems have yielded significant progress, understanding the impact of non-equilibrium effects on open systems remains a challenge. These effects, characterized by the exchange of energy, information, and materials with the external environment, can influence both mutual information and information capacity. Here, we delve into this challenge by exploring non-equilibrium effects using the memoryless channel model, a cornerstone of information channel coding theories and methodology development. Our findings reveal that mutual information exhibits a convex relationship with non-equilibriumness, quantified by the non-equilibrium strength in transmission probabilities. Notably, channel information capacity is enhanced by non-equilibrium effects. Furthermore, we demonstrate that non-equilibrium thermodynamic cost, characterized by the entropy production rate, can actually improve both mutual information and information channel capacity, leading to a boost in overall information transmission efficiency. Our numerical results support our conclusions.

1. Introduction

Information transmission is critical in a physical/biological world and in human society. Examples include communication [1,2,3], cryptography [4,5,6], information transmission in cells [7,8,9], information processing in the brain [10,11,12,13], and information propagation in social media and networks [14,15]. Recently, quantum information transmission has received much attention for the new generation of information communications and design for quantum computers [3]. The efficiency of information transmission is often measured by the mutual information between two physical systems. Information capacity is defined as the upper bound of mutual information. The study of information transmission has focused on various closed systems. However, most of the practical examples are open systems. Despite its huge successes in closed systems, there are still challenges for open systems on how non-equilibrium effects characterized by the exchange of energy, information, and materials from outside environments influence mutual information and information capacity and, therefore, information transmission efficiency [16].
The memoryless channel is the simplest information transmission model in information and communication theory [17]. It depicts how a noisy channel distorts the transmitted symbols with a transmission (conditional) probability distribution, which is permuted according to the transmitted symbol. This model is also important because it is the cornerstone of the information channel coding theories and associated methodology development that are concerned with how to transmit bit information more efficiently. It is noteworthy that physically noisy channels often lead to information dissipation or information loss of the transmitted messages.
A memoryless channel can be viewed as an ensemble of non-equilibrium open systems. This is because the information transmission can be through different noisy channels or the same channel with difference transmissions on the signal components. By quantifying the non-equilibrium strength [18,19] for information dynamics and non-equilibrium entropy production [20,21,22] or the dissipation cost for information thermodynamics [23,24,25,26] we investigate the non-equilibrium dynamics and the thermodynamics of this class of channels. The channel system is shown to be at equilibrium state if and only if the non-equilibrium strength, in terms of transmission probabilities and the corresponding entropy production rate, is zero. Otherwise, the system is in a non-equilibrium state.
Our investigation unveils a connection between the model’s mutual information, non-equilibrium strength, and entropy production. We reveal that mutual information exhibits a convex relationship with non-equilibriumness, quantified by non-equilibrium strength via probability flux. Notably, channel information capacity, which reflects the maximum achievable information transfer, is also enhanced by non-equilibrium effects. Furthermore, information cost, captured by the entropy production rate, surprisingly boosts both mutual information and information capacity, leading to a net gain in information transmission efficiency. This implies that non-equilibrium strength, mediated by probability flux, and the information-theoretic cost of non-equilibrium thermodynamics can both improve communication efficiency. However, at equilibrium, the absence of non-equilibrium effects results in vanishing mutual information and information capacity, signifying no effective information transmission. Our numerical simulations corroborate these findings, validating the model’s key features.

2. Non-Equilibrium Information Dynamics for Memoryless Channel

2.1. Information Dynamics for Memoryless Channel

Let us suppose that an information sender sends a message randomly to an information receiver within time t. The message is denoted by S = { s 1 , s 2 , . . . , s t } , where s i is the input symbol chosen from the code book S . Here, we assume that the input symbols s in one message are sent independently according to an identical distribution P ( s ) , which is the so-called input distribution. During the transmission, the message S is usually interfered with by random environments, it is received as a noisy message X = { x 1 , x 2 , . . . , x t } (where x i is the noisy output symbol), and it forms a noisy code book X , which must not be the same as the original code book S . The random environments are defined as the communication channel in the information theory [27,28]. The influence of the channel on the messages is characterized by the conditional distribution P ( X | S ) . The channel is said to be memoryless if it distorts each symbol independently of other symbols. Then, we can quantify the channel by using the time-invariant (conditional) transmission probability distribution q ( x | s ) for the output and input symbols x and s (see Figure 1). Thus, the conditional probability distribution P ( X | S ) for the messages can be rewritten into the products of the conditional transmission probabilities q ( x i | s i ) for the symbols at each time, P ( X | S ) = i = 1 t q ( x i | s i ) . The probability of the output symbol x can be given by the input distribution and the transmission distribution:
P ( x ) = s q ( x | s ) P ( s ) .
The P ( x ) is the so-called output distribution. In this model, we assume that both the sent and received messages S and X are stationary, independently and identically distributed processes with stationary distributions of symbols P ( s ) and P ( x ) , respectively.
From another point of view, information transmission can be viewed as a process for information processing (see Figure 2a). Here, the channel can be regarded as an information processor controlled by random environments. When the sender inputs a symbol s into the channel at time t, the channel processes the symbol randomly for a period or delay δ t = 1 and then outputs a new symbol x to the receiver at time t + 1 . With this perspective, we can establish Markovian information dynamics for the memoryless channel model. The symbols s t and x t can be viewed as the information state of the composite system of the sender and the channel, z t = ( s t , x t ) . The output symbol x t is decided by the channel within the information processing, according to the transmission probability q ( x t | s t 1 ) . The input symbol s t is independent of s t 1 and is generated by the sender according to the probability P ( s t ) . In addition, both x t and s t are independent of x t 1 if the channel is memoryless. Then, the transition probability of the state z t given z t 1 can be written as follows:
Q ( z t | z t 1 ) Q ( s t , x t | s t 1 ) = P ( s t ) q ( x t | s t 1 ) .
With this transition probability, the time evolution of the state distribution P ( z ) can be given by the following Markov master equation:
P ( z t ) = z t 1 Q ( z t | z t 1 ) P ( z t 1 ) .
Here, the transition probability Q in Equation (2) is recognized as the information driving force behind the composite system [18,19]. The driving force Q describes how the channel processes an input symbol s t 1 and generates an output x t ; Q also determines the evolution of the probability distribution P ( z ) in the dynamics. It can be verified that given an arbitrary initial distribution P ( z 1 ) the composite system finally goes to the unique stationary joint distribution π ( z ) = P ( s ) P ( x ) , which is independent of time, such that π ( z ) = z Q ( z | z ) π ( z ) .
Although the information driving force Q is associated with the input distribution P ( s ) , the efficiency of the information transfer is more related to the transmission probability of the channel, q ( x | s ) . Thus, we focus our discussions and conclusions mainly on how q ( x | s ) influences information transfer with an arbitrary given input P ( s ) .

2.2. Non-Equilibrium Strength

The environments of the channel are usually complex, which can give rise to non-equilibriumness in the channel, reflected in the transmission probabilities q ( x | s ) . This is equivalent to the measure of the broken time reversal symmetry or time-irreversibility in the composite dynamics. That is to say, the probability of the time sequence of the information states Z = { z 1 , z 2 , . . . , z t } , as shown in Equation (2), is usually not equal to that of the corresponding time reversal sequence Z ˜ = { z t , z t 1 , . . . , z 1 } . It is noteworthy that according to the information dynamics in Equation (3) each output symbol x i is determined by the input symbol s i 1 via the transmission probability q ( x i | s i 1 ) in the forward time sequence Z; and x i is decided by the input s i + 1 , according to the transmission probability q ( x i | s i + 1 ) in the backward time sequence Z ˜ (see Figure 2b). Then, the time-irreversibility of these information dynamics depends on the difference between the probability of the transition from s i 1 to x i in the forward time direction and that of the transition from s i + 1 to x i in the backward time direction. Furthermore, the output x i is independent of the previous output x i 1 in the time forward transition and the time successive output x i + 1 in the time backward transition; also, x i is independent of the simultaneous input s i , due to the information dynamics in Equation (3). Thus, the influences of the outputs x i 1 and x i + 1 , including the input s i , on the time-irreversibility can be negligible. The above observation yields the quantification of the time-irreversibility represented by the probability flux, which can be given by the difference between the probability of a time sequence and that of the time-reversal time sequence in the following equation:
J x i ( s i 1 , s i + 1 ) = x i 1 , x i + 1 , s i [ P ( Z ) P ( Z ˜ ) ] = 2 P ( s i 1 ) P ( s i + 1 ) d x i ( s i 1 , s i + 1 ) ,
with
d x i ( s i 1 , s i + 1 ) = 1 2 [ q ( x i | s i 1 ) q ( x i | s i + 1 ) ] .
Here, Z = { z i 1 , z i , z i + 1 } = { ( s i 1 , x i 1 ) , ( s i , x i ) , ( s i + 1 , x i + 1 ) } is a time sequence that contains the transition from s i 1 to x i in the forward time direction. The final state z i + 1 in the sequence Z is needed in the quantification of the time-irreversibility, because by using Z the transition form s i + 1 to x i in the backward time direction can be contained in the corresponding time-reversal time sequence Z ˜ = { ( s i + 1 , x i + 1 ) , ( s i , x i ) , ( s i 1 , x i 1 ) } . Then, the probabilities of Z and Z ˜ , which can be given by Equation (2), guarantee the dynamical connection between the transitions from s i 1 to x i and form s i + 1 to x i , which is decided by the information dynamics in Equation (3). Furthermore, the input symbols x i 1 and x i + 1 and the output symbol s i are summed away from the probabilities of Z and Z ˜ because they are not important to the time-irreversibility, as discussed in the above.
The probability flux J can be used to characterize the time-irreversibility of the composite information dynamics in Equation (3) because of term d in Equation (5), which reflects the non-equilibrium strength of the information dynamics (detailed discussion can be found in Appendix A). The dynamics reach equilibrium if and only if the probability flux J or the non-equilibrium strength d vanishes; otherwise, the dynamics are in non-equilibrium.

2.3. Non-Equilibrium Decomposition of Transmission Probabilities

From the perspective of non-equilibrium information dynamics, the non-equilibrium strength d in Equation (5) suggests a decomposition of the transmission probabilities q ( x | s ) , as follows:
q ( x | s ) = m x ( s , s ) + d x ( s , s ) , for s s .
where m and d can be given in the consistent forms in the force decomposition shown in Appendix A:
m x ( s , s ) = 1 2 [ q ( x | s ) + q ( x | s ) ] , d x ( s , s ) = 1 2 [ q ( x | s ) q ( x | s ) ] ,
Here, m is recognized as the equilibrium strength of the information dynamics, in contrast to the effect of d.
In Equation (6), s represents an input symbol which differs from the input symbol s. Due to this decomposition, both m and d can change independently of each other within the ranges of constraints that are properly given. Consequentially, q ( x | s ) can be changed by changing m and d. This can improve the information transfer as discussed in the above.
Due to the non-negativity and normalization of the transmission probabilities q ( x | s ) , the constraints can be given as follows:
0 m x ( s , s ) + d x ( s , s ) 1 0 m x ( s , s ) 1 x m x ( s , s ) + d x ( s , s ) = 1 ,
The first and second constraints are originated from the non-negativity of q ( x | s ) ; that is: 0 q ( x | s ) = m x ( s , s ) + d x ( s , s ) 1 and 0 m x ( s , s ) = 1 2 [ q ( x | s ) + q ( x | s ) ] 1 . The third constraint comes from the normalization of q ( x | s ) , i.e., 1 = x q ( x | s ) = x m x ( s , s ) + d x ( s , s ) . It is noteworthy that for an appropriate set of the decomposition that satisfies the constraints in Equation (8), all the possible non-equilibrium strengths d x ( s , s ) form a convex set by fixing the equilibrium strength 0 m x ( s , s ) 1 at each x (see Appendix B).
A strong connection exists between non-equilibrium strength and information transfer. Intuitively, when composite dynamics reach equilibrium (detailed discussion in Appendix A) the non-equilibrium strength d becomes zero. This signifies that the channel generates output symbols (x) independent of the sender’s input. In other words, the transmission probabilities q ( x | s ) become equivalent to the output distribution P ( x ) , regardless of the input distribution P ( s ) . This implies that under equilibrium, no matter the input distribution, the channel transmits no useful information from the sender. Conversely, a non-equilibrium state exists only when the non-equilibrium strength d is not zero. This can lead to a scenario where q ( x | s ) deviates from P ( x ) for any input distribution. Consequently, information transfer can only occur under non-equilibrium conditions.

3. Information Transmission Enhanced by Non-Equilibrium Strength

3.1. Mutual Information and Non-Equilibrium Strength

Information theory gives a fundamental quantification of the useful information transferred from the information sender to the receiver through the memoryless channel. This is the so-called information transfer rate or mutual information rate [29,30] per symbol for the two messages X = { x 1 , x 2 , . . . , x t } and S = { s 1 , s 2 , . . . , s t } :
I = lim t 1 t S , X P ( S , X ) log P ( S , X ) P ( S ) P ( X ) = z , z π ( z ) Q ( z | z ) log π ( z ) Q ( z | z ) π ( z ) π ( z ) = x , s P ( s ) q ( x | s ) log q ( x | s ) P ( x ) 0 .
In the second equality in Equation (9), z = ( x , s ) and z = ( x , s ) ; here, I is written in terms of the Markovian information dynamics in Equation (3), where the channel can be viewed as an information processing device and the stationary distribution π ( s , x ) = P ( s ) P ( x ) . In the third equality, I is in the ordinary form of the mutual information between the input and output symbols. It can be proved that these two expressions of the mutual information rate are equivalent to each other from the descriptions in Equations (2)–(4). Here, I > 0 means that although the channel is noisy the useful information from the sender can be received by the receiver. Otherwise, the vanishing I indicates that the useful information is totally corrupted through the channel. Here, the mutual information achieves its minimum ( I = 0 ) if and only if the input and output symbols are independent of each other, i.e., q ( x | s ) = P ( x ) holds for every x and s. This observation is consistent with the perspective of non-equilibrium information dynamics: ineffective information transfer ( I = 0 ) implies a detailed balance condition in the channel, where the non-equilibrium strength d = 0 in this situation. On the other hand, when in equilibrium the information cannot be transferred efficiently. In other words, the information from the sender can be received by the receiver if and only if the channel model is in non-equilibrium. This inspires us to investigate the relationship between non-equilibriumness and information transfer.
We catch the first glimpse of this relationship by noting that the mutual information I in Equation (9) can be rewritten as the function with respect to the non-equilibrium strength d with the decomposition in Equation (6):
I ( d ) = x , s P ( s ) q ( x | s ) log q ( x | s ) P ( x ) = x , s P ( s ) [ m x ( s , s ) + d x ( s , s ) ] log [ m x ( s , s ) + d x ( s , s ) ] s P ( s ) [ m x ( s , s ) + d x ( s , s ) ]
Here, P ( x ) = s P ( s ) q ( x | s ) = s P ( s ) [ m x ( s , s ) + d x ( s , s ) ] shown in Equation (1). According to Equation (A3), the model is in equilibrium if and only if every strength d = 0 , and this is also the necessary and sufficient condition of ineffective information transfer ( I = 0 ), no matter how the input P ( s ) is chosen. Thus, with the arbitrarily chosen P ( s ) , we can describe the impact of the channel-induced non-equilibriumness, characterized by the strength d, on the information transfer, as follows.
By noting the constraints on m and d in Equation (8), we can conclude that I is a convex function of the non-equilibrium strength d at each fixed equilibrium strength m and input P ( s ) (see Appendix C). The convexity of I ( d ) indicates that I has the unique minimum I = 0 at d = 0 , where the channel model is under equilibrium; in addition, I is a monotonously increasing function as the absolute value of | d x ( s , s ) | increases. This uncovers the intrinsic relationship between the information transfer and the non-equilibriumness: the degree of the non-equilibriumness d determined by the channel enhances the information transfer I. This conclusion can be independent of the input distribution P ( s ) .

3.2. Channel Capacity and Non-Equilibrium Strength

The channel coding theorem [26,27] reveals that the information transfer rate cannot exceed an upper bound, the channel capacity, which is exactly the maximum of the mutual information I, and merely depends on the channel. This yields a unique optimal input distribution P * ( s ) at which the capacity can be achieved. On the other hand, as with the relationship between the mutual information and non-equilibriumness revealed in Equation (7), we are also interested how the non-equilibriumness influences the capacity. Mathematically, the information capacity is formulated as the maximization of the mutual information I:
C = max P ( s ) I ; subject to s P ( s ) = 1 , and P ( s ) 0 .
Since the information capacity C is a concave function of P ( s ) then it has an unique maximum with respect to an unique optimal input P * ( s ) . Although this optimization problem, in general, does not have an analytical solution [31] we can investigate the relationship between C and the non-equilibrium strength d, which characterizes the channel-induced non-equilibriumness in the analytically solvable cases or we can perform a numerical test for more general cases. For example, the (weak) symmetric channel model is solvable in the capacity problem, and it is usually used as an explanatory case in the information theory. In the symmetric channel model, the transmission distribution q ( x | s ) at each input s is a permutation of that at the other input s. Thus, the information capacity can be achieved at the uniform distribution of s, P * ( s ) = 1 / N s ( N s is the number of the symbols s). The information capacity can then be given as follows:
C = log N x S ,
where N x is the number of the symbol x, and where S = x q ( x | s ) log q ( x | s ) is the entropy of the arbitrary transmission distribution. Due to the symmetry in the channel, all the entropies S are equal to each other. To see the relationship between the capacity and the non-equilibrium strength d in this model, we re-express C as the function of d:
C = log N x + x [ m x ( s , s ) + d x ( s , s ) ] log [ m x ( s , s ) + d x ( s , s ) ] ,
where we decompose q ( x | s ) by using Equation (6). Then, by fixing the equilibrium strength m in Equation (12) the information capacity C is a convex function of d. Meanwhile, the transmission probabilities are q ( x | s ) = P ( x ) = 1 N x , and there is no useful information that can be transferred by the channel. Otherwise, the capacity can be enhanced (monotonously increased) by the increasing absolute values of | d x ( s , s ) | in the symmetric model.

4. Better Information Transfer Efficiency under Larger Dissipation

4.1. Information Dissipation Enlarges Mutual Information

The larger strength of the non-equilibrium force d, which characterizes a non-equilibrium information dynamic (Equation (3)), can increase the efficiency or the rate of the transferred information I in the channel. On the other hand, a non-equilibrium information dynamic usually gives rise to information dissipation. Here, we are interested in how information dissipation influences information transfer.
In thermodynamics, the rate of thermodynamic dissipation is characterized by the entropy production rate (EPR). In this model, EPR can be given by the averaged log ratio of the probabilities of the time sequences and that of the time-reversal sequences and by taking the time average as follows, according to the fluctuation theorem [21]:
R Z = lim t 1 t Z P ( Z ) log P ( Z ) P ( Z ˜ ) ,
where Z = { z 1 , z 2 , . . . , z t } = { ( s 1 , x 1 ) , ( s 2 , x 2 ) , . . . , ( s t , x t ) } is a time sequence of the composite dynamics described by Equation (3) and Z ˜ ) is the time-reversal sequence of Z. Since the information dynamics are assumed to be Markovian, the expression of the EPR can be evaluated by using the probability of the time sequences shown in Equation (2):
R Z = 1 2 s i 1 , s i + 1 , x i J x i ( s i 1 , s i + 1 ) log Q ( s i , x i | s i 1 ) Q ( s i , x i | s i + 1 ) = s , s , x P ( s ) P ( s ) q ( x | s ) log q ( x | s ) q ( x | s ) 0 .
Here, the first equality in Equation (14) is written from the perspective of non-equilibrium Markov information dynamics, where the time-irreversible probability flux J x i ( s i 1 , s i + 1 ) is given by Equation (4) and the transition probability Q is given by Equation (2). From this perspective, the EPR is quantified by the averaged information difference or information loss between the transition in the backward time direction and the transition in the forward time direction, given by log Q ( s i , x i | s i + 1 ) and log Q ( s i , x i | s i 1 ) , respectively. The information loss l = log Q ( s i , x i | s i 1 ) Q ( s i , x i | s i + 1 ) can be positive if the forward transmission probability is larger than the backward probability, Q ( s i , x i | s i 1 ) > Q ( s i , x i | s i + 1 ) . This means that the forward transition happening is more likely than the backward transition in time with a positive information loss. The information loss l < 0 or, equivalently, Q ( s i , x i | s i 1 ) < Q ( s i , x i | s i + 1 ) indicates that the forward transition happening is less likely than the backward transition in time with a negative information loss. Otherwise, l = 0 if and only if q ( x i | s i 1 ) = q ( x i | s i + 1 ) ; then, the forward and backward transition happens with the same probability and with no information loss. Physically, the EPR can be regarded as the total entropy change consisting of the system and the environment [22]. Due to the Second thermal law, the total entropy change or the EPR cannot be negative, while some individual information losses l in the EPR can be negative. In other words, individual losses can be negative while the average losses must be non-negative. Mathematically, the non-negativity of the EPR is guaranteed by the form of the relative entropy D K L ( P ( Z ) | | P ( Z ˜ ) ) 0 in Equation (13). The non-negativity indicates that, driven by the non-vanishing probability flux J, the EPR is always positive in non-equilibrium conditions. Otherwise, the EPR vanishes at the equilibrium, and the probability flux J vanishes. Thus, the non-equilibriumness or time-irreversibility of the information dynamics can lead to positive information loss or thermodynamic dissipation quantified by the EPR.
The second equality in Equation (14) is written from the perspective of the information transfer. In this perspective, the information i of the input symbol can reduce the uncertainty of the output symbol; thus, the information i of an input symbol s carried by an output symbol x can be given by the reduction of the uncertainty of x, i s = log q ( x | s ) P ( x ) . Here, log P ( x ) quantifies the prior uncertainty of x without the information of s, and log q ( x | s ) quantifies the posterior uncertainty of x, which has been reduced by the information i s . Then, the entity l = log q ( x | s ) q ( x | s ) = log q ( x | s ) P ( x ) log q ( x | s ) P ( x ) = i s i s in Equation (14) quantifies the difference between the efficient information i s of an input s carried by the output x and the efficient information i s of another input s s contained in x. If the output x carries more information of s than that of s , then l = i s i s > 0 . On the contrary, l < 0 indicates that x carries more information of s than that of s. Either the positive or negative information difference l implies that if the output x carries more information of one input s than that of another input s then s can be recovered at the receiver more efficiently than s . This indicates that the positions of s and s in the (second equality) EPR are symmetric. Then, a negative information difference l is equivalent to a positive l by exchanging the positions of s and s . Thus, the averaged information difference or the EPR must be non-negative. The form of the relative entropy D K L ( P ( Z ) | | P ( Z ˜ ) ) 0 in Equation (13) also guarantees the non-negativity of the EPR. Otherwise, l = 0 refers to the worst situation, where the x carries equivalent information of both s and s. Thus, s and s cannot be distinguished by the receiver when the output x is observed. The worst situation can happen when the information dynamics are in equilibrium, i.e., the detailed balance condition in Equation (A3) is achieved. Furthermore, the information dissipation characterized by the EPR vanishes at equilibrium. Otherwise, the useful information of the input symbols can be transferred by the output symbols only in non-equilibrium conditions, which is characterized by the positive EPR in the thermodynamics. Thus, the EPR in Equation (14) uncovers the intrinsic connection between the non-equilibrium information thermodynamics and the information transfer.
To see the relationship between the EPR and the information transfer more clearly, we note that the information transfer rate in Equation (9) can be decomposed into a time-reversible part I M and a time-irreversible part I D with the decomposition in Equation (6), as follows:
I = I M + I D
with
I M = x , s P ( s ) m x ( s , s ) log [ m x ( s , s ) + d x ( s , s ) ] s P ( s ) [ m x ( s , s ) + d x ( s , s ) ] I D = x , s P ( s ) d x ( s , s ) log [ m x ( s , s ) + d x ( s , s ) ] s P ( s ) [ m x ( s , s ) + d x ( s , s ) ] .
Here, I M and I D correspond to the equilibrium strength m and the non-equilibrium strength d, respectively. The direct connection between the information transfer rate and the EPR can be uncovered via the time-irreversible mutual information rate I D by noting the identity [18,19]:
R Z = R X + R S + 2 I D .
Here, R X and R S are the EPR of the information sender and the receiver, respectively, and they can be negligible if both the sent messages and received messages are independently and identically distributed processes. Thus, the information dissipation in the information processing can be exactly quantified by the time-irreversible mutual information rate I D :
I D = 1 2 R Z .
where the factor 1 / 2 is due to the symmetry between the forward and backward time sequences in the expression of the EPR in Equation (13). By noting the relation in Equation (17), we obtain
I = I M + I D = I M + 1 2 R Z .
This equation implies that the efficiency of the information transfer can be enhanced by increasing the information dissipation. As shown in Equation (15), the time-reversible mutual information rate I M and the time-irreversible mutual information rate I D are in the form of the functions of the degree of the non-equilibriumness d. By fixing m and the input P ( s ) , I D is a convex function of d, and it monotonously increases as the absolute value | d x ( s , s ) | increases. The convexity of I D can be verified by the convexity of the information transfer rate I, where the proof has been given in Appendix C.
By noting the same nature of the mutual information rate I as the function of d in Equation (10), we conclude that an increase in the efficiency of the information transfer requires an increase in the information dissipation cost. Thus, thermodynamic cost can be used to boost information transfer efficiency.

4.2. Better Information Capacity under Larger Dissipation

As shown in Equation (12), information capacity can be enhanced by the non-equilibriumness of the symmetric model. Also, it can be seen that the EPR or the information dissipation at the optimal input P * ( s ) = 1 N s is a convex function of the degree non-equilibriumness of the model. By inserting P * ( s ) into Equation (14) or Equation (15), we have the following:
I D * ( d ) = x d x ( s , s ) log [ m x ( s , s ) + d x ( s , s ) ] .
By fixing the averaged equilibrium strength m, the convexity of I D * with respect to the averaged non-equilibrium strength d can be verified by noting the convexity of the information capacity clearly.
Through observation of the information capacity expression in Equation (12) and the EPR expression in Equation (14), one can obtain the relationship between the information capacity and the EPR:
C = I M * + I D * .
where I M * = log N x + x m x ( s , s ) log [ m x ( s , s ) + d x ( s , s ) ] is the time-reversible part in the information capacity. We can see that in the symmetrical channel the information capacity increases with respect to the increase of the thermodynamic information dissipation cost. Therefore, the thermodynamic cost can be used to increase the information capacity.

5. A Case with a Binary Memoryless Channel

To illustrate our concept, let us consider a simple model with a binary information measurement, which can be viewed as a fundamental form of information transmission. All the elements are explicitly defined and readily interpretable. In this model, the object, acting as the information sender, can exist in two states: s = 0 or s = 1 , which can be viewed as the input symbols. These states (symbols) influence the measurement device, acting as the information receiver, through a potential field within a confined area. This potential landscape has two “flat” wells: a higher one with height H > 0 and a lower one with height 0. The well positions depend on the object’s state. When the object is in state s = 0 , the lower well is on the left half and the higher well is on the right. Conversely, when the object is in state s = 1 , the well positions are flipped.
We use the device’s position (x) within the area to represent the measurement outcome or, for example, the output symbol. If the device settles in the left half ( x = 0 ), it suggests that the input symbol (state) s = 0 . Conversely, if it ends up in the right half ( x = 1 ), it suggests that the input symbol s = 1 . In essence, the object’s state is transmitted to the device through the potential field, and the device’s final position acts as the received message. The measurement device interprets the potential landscape and settles in the well corresponding to the object’s state. By settling in a specific location ( x = 0 or x = 1 ), the device receives and reports the information about the object’s state. Based on the above descriptions, the potential profile with respect to the input (s) and output (x) symbols can be given as follows (see Figure 3):
U ( x | s ) = 0 , x = 0 , s = 0 H 0 , x = 1 , s = 0 H 1 , x = 0 , s = 1 0 , x = 1 , s = 1
Here, since the potential profiles may not be symmetrical, we employ distinct notations, H 0 and H 1 , to represent the heights corresponding to different values of s. Based on the potential profile in Equation (21), finding the device in the lower well indicates a precise measurement or correct transmission ( x = s ). Conversely, if the device resides in the higher well, the measurement outcome or received symbol is imprecise ( x s ).
The device interacts with its environment at a constant temperature ( β = 1 ). This environmental noise allows the device to escape local potential minima (wells) and overcome potential barriers within the confined area. As a result, the device reaches equilibrium distributions within the area. These equilibrium distributions describe the probability of finding the device at each location. This is represented by the following equation:
p l 0 = 1 exp ( H 0 ) + 1 or p l 1 = 1 exp ( H 1 ) + 1 , probabilities of the device at the lower wells ; p h 0 = 1 p l 0 or p h 1 = 1 p l 1 , probabilities of the device at the higher wells .
Considering the described measurement model, we can interpret the equilibrium distributions (Equation (22)) in terms of measurement outcomes. The probabilities, p l 0 and p l 1 , directly reflect the probability of a precise measurement ( x = s ). Conversely, p h 0 and p h 1 represent the probabilities of an imprecise measurement ( x s ). With this understanding of the equilibrium distributions, we can now determine the transmission probabilities q ( x | s ) as follows:
q ( x | s ) = p l 0 , x = 0 , s = 0 p h 0 , x = 1 , s = 0 p l 1 , x = 1 , s = 1 p h 1 , x = 0 , s = 1
During sequential measurements, as depicted in Figure 2, the measurement device remains in a non-equilibrium state as the potential profiles switches based on the state being measured, as shown in Figure 4. In this way, the non-equilibrium strength or the time-irreversible probability flux according to Equation (7) can be used to represent the degree of non-equilibriumness during the measurement process:
d = 1 2 [ q ( x = 0 | s = 0 ) q ( x = 0 | s = 1 ) ] = 1 2 ( p l 0 p h 1 ) ,
and
m = 1 2 ( p l 0 + p h 1 ) ,
is recognized as the equilibrium strength.
If the states of the object are independently and identically distributed with distribution { P ( s = 0 ) = p , P ( s = 1 ) = 1 p } then the measurement device outputs symbols x at t + 1 according to distribution (Equation (1)):
P ( x ) = p l 0 p + p h 1 ( 1 p ) , x = 0 p h 0 p + p l 1 ( 1 p ) , x = 1 .
The decomposition of the transmission probabilities in this case can be given by Equation (6) as
p l 0 = m + d , and p h 1 = m d .
The constraints on m and d can be given by Equation (8) as
0 m + d 1 0 m d 1 0 m 1 ,
Then, d can be chosen from the convex set D = [ a , b ] with fixed m, where the lower bound a = max ( m 1 , m ) < 0 and the upper bound b = min ( 1 m , m ) . More explicitly, D = [ m , m ] for 0 < m 1 / 2 , and D = [ m 1 , 1 m ] for 1 / 2 < m < 1 . Here, for numerical illustration, we select four sets of equilibrium strengths m and input probabilities P ( s = 0 ) = p , which are shown in Table 1.
We show that more non-equilibrium or probability flux leads to more effective information transfer during the measurement. The mutual information between the input and output symbols s and x as the function of the characterization of the non-equilibrium d as the non-equilibrium strength can be derived from Equation (10):
I = S ( P ( x = 0 ) ) p S ( p l 0 ) ( 1 p ) S ( p h 1 ) = S ( m + d ( 2 p 1 ) ) p S ( m + d ) ( 1 p ) S ( m d )
where S ( p ) = p log p ( 1 p ) log ( 1 p ) is the entropy for a binary distribution. In the special case of a symmetric channel, where the potential profiles in Equation (21) have equal heights ( p h = p h 1 = p h 0 in Equation (22)), the information capacity simplifies to Equation (12):
C = 1 S ( p h ) = 1 S ( 1 / 2 d ) .
Here, the optimal input distribution can be given as P * ( 0 ) = P * ( 1 ) = 1 / 2 ; the non-equilibrium strength or the probability flux d = 1 / 2 p h and the equilibrium strength m = 1 / 2 in the symmetric model. We can see from the figure that both the mutual information I (Figure 5) and the information capacity C (Figure 6a) increase, while the absolute value | d | increases by fixing m and the input probability p. The minima of I and C are both achieved at the equilibrium point d = 0 , and the minima are both zero. This indicates that the transfer efficiency is zero at equilibrium. To see how the information dissipation influences the efficiency of the information transfer, we first evaluate the EPR in Equation (14) as
R = 2 p ( 1 p ) d [ ( log ( m + d ) log ( m d ) ) ( log ( 1 m d ) log ( 1 m + d ) ) ] .
By fixing m and p, R has a convex shape with a global minimum at d = 0 , while it monotonously increases when the absolute value | d | increases (see Figure 7; see also Figure 6b for the information capacity versus the EPR). Therefore, we conclude that in regard to the larger absolute value of the asymmetry of the noise probability | d | or the non-equilibrium strength we characterize the larger non-equilibrium results in both larger mutual information (see Figure 8) and larger information capacity (see Figure 6c) in this model, according to Equations (19) and (20), respectively. This leads to larger dissipation cost. In other words, higher mutual information and higher capacity require more dissipation cost to support and sustain.

6. Conclusions

This study investigated how non-equilibrium effects influence information transfer efficiency and capacity, quantified by mutual information and information capacity, respectively. We employed the memoryless channel model as a foundation for our exploration. Non-equilibriumness was characterized by the non-equilibrium information driving force, quantified by the probability flux that breaks detailed balance (a key principle in equilibrium systems). Our findings reveal that both mutual information and channel information capacity can be enhanced as the non-equilibrium strength, mediated by the probability flux, increases. This suggests that a stronger non-equilibrium force, from a dynamical perspective, improves both the efficiency and the capacity of information transfer. Additionally, our findings demonstrate that the entropy production rate, which characterizes the thermodynamic cost of dissipation, surprisingly boosts both mutual information and information channel capacity. To illustrate this, we considered a binary channel example. Here, we showed that a more asymmetric noise probability distribution leads to a larger non-equilibrium strength and higher entropy production, ultimately enhancing information transfer efficiency and capacity.
We plan to generalize our model to apply to different physical and biological systems, such as in [32].

Author Contributions

J.W. and Q.Z. conceived the original idea. Q.Z. performed the research and wrote the manuscript. J.W. revised the mathematical method and the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

Q.Z. thanks the support in part by National Natural Science Foundation of China (NSFC-91430217).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Probability Flux, Equilibrium and Non-Equilibrium Forces, and Time-Irreversibility

The probability flux J in Equation (4) can be used to characterize the time-irreversibility of the composite information dynamics in Equation (3) because of the information driving force Q in Equation (2). The force Q can be decomposed into a detailed balance-preserving part or equilibrium driving force Q m and a detailed balance-breaking part or non-equilibrium driving force Q d , as follows:
Q ( z i | z i 1 ) = Q m ( z i 1 , z i , z i + 1 ) + Q d ( z i 1 , z i , z i + 1 )
with
Q m ( z i 1 , z i , z i + 1 ) = P ( s i ) m x i ( s i 1 , s i + 1 ) , Q d ( z i 1 , z i , z i + 1 ) = P ( s i ) 2 P ( s i 1 ) P ( s i + 1 ) J x i ( s i 1 , s i + 1 ) = P ( s i ) d x i ( s i 1 , s i + 1 ) , m x i ( s i 1 , s i + 1 ) = 1 2 [ q ( x i | s i 1 ) + q ( x i | s i + 1 ) ] .
Here, the entities m x i ( s i 1 , s i + 1 ) and d x i ( s i 1 , s i + 1 ) (see Equation (5)) quantify the strengths of the equilibrium force Q m and the non-equilibrium force Q d , respectively. Furthermore, d can also be recognized as the strength of the probability flux J, as seen from Equation (4). The roles of the equilibrium driving force Q m and the non-equilibrium driving force Q d can be seen as follows. The probability of a time sequence Z given in Equation (2) can be written in terms of Q m and Q d , as follows:
P ( Z ) = π ( z 1 ) i = 2 t Q ( z i | z i 1 ) = P ( s 1 ) P ( x 1 ) Q ( z t | z t 1 ) i = 2 t 1 [ Q m ( z i 1 , z i , z i + 1 ) + Q d ( z i 1 , z i , z i + 1 ) ] = P ( s 1 ) P ( x 1 ) Q ( z t | z t 1 ) i = 2 t 1 P ( s i ) [ m x i ( s i 1 , s i + 1 ) + d x i ( s i 1 , s i + 1 ) ] .
Due to the description of the time-irreversibility, the composite dynamic is at equilibrium or is time-reversible if and only if P ( Z ) = P ( Z ˜ ) for all possible sequences Z. This corresponds to the non-equilibrium driving force Q d = 0 in Equation (A2), which gives the detailed balance condition
q ( x i | s i 1 ) = q ( x i | s i + 1 ) , for arbitrary x i , s i 1 , s i + 1 .
This detailed balance condition indicates that the output x i at each time is generated independently of the inputs s i 1 and s i + 1 in both forward and backward transitions in time. Namely, the output x i is generated according to the output distribution P ( x ) , which is independent of the output distribution P ( s ) . Then, the information driving force Q can be reduced into Q ( z i | z i 1 ) = Q m ( z i 1 , z i , z i + 1 ) = P ( s i ) P ( x i ) according to Equation (A1). This gives rise to
P ( Z ) = P ( s 1 ) P ( x 1 ) Q ( z t | z t 1 ) i = 2 t 1 Q m ( z i 1 , z i , z i + 1 ) = i = 1 t P ( s i ) P ( x i ) = P ( Z ˜ ) ,
which indicates the equilibrium state of the dynamics. Thus, the equilibrium force Q m preserves the detailed balance condition for the composite dynamics. On the other hand, driven by the non-vanishing non-equilibrium force Q d , the composite dynamic is away from the equilibrium, and it generates a non-vanishing probability flux J, appearing in Q d for all possible P ( Z ) P ( Z ˜ ) . With these discussions, we can see that the probability flux J is exactly the characterization of the non-equilibrium behavior of the composite information dynamics. Furthermore, together the strengths of the equilibrium and non-equilibrium forces m in Equation (A2) and d in Equation (5) can be regarded as the characterization of the non-equilibrium behavior of the information dynamics, which only depends on the channel. In particular, the strength d 0 characterizes the degree of the non-equilibriumness induced by the channel; meanwhile, d = 0 implies the detailed balance condition and the equilibrium state of the information dynamics.

Appendix B. Convexity of the Set of d

It is noteworthy that for an appropriate set of the decomposition that satisfies the constraints in Equation (8) all the possible strengths of the non-equilibrium driving force d x ( s , s ) form a convex set by fixing the strength of the equilibrium driving force 0 m x ( s , s ) 1 at each x. To see this, we first decompose the transmission probabilities q ( x | s ) at each output x and input s in the form of Equation (6) with a given set of the input symbols s s . Let D x be the set of all possible d that satisfy the constraints with m fixed. Then, arbitrarily choose two different d ( 1 ) , d ( 2 ) D x . Consequentially, both d ( 1 ) and d ( 2 ) satisfy the constraints in Equation (8) (all d D x satisfying Equation (8)), which can be given explicitly as follows:
0 m x ( s , s ) + d x ( 1 ) ( s , s ) = q ( 1 ) ( x | s ) 1 , and x m x ( s , s ) + d x ( 1 ) ( s , s ) = x q ( 1 ) ( x | s ) = 1 0 m x ( s , s ) + d x ( 2 ) ( s , s ) = q ( 2 ) ( x | s ) 1 , and x m x ( s , s ) + d x ( 2 ) ( s , s ) = x q ( 2 ) ( x | s ) = 1 . ,
Here, q ( 1 ) ( x | s ) and q ( 2 ) ( x | s ) are two transmission probabilities given by Equation (6). They can be different from each other at each x, s, and s because they are decided by the different strengths d ( 1 ) and d ( 2 ) , respectively. Let d ( λ ) = λ d ( 1 ) + ( 1 λ ) d ( 1 ) ( 0 λ 1 ) at each x, s, and s . We can check that d ( λ ) follows the constraints of D x in Equation (8), as follows:
m x ( s , s ) + d x ( λ ) ( s , s ) = λ q ( 1 ) ( x | s ) + ( 1 λ ) q ( 2 ) ( x | s ) = q ( λ ) ( x | s ) [ 0 , 1 ] , x m x ( s , s ) + d x ( λ ) ( s , s ) = x q ( λ ) ( x | s ) = λ x q ( 1 ) ( x | s ) + ( 1 λ ) x q ( 2 ) ( x | s ) = 1 .
Here, q ( λ ) ( x | s ) = m x ( s , s ) + d x ( λ ) ( s , s ) is the transmission probability decided by d ( λ ) . Thus, d ( λ ) can be a proper non-equilibrium strength that satisfies the constraints in Equation (8) for the arbitrary value of λ [ 0 , 1 ] . This also implies d ( λ ) D x . This result shows the convexity of D x by the definition of a convex set [33]. The convexity of D x is useful for the following discussion.

Appendix C. Convexity of Mutual Information and Channel Capacity

To prove the convexity of the mutual information, we first decompose q ( x | s ) in the form of Equation (6), and we fix m and P ( s ) . Second, all the possible strengths d form a convex set D x , which is decided by the fixed m and the constraints in Equation (8), as proved in Appendix B. Finally, we can prove that the mutual information I is convex on D . We arbitrarily choose two different strengths d ( 1 ) , d ( 2 ) from D x . Then, two different transmission probabilities can be determined by d ( 1 ) , d ( 2 ) , with the fixed m as q ( 1 ) ( x | s ) = m x ( s , s ) + d x ( 1 ) ( s , s ) and q ( 2 ) ( x | s ) = m x ( s , s ) + d x ( 2 ) ( s , s ) , respectively. Consequentially, two different output distributions can be determined by the transmission probabilities q ( 1 ) ( x | s ) and q ( 2 ) ( x | s ) as P ( 1 ) ( x ) = s P ( s ) q ( 1 ) ( x | s ) and P ( 2 ) ( x ) = s P ( s ) q ( 2 ) ( x | s ) with the fixed input distribution P ( s ) , respectively. By applying the log sum inequality [28] to Equation (10), we ascertain that
λ I ( d ( 1 ) ) + ( 1 λ ) I ( d ( 1 ) ) = λ x , s P ( s ) q ( 1 ) ( x | s ) log q ( 1 ) ( x | s ) P ( 1 ) ( x ) + ( 1 λ ) x , s P ( s ) q ( 2 ) ( x | s ) log q ( 2 ) ( x | s ) P ( 2 ) ( x ) x , s P ( s ) [ λ q ( 1 ) ( x | s ) + ( 1 λ ) q ( 2 ) ( x | s ) ] log λ q ( 1 ) ( x | s ) + ( 1 λ ) q ( 2 ) ( x | s ) λ P ( 1 ) ( x ) + ( 1 λ ) P ( 2 ) ( x ) = I ( d ( λ ) )
where λ [ 0 , 1 ] and d ( λ ) = λ d ( 1 ) + ( 1 λ ) d ( 2 ) D x . This completes the proof.
The channel capacity given in Equation (12) is a convex function because of the convexity of the negative entropy S in Equation (11). This can be seen from the Jensen’s inequality for the negative entropy, λ S ( d ( 1 ) ) ( 1 λ ) S ( d ( 2 ) ) S ( λ d ( 1 ) + ( 1 λ ) ( d ( 2 ) ) ) , where d ( 1 ) and d ( 2 ) are the two possible non-equilibrium strengths from the convex set D x . Here, the set D x is decided by the fixed equilibrium strength m, as described in Appendix B. Then, the information capacity C has a global minimum C = 0 at the equilibrium point of the model ( d = 0 ).

References

  1. Andersen, J.B. Antenna arrays in mobile communications: Gain, diversity, and channel capacity. Phys. Rev. Lett. 2000, 84, 12–16. [Google Scholar] [CrossRef]
  2. Bollt, E.; Lai, Y.C.; Grebogi, C. Coding, Channel Capacity, and Noise Resistance in Communicating with Chaos. Phys. Rev. Lett. 1997, 79, 3787–3790. [Google Scholar] [CrossRef]
  3. Lee, W.C.Y. Estimate of Channel Capacity in Rayleigh Fading Environment. Phys. Rev. Lett. 2002, 39, 187–189. [Google Scholar] [CrossRef]
  4. Prouff, E.; Rivain, M. Theoretical and practical aspects of mutual information-based side channel analysis. Phys. Rev. Lett. 2010, 104, 121–138. [Google Scholar] [CrossRef]
  5. Bennett, C.H.; Mor, T.; Smolin, J.A. Parity bit in quantum cryptography. Phys. Rev. A 1996, 54, 2675–2684. [Google Scholar] [CrossRef] [PubMed]
  6. Wolf, S.; Wultschleger, J. Zero-error information and applications in cryptography. In Proceedings of the Information Theory Workshop, San Antonio, TX, USA, 24–29 October 2004; IEEE: Piscataway, NJ, USA, 2004; pp. 1–6. [Google Scholar]
  7. Krishnaswamy, S.; Spitzer, M.H.; Mingueneau, M.; Bendall, S.C.; Litvin, O.; Stone, E.; Pe’er, D.; Nolan, G.P. Conditional density-based analysis of T cell signaling in single-cell data. Science 2014, 346, 1250689. [Google Scholar] [CrossRef] [PubMed]
  8. Voliotis, M.; Perrett, R.M.; McWilliams, C.; McArdle, C.A.; Bowsher, C.G. Information transfer by leaky, heterogeneous, protein kinase signaling systems. Proc. Natl. Acad. Sci. USA 2014, 111, E326–E333. [Google Scholar] [CrossRef] [PubMed]
  9. Andrews, B.W.; Iglesias, P.A. An information-theoretic characterization of the optimal gradient sensing response of cells. PLoS Comput. Biol. 2007, 3, e153. [Google Scholar] [CrossRef]
  10. Pregowska, J. Szczepanski, and E. Wajnryb, Mutual information against correlations in binary communication channels. BMC Neurosci. 2015, 16, 32. [Google Scholar] [CrossRef]
  11. Kim, B.; Boes, J.L.; Frey, K.A.; Meyer, C.R. Mutual Information for Automated Unwarping of Rat Brain Autoradiographs. Neuroimage 1997, 5, 31–40. [Google Scholar] [CrossRef]
  12. Chen, C.C.; Hsieh, J.C.; Wu, Y.Z.; Lee, P.L.; Chen, S.S.; Niddam, D.M.; Yeh, T.C.; Wu, Y.T. Mutual-Information-Based Approach for Neural Connectivity during Self-Paced Finger Lifting Task. Human. Brain Mapp. 2008, 29, 265–280. [Google Scholar] [CrossRef] [PubMed]
  13. Haueisen, J.; Leistritz, L.; Süsse, T.; Curio, G.; Witte, H. Identifying mutual information transfer in the brain with differential-algebraic modeling: Evidence for fast oscillatory coupling between cortical somatosensory areas 3b and 1. Neuroimage 2007, 37, 130–136. [Google Scholar] [CrossRef] [PubMed]
  14. Cai, K.; Spangler, S.; Chen, Y.; Zhang, L. Leveraging sentiment analysis for topic detection. Web Intell. Agent. Syst. 2008, 1, 291–302. [Google Scholar]
  15. De Choudhury, M.; Sundaram, H.; John, A.; Seligmann, D.D. What makes conversations interesting? Themes, participants and consequences of conversations in online social media. In Proceedings of the 18th International Conference on World Wide Web, WWW 2009, Madrid, Spain, 20–24 April 2009; ACM: New York, NY, USA, 2009. [Google Scholar]
  16. Hänggi, P. Stochastic Resonance in Biology: How Noise Can Enhance Detection of Weak Signals and Help Improve Biological Information Processing. ChemPhysChem 2002, 3, 285–290. [Google Scholar] [CrossRef] [PubMed]
  17. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  18. Zeng, Q.; Wang, J. Information Landscape and Flux, Mutual Information Rate Decomposition and Connections to Entropy Production. Entropy 2017, 19, 678. [Google Scholar] [CrossRef]
  19. Zeng, Q.; Wang, J. Non-Markovian nonequilibrium information dynamics. Phys. Rev. E 2018, 98, 032123. [Google Scholar] [CrossRef]
  20. Jarzynski, C. Nonequilibrium Equality for Free Energy Differences. Phys. Rev. Lett. 1997, 78, 2690. [Google Scholar] [CrossRef]
  21. Crooks, E.G. Entropy production fluctuation theorem and the nonequilibrium work relation for free energy differences. Phys. Rev. E 1999, 60, 2721–2726. [Google Scholar] [CrossRef] [PubMed]
  22. Seifert, U. Entropy Production along a Stochastic Trajectory and an Integral Fluctuation Theorem. Phys. Rev. Lett. 2005, 95, 040602. [Google Scholar] [CrossRef]
  23. Sagawa, T.; Ueda, M. Generalized Jarzynski Equality under Nonequilibrium Feedback Control. Phys. Rev. Lett. 2010, 104, 090602. [Google Scholar] [CrossRef] [PubMed]
  24. Parrondo, J.M.R.; Horowitz, J.M.; Sagawa, T. Thermodynamics of information. Nat. Phys. 2015, 11, 131–139. [Google Scholar] [CrossRef]
  25. Sagawa, T.; Ueda, M. Fluctuation theorem with information exchange: Role of correlations in stochastic thermodynamics. Phys. Rev. Lett. 2012, 109, 180602. [Google Scholar] [CrossRef]
  26. Horowitz, J.M.; Esposito, M. Thermodynamics with Continuous Information Flow. Phys. Rev. X 2014, 4, 031015. [Google Scholar] [CrossRef]
  27. Gallager, R.G. Information Theory and Reliable Communication; Wiley: Hoboken, NJ, USA, 1968. [Google Scholar]
  28. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  29. Holliday, T.; Goldsmith, A.; Glynn, P. Capacity of Finite State Channels Based on Lyapunov Exponents of Random Matrices. IEEE Trans. Inf. Theory 2006, 52, 3509. [Google Scholar] [CrossRef]
  30. Barato, C.; Hartich, D.; Seifert, U. Rate of Mutual Information between Coarse-Grained Non-Markovian Variables. J. Stat. Phys. 2013, 153, 460–478. [Google Scholar] [CrossRef]
  31. Verdu, S.; Han, T.S. A general formula for channel capacity. IEEE Trans. Inf. Theory 1994, 40, 1147–1157. [Google Scholar] [CrossRef]
  32. Kirchberg, H.; Nitzan, A. Energy Conversion and Entropy Production in Biased Random Walk Processes—From Discrete Modeling to the Continuous Limit. Entropy 2023, 25, 1218. [Google Scholar] [CrossRef]
  33. Boyd, S.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
Figure 1. Binary memoryless channel. The input symbol s from the information sender is sent to the receiver through the channel. The output symbol is x. The channel is characterized by the transmission probabilities q ( x | s ) .
Figure 1. Binary memoryless channel. The input symbol s from the information sender is sent to the receiver through the channel. The output symbol is x. The channel is characterized by the transmission probabilities q ( x | s ) .
Entropy 26 00581 g001
Figure 2. Information processing in the channel: (a) A processing time sequence within a time interval t is represented by Z = { ( s 1 , x 1 ) , ( s 2 , x 2 ) , . . . , ( s t , x t ) } . In the forward process (blue arrows), the channel processes the input symbols s, and every output symbol x i is determined by the previous symbols s i 1 in time, according to the probabilities q ( x i | s i 1 ) . In the backward process (red arrows), the output symbols x i are determined by the “future” symbols s i + 1 in time, according to the probabilities q ( x i | s i + 1 ) . (b) The minimum graph of the two dynamically correlated transitions s i 1 x i and s i + 1 x i in the forward and backward time directions, which characterize the time-irreversibility of the dynamics.
Figure 2. Information processing in the channel: (a) A processing time sequence within a time interval t is represented by Z = { ( s 1 , x 1 ) , ( s 2 , x 2 ) , . . . , ( s t , x t ) } . In the forward process (blue arrows), the channel processes the input symbols s, and every output symbol x i is determined by the previous symbols s i 1 in time, according to the probabilities q ( x i | s i 1 ) . In the backward process (red arrows), the output symbols x i are determined by the “future” symbols s i + 1 in time, according to the probabilities q ( x i | s i + 1 ) . (b) The minimum graph of the two dynamically correlated transitions s i 1 x i and s i + 1 x i in the forward and backward time directions, which characterize the time-irreversibility of the dynamics.
Entropy 26 00581 g002
Figure 3. The potential profiles of U ( x | s ) with two wells described by Equation (21).
Figure 3. The potential profiles of U ( x | s ) with two wells described by Equation (21).
Entropy 26 00581 g003
Figure 4. An illustration of the sequential information measurement or transmission. The time sequence of the input symbols is given by { s t 1 = 1 , s t = 0 } . The sequence of the output symbols is given by { x t , x t + 1 } : (a) the device measures or receives correct messages as { x t = 1 , x t + 1 = 0 } , where all the output symbols appear at the correct locations of the lower wells in the potential profiles. (b) The device obtains wrong messages as { x t = 0 , x t + 1 = 1 } , where all the received symbols appear at the locations of the higher wells in the potential profiles.
Figure 4. An illustration of the sequential information measurement or transmission. The time sequence of the input symbols is given by { s t 1 = 1 , s t = 0 } . The sequence of the output symbols is given by { x t , x t + 1 } : (a) the device measures or receives correct messages as { x t = 1 , x t + 1 = 0 } , where all the output symbols appear at the correct locations of the lower wells in the potential profiles. (b) The device obtains wrong messages as { x t = 0 , x t + 1 = 1 } , where all the received symbols appear at the locations of the higher wells in the potential profiles.
Entropy 26 00581 g004
Figure 5. The mutual information rate of the binary channel as the convex function of the non-equilibrium strength d, I ( d ) , with fixed equilibrium strength m and input probability P ( s = 0 ) = p in Table 1: (a) Set 1 gives m and p. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p. The minimum of all the functions I ( d ) is zero at the equilibrium point where d = 0 , and I ( d ) increases monotonously as the absolute value | d | increases.
Figure 5. The mutual information rate of the binary channel as the convex function of the non-equilibrium strength d, I ( d ) , with fixed equilibrium strength m and input probability P ( s = 0 ) = p in Table 1: (a) Set 1 gives m and p. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p. The minimum of all the functions I ( d ) is zero at the equilibrium point where d = 0 , and I ( d ) increases monotonously as the absolute value | d | increases.
Entropy 26 00581 g005
Figure 6. (a) The information capacity of the binary channel as the convex function of the non-equilibrium strength d, C ( d ) . (b) The entropy production rate of the binary channel when the information capacity is achieved; it is the convex function of the non-equilibrium strength d, R * ( d ) . (c) The information capacity as the function of the entropy production rate C ( R * ) at optimal input distribution.
Figure 6. (a) The information capacity of the binary channel as the convex function of the non-equilibrium strength d, C ( d ) . (b) The entropy production rate of the binary channel when the information capacity is achieved; it is the convex function of the non-equilibrium strength d, R * ( d ) . (c) The information capacity as the function of the entropy production rate C ( R * ) at optimal input distribution.
Entropy 26 00581 g006
Figure 7. The entropy production rate of the binary channel as the convex function of the non-equilibrium strength d, R ( d ) , with fixed equilibrium strength m and input probability P ( s = 0 ) = p in Table 1: (a) Set 1 gives m and p. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p. The minimum of all the functions R ( d ) is zero at the equilibrium point where d = 0 , and R ( d ) increases monotonously as the absolute value | d | increases.
Figure 7. The entropy production rate of the binary channel as the convex function of the non-equilibrium strength d, R ( d ) , with fixed equilibrium strength m and input probability P ( s = 0 ) = p in Table 1: (a) Set 1 gives m and p. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p. The minimum of all the functions R ( d ) is zero at the equilibrium point where d = 0 , and R ( d ) increases monotonously as the absolute value | d | increases.
Entropy 26 00581 g007
Figure 8. The mutual information as the function of the entropy production rate, I ( R ) : (a) The parameters m and p are given by Set 1 in Table 1. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p.
Figure 8. The mutual information as the function of the entropy production rate, I ( R ) : (a) The parameters m and p are given by Set 1 in Table 1. (b) Set 2 gives m and p. (c) Set 3 gives m and p. (d) Set 4 gives m and p.
Entropy 26 00581 g008
Table 1. Sets of fixed equilibrium strength m and input probability P ( s = 0 ) = p , used for numerical illustrations.
Table 1. Sets of fixed equilibrium strength m and input probability P ( s = 0 ) = p , used for numerical illustrations.
Set 1Set 2Set 3Set 4
m 0.4018 0.0760 0.2399 0.1233
p 0.8530 0.6221 0.3510 0.5132
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, Q.; Wang, J. Non-Equilibrium Enhancement of Classical Information Transmission. Entropy 2024, 26, 581. https://doi.org/10.3390/e26070581

AMA Style

Zeng Q, Wang J. Non-Equilibrium Enhancement of Classical Information Transmission. Entropy. 2024; 26(7):581. https://doi.org/10.3390/e26070581

Chicago/Turabian Style

Zeng, Qian, and Jin Wang. 2024. "Non-Equilibrium Enhancement of Classical Information Transmission" Entropy 26, no. 7: 581. https://doi.org/10.3390/e26070581

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop