Literature DB >> 36241690

Fluctuation relations for irreversible emergence of information.

J Ricardo Arias-Gonzalez1.   

Abstract

Information theory and Thermodynamics have developed closer in the last years, with a growing application palette in which the formal equivalence between the Shannon and Gibbs entropies is exploited. The main barrier to connect both disciplines is the fact that information does not imply a dynamics, whereas thermodynamic systems unfold with time, often away from equilibrium. Here, we analyze chain-like systems comprising linear sequences of physical objects carrying symbolic meaning. We show that, after defining a reading direction, both reversible and irreversible informations emerge naturally from the principle of microscopic reversibility in the evolution of the chains driven by a protocol. We find fluctuation equalities that relate entropy, the relevant concept in communication, and energy, the thermodynamically significant quantity, examined along sequences whose content evolves under writing and revision protocols. Our results are applicable to nanoscale chains, where information transfer is subject to thermal noise, and extendable to virtually any communication system.
© 2022. The Author(s).

Entities:  

Year:  2022        PMID: 36241690      PMCID: PMC9568592          DOI: 10.1038/s41598-022-21729-9

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

Information management involves dynamic operations on a chain of physical objects carrying symbolic meaning. Such operations, technologically revolutionizing at atomic scales, comprise both physical manipulations of the objects, including mechanical maneuvering of or electromagnetic interactions with them, and purely configurational changes in the resulting sequence of symbols[1-4]. As a consequence, the chain-like system exchanges energy with the environment. Physical operations involve a thermodynamic analysis, whereas the computational aspects are often uniquely treated through information theory. However, information is also a physical entity amenable to be inspected through thermodynamic analysis[5]. The links between information theory and statistical mechanics are expanding the thermodynamic view of a physical system and its environment[6-8], with experimental verifications of the Landauer’s principle[9-11] and evidence for the conversion of information to work[12-14]. A paradigm of the link between information and thermodynamics is the so-called central dogma of molecular biology, which describes the information flow from DNA to proteins. These ancient, natural systems, besides involving polymer chains, are nanoscopic and dynamic, often evolving away from equilibrium[15-20], a feature that transcends to macroscopic scales in living beings[21]. Information variations in a chain-like system are associated to energy transactions with the environment, which can take place reversibly or irreversibly, with a lower theoretical energy limit[22,23]. Fluctuations as a consequence of pure computations are on the order of the thermal level (i.e., similar to kT, being k the Boltzmann constant and T the absolute temperature), according to Landauer’s principle. Such energies are negligible at routine human scales but become significant when the size of the system is nanoscopic or smaller, because the work and heat it generates also compare with the thermal level. Small systems are based on nanostructures, including individual molecules and arrangements of atoms, such as biological and quantum systems. Fluctuation theorems have appeared in recent years explaining quantitatively energy imbalances between forward and reverse pathways or between equilibrium and non-equilibrium processes[24,25]. They have been tested experimentally[26-28], mostly in biomolecular systems analyzed on a one-by-one basis[29]. Most of these theorems establish relations among thermodynamic potentials for general systems, often with no specific insight into information theory. This theory, in turn, deals with spatially-indexed, 1-dimensional arrangements of symbols, which may not be necessarily associated to a time order. Recent generalizations separate the role of information and feedback control[30,31], but still the interpretation of non-Markovianity, irreversibility and reversibility in terms of purely informational operations such as reading, writing and error correction[32,33] remains obscured. Here, we analyze energy exchanges associated to the symbolic management of a sequence of characters, without reference to the physical construction of the chain. Just by considering reversibility at the single sequence level and conservation laws, we next present two pairs of fluctuations equalities in the creation of information sequences, which use depends on energy exchange constraints. Our analysis integrates key information concepts, namely, reading, writing, proofreading and editing in the thermodynamic description of a string of symbols with information.

Results and discussion

The system under study is a chain made up of physical objects, represented by random variables (, being the alphabet of the symbols or domain of the discrete variables with cardinality , and the linear sequence). A microstate of the chain-like system builds up as a single microscopic sequence of values, , taken by the physical subunits of the chain, see Fig. 1.
Figure 1

Sketch of the chain-like system, which represents a segment of a tape with a finite number of cells. At each computation step, i, a machine working on the tape follows instructions from a program, transition table or polymer template, considers the symbols already incorporated (from 1 to ) and its ‘head’ changes the value of the current cell. The segment sequence, , is divided into an initial sequence and a remainder sequence . The remainder segment may be initially inexistent or with a null sequence (meaningless values for the cells from to n).

The chain has an associated reading direction, which is the one that makes the information along the chain meaningful (from left to right in Fig. 1), according to the language rules[33]. For example, a DNA template strand is copied (transcribed) from its 3’ end to its 5’ end during replication (transcription) and English language or sheet music are interpreted from left to right. The reading direction, which is characterized by order index i, imposes naturally a memory among (non-Markovian dependence on) the neighboring subunits, which restricts interactions to only one side of each subunit (left part of each current position i; i.e. neighbors j, such that , in Fig. 1). This is in contrast to general, one-dimensional chains, in which physical interactions take place on neighbors at both sides of each subunit. In a general information setup, sequence index i might not be necessarily related either to a temporal order followed to set the symbolic values in the chain or to the spatial arrangement of the symbols in the physical system. Sketch of the chain-like system, which represents a segment of a tape with a finite number of cells. At each computation step, i, a machine working on the tape follows instructions from a program, transition table or polymer template, considers the symbols already incorporated (from 1 to ) and its ‘head’ changes the value of the current cell. The segment sequence, , is divided into an initial sequence and a remainder sequence . The remainder segment may be initially inexistent or with a null sequence (meaningless values for the cells from to n). The construction of the chain is subjected to a protocol, , which stochastically drives its growth through privileged sequences. Information systems entail external mechanisms acting on the physical chain, which are specific of the device participating in the protocol. For example, DNA replication and transcription involves mechano-chemical forces performed by DNA/RNA polymerases (acting as the device head in Fig. 1) and artificial systems may use optical burning or magnetic setting of the characters on a tape. In the following, we will focus on the management of the sequence of symbols to find universal relations inherent to all information devices, i.e., without reference to the details of the device. Such details can be appropriately modeled for a complete thermodynamic analysis of information management in a specific system. A general information protocol, then, combines two main actions: writing, which entails the incorporation of characters along the reading direction, and revision, which comprises the proofreading + editing of the characters[33]. More in depth, writing involves testing characters within the alphabet at each chain subunit in turn, allowing one to equilibrate with respect to the previously incorporated neighbors, and then moving on, freezing that character permanently; and revision entails a global equilibration of the chain, during which characters can fluctuate non-sequentially (within the alphabet values) at each chain subunit. As studied earlier, writing is described by a directional, stochastic chain with memory, whereas revision entails a standard equilibrium description of the full chain[34,35]. In line with this analysis, we will use superindex for thermodynamic functions associated to the writing protocol and no superindex when referring to the proofreading and editing of the information. We next provide four fluctuations equalities for chain-like systems that evolve irreversibly from an intial information segment. When the energy of the chain is exactly specified, the entropy change between two revised growth states, , is related to the entropy change of the chains generated by writing, , as follows:where the expected value is taken over the ensemble of sequences produced under the writing protocol (). In this regard, the difference between the entropy change of a written chain, , and that of its revised version, , is related to the ratio of the writing (forward) and resetting (reverse) probabilities as follows: Using Jensen’s inequality (the exponential is a convex function) and , it follows from Eq. (1) thatwhich involves that when the chain can only exchange a defined energy with the environment, the information in the ensemble-average written chain is always greater than that in the revised chain. More in depth, by introducing the mutual information, and , in the way done earlier[36], this inequality imposes that . If the initial state were the same, independently of the protocol, then , which indicates that the entropy of the revised chain is larger than that of the ensemble-average written chain (lower fidelity in a copying process, for example). This consequence is reminiscent of the extended version of the Clausius theorem, as we will analyze further in this paper. It is important to note that the expected value of a writing thermodynamic function (, among others used below) must be interpreted as the ensemble-average value of that quantity over a sufficiently large number of printed chains in this context. In contrast, the expected value of a thermodynamic function generated after revision (S used so far) is equivalent to that of the chain printed once the proofreading and the editing have been applied. For the sake of clarity, we use the next analogy (Fig. 2): the writing expected value of a thermodynamic potential corresponds to the ensemble-average taken over the replicas of a text printed with a typewriting machine, whereas the revision expected value corresponds to the value of that quantity over the text generated by using a word processor on a computer, where one can proofread and edit the text as many times as desired before it is printed out.
Figure 2

Analogy for the writing and revision statistics. The image on the left corresponds to the generation of a sequence of characters according to the directional statistics (writing): following the reading direction (from left to right in, for example, English language and sheet music, or from the 3’ end to the 5’ end of a DNA template strand during its replication), the characters are incorporated one by one, without backtracking or deletions; although errors may be detectable are not editable in a typewriting machine. The image on the right corresponds to the standard statistics (revision): the information is again interpreted according to the reading direction but now it is possible to revise it—a process that includes error detection (proofreading) and correction (editing) in a computer with a word processor—before it is printed out.

Analogy for the writing and revision statistics. The image on the left corresponds to the generation of a sequence of characters according to the directional statistics (writing): following the reading direction (from left to right in, for example, English language and sheet music, or from the 3’ end to the 5’ end of a DNA template strand during its replication), the characters are incorporated one by one, without backtracking or deletions; although errors may be detectable are not editable in a typewriting machine. The image on the right corresponds to the standard statistics (revision): the information is again interpreted according to the reading direction but now it is possible to revise it—a process that includes error detection (proofreading) and correction (editing) in a computer with a word processor—before it is printed out. The above theorems arise from microcanonical assumptions (Supplementary Information). This description is too limited when the information chain can indeed exchange arbitrary energies with the environment. More importantly, allowing to break this energetic limitation enables revision as an effective process (i.e. capable of decreasing the entropy of the chains generated by writing), as we analyze next. When the chain-like system is in contact with a thermal reservoir at temperature T, we find from canonical assumptions:where is the energy change between the initial and final microstates of a chain with sequence and is the energy change between the corresponding revised initial and final states. The difference between the entropy change of a written chain and that of its revised version, given the difference between the energy change of that chain relative to its revised version, is related to the ratio of the writing and resetting probabilities at temperature T as follows: In line with Eq. (3), the corresponding inequality for information chains in contact with a thermal reservoir is obtained by applying Jensen’s inequality to Eq. (4), namely: , in other words,with . Using the mutual information quantity, we find , which allows the information of the revised chain to be higher than that of the written chain ( provided that the expected energy of the revised chain is lower than that of the written chain (, necessary but not sufficient condition). Inequality (6) is reminiscent of an earlier result demonstrated independently[35]: by setting equal initial states for the written and revised chains, it is found that , which was used to formulate necessary and sufficient conditions for effective proofreading and editing in information theory[33]. It is straightforward to see that when the system is only allowed to exchange a specific energy with the environment (, ), we recover Eqs. (1)–(3) as direct consequences of Eqs. (4)–(6). Theorems (1), (2) and (4), (5) are the main results of this paper, which integrate humanly-defined conceptual operations—namely, reading, writing and revision (as comprising proofreading and editing)—in information theory. In the following, we establish the notation and theoretical framework to analyze the irreversible construction of information chains, which are used to prove these results, to address thermodynamic meaning and to deduce consequences.

Thermodynamic analysis of information chains

We study an information chain, , as depicted in Fig. 1. Given two sequence lengths , the microstates and (see Fig. 1), with respective configurationsshare a common segment. The remainder sequence from m to n spells: The number of configurations for sequences , and are , and , respectively. The initial state of the chain-like system is conformed by sequences of length m, constructed with an undefined protocol. More in depth, the system is considered to start isolated and with a defined energy for the first pair of fluctuation relation and in an equilibrium state conformed by sequences of length m for the second pair of relations. An initial microstate of the chain, , is taken from such an ensemble, according to the surrounding conditions. Initial sequences are found, for example, as origins of DNA replication, gene promoters in transcription or the start codon in translation in molecular biology. Initial sequences may be absent in a general information system though. Then, the chain evolves irreversibly to a microstate of length n; the final sequence thus represents a microstate, , from an ensemble of written sequences. Alternatively, such microstate may belong to an ensemble of revised sequences (with a different probability). In the reversed evolution, the chain-like system starts at in equilibrium and ends at by attempting the recovery of the intial cell values of , either irreversibly or reversibly. The mathematical details of this analysis and proofs for the herein presented fluctuation equalities proceed in the Supplementary Information, where we consider thermodynamic functions (“A”) for the initial and final micro(states). More in depth, we evaluate potentials, on the one hand, at the single-sequence level (either or ) and, on the other hand, at the ensemble-average level (either computed with probability or computed with ), following the formalism derived elsewhere[34,35]. In this regard, the entropies for the writing and revision protocols, and , respectively, are defined by using probabilities , being the probability of introducing a specific character at each writing step when the previous incorporations are frozen permanently, and , which is the canonical probability distribution for a chain of symbols that fluctuate non-sequentially until equilibrium, with the energy of the chain. It is important to note that the writing protocol is supposed to be quasistatic: each character incorporation takes place in equilibrium conditions with respect to the previously incorporated characters, which remain fixed as the writing protocol proceeds. The fact that the writing protocol implies freezing each incorporated character introduces naturally the irreversibility of this process. Probabilities of writing, , and revision, , are equivalent only in the limit in which the memory effects are negligible (Independence Limit theorem)[34].

Thermodynamic meaning

To expand the physical interpretation of the above results, we study their coherence with more general results on stochastic thermodynamics. We next consider heat and work quantities for chain-like systems. The reversible, microscopic heat between two microstates of the chain with definite energy, (in equilibrium) and , in a monothermal evolution under protocol is[37,38]: The protocol-dependent ensemble-average heat becomes:and the equilibrium heat (Supplementary Information). We consider the heat positive when it is absorbed by the system and negative when it is released from the system. It follows from Eqs. (3) and (11) thatwhich is the extended version of the Clausius theorem[37] at constant temperature T, namely, ; note that Q is the heat exchanged in a reversible process and is the heat exchanged in a directional (irreversible) process. Inequality (12) sets a maximum to the the revised information in terms of the heat generated by the chain under (irreversible) writing, namely Like for the dissipated work, (see, for example,[39]), it is possible to define the dissipated heat as , which better illustrates this trade-off. is the heat that cannot be spent in changing the entropy of the chain, eventually released to the environment. This definition can be used to show compatibility with the Gallavotti–Cohen fluctuation theorem[40] in information chains (Supplementary Information). In a cycle, , since the reversible heat is zero. The fact that the dissipated heat in a closed loop is coherent with the fact that the information chain does not necessarily return to the initial microstate, , by resetting the remainder sequence, , to its exact initial values. The above analysis considers chain-like systems that exchange heat as a trade for information with the environment, which remains at constant temperature. This is in contrast to the heat exchange fluctuation theorem[41], which considers the irreversible evolution of two general systems put into contact from initially equilibrated, different temperatures. We now set the reversible, microscopic work between two microstates of the chain as the Helmholtz free energy change[37,38]: The protocol-dependent ensemble-average work yields: The equilibrium, ensemble-average (macroscopic) work, , emerges similarly (Supplementary Information). We consider the work positive when it is supplied to the system and negative when it is generated by the system. The expansion of the Helmholtz free energy in terms of the entropy and the energy of the chain for single sequences[35] along with the use of Eq. (14) makes evident the compatibility of Eqs. (4) and (5) with Jarzynski and Crooks theorems, respectively[39,42]. In addition, inequality (6), which can be expressed as by using Eq. (15), is coherent with Jarzynski equality after applying Jensen’s theorem, thus stating that the ensemble-average work to build information chains always supersedes the equilibrium free energy difference. In other words,which sets a maximum to the revised information in terms of the (irreversible) writing work supplied to the chain. Unlike the boundary Eq. (13), in which the nature of the system limits its capacity to release heat irreversibly, if the chain can exchange any energy with the thermal reservoir, the information can be made high just by increasing the work over the chain irreversibly. Equations (4) and (16) are in agreement with earlier fluctuation generalizations[30,43], by noting herein that I is the revised information in the chain and that we are just considering purely informational entropy.

Conclusions

We have derived fluctuation relations in the build-up of information chains. To do that, we have analyzed single sequences of characters by using the principle of microscopic reversiblity and conservation laws through a general (non-Markovian) formalism. Our results support the interpretation of thermodynamic irreversibility in terms of writing and equilibrium in terms of revision, as comprising the proofreading and editing of the information. We have shown that the connection of entropy changes, which involve structural order in thermodynamics but information in communication systems, to heat and work makes our results coherent with well-known non-equilibrium theorems. Irreversibility in an information chain appears due to the unlikelihood to grow and reset through coincident sequences. In other words, energy of a written sequence of characters (forward pathway) is dissipated as a consequence of the fact that the deletion of its associated information (backward pathway) does not necessarily return those characters to their initial symbolic values. Our results are directly exploitable in nanoscale information setups, where the thermal noise is represented by kT. In this regard, fidelity during replication, transcription or translation are analyzable with these theorems, where the homeostatic control on a DNA molecule, a messenger RNA or a protein in the cell presets temperature variations in a narrow interval. The application of our results to more general communication channels requires the description of noise by specific parameters[33], other than the thermal level. Supplementary Information.
  26 in total

1.  Entropy production fluctuation theorem and the nonequilibrium work relation for free energy differences.

Authors:  G E Crooks
Journal:  Phys Rev E Stat Phys Plasmas Fluids Relat Interdiscip Topics       Date:  1999-09

2.  Experimental verification of Landauer's principle linking information and thermodynamics.

Authors:  Antoine Bérut; Artak Arakelyan; Artyom Petrosyan; Sergio Ciliberto; Raoul Dillenschneider; Eric Lutz
Journal:  Nature       Date:  2012-03-07       Impact factor: 49.962

3.  Dynamical Ensembles in Nonequilibrium Statistical Mechanics.

Authors: 
Journal:  Phys Rev Lett       Date:  1995-04-03       Impact factor: 9.161

4.  In singulo biochemistry: when less is more.

Authors:  Carlos Bustamante
Journal:  Annu Rev Biochem       Date:  2008       Impact factor: 23.643

5.  Stochastic thermodynamics with information reservoirs.

Authors:  Andre C Barato; Udo Seifert
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2014-10-31

6.  Information management in DNA replication modeled by directional, stochastic chains with memory.

Authors:  J Ricardo Arias-Gonzalez
Journal:  J Chem Phys       Date:  2016-11-14       Impact factor: 3.488

7.  Fluctuation theorem with information exchange: role of correlations in stochastic thermodynamics.

Authors:  Takahiro Sagawa; Masahito Ueda
Journal:  Phys Rev Lett       Date:  2012-11-02       Impact factor: 9.161

Review 8.  Environmental sensing, information transfer, and cellular decision-making.

Authors:  Clive G Bowsher; Peter S Swain
Journal:  Curr Opin Biotechnol       Date:  2014-05-19       Impact factor: 9.740

Review 9.  Central dogma at the single-molecule level in living cells.

Authors:  Gene-Wei Li; X Sunney Xie
Journal:  Nature       Date:  2011-07-20       Impact factor: 49.962

Review 10.  Revisiting the central dogma one molecule at a time.

Authors:  Carlos Bustamante; Wei Cheng; Yara X Mejia; Yara X Meija
Journal:  Cell       Date:  2011-02-18       Impact factor: 41.582

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.