Brought to you by:
Paper

Stochastic thermodynamics of bipartite systems: transfer entropy inequalities and a Maxwell's demon interpretation

, and

Published 27 February 2014 © 2014 IOP Publishing Ltd and SISSA Medialab srl
, , Citation D Hartich et al J. Stat. Mech. (2014) P02016 DOI 10.1088/1742-5468/2014/02/P02016

1742-5468/2014/2/P02016

Abstract

We consider the stationary state of a Markov process on a bipartite system from the perspective of stochastic thermodynamics. One subsystem is used to extract work from a heat bath while being affected by the second subsystem. We show that the latter allows for a transparent and thermodynamically consistent interpretation of a Maxwell's demon. Moreover, we obtain an integral fluctuation theorem involving the transfer entropy from one subsystem to the other. Comparing three different inequalities, we show that the entropy decrease of the first subsystem provides a tighter bound on the rate of extracted work than either the rate of transfer entropy from this subsystem to the demon or the heat dissipated through the dynamics of the demon. The latter two rates cannot be ordered by an inequality, as shown with the illustrative example of a four state system.

Export citation and abstract BibTeX RIS

1. Introduction

Thermodynamics of information processing started a long time ago with a thought experiment about 'violations' of the second law achieved by Maxwell's demon [1]. Among the seminal contributions to this field (see [2] for a collection of papers) are Szilard's engine [3], Landauer's principle [4] and Bennett's work [5].

More broadly, access to small systems where fluctuations are not negligible is now possible and an understanding of the relation between thermodynamics and information has become a problem of practical interest. For example, experimental verifications of the conversion of information into work [6] and of Landauer's principle [7] have been realized. Moreover, considerable theoretical progress has been made recently with the derivation of second law inequalities [8]–[11] and fluctuation relations [12]–[18] for feedback driven systems. The study of simple models has also played an important role [19]–[33]. In particular, Mandal and Jarzynski [34] (see also [35]–[38]) have introduced a model that clearly demonstrates an idea expressed by Bennett [5]: a tape can be used to do work as it randomizes itself.

In related work [39, 40], we have studied the relation between the rate of mutual information and the thermodynamic entropy production in bipartite systems. By a bipartite system we mean a Markov process with states that are determined by two variables such that in a transition between states only one of the variables can change. We have obtained an analytical upper bound on the rate of mutual information and developed a numerical method to estimate the Shannon entropy rates of continuous time series [40].

In this paper we take the view that a bipartite system provides a simple and convenient description of a Maxwell's demon. More precisely, considering a subsystem y as a Maxwell's demon, we show that work extraction by the other subsystem x leads to an entropy decrease of the external medium, with this entropy decrease being bounded by the entropy reduction of x due to its coupling with y. As an advantage of this Maxwell's demon realization, the full thermodynamic cost is easily accessible, being given by the standard entropy production of the bipartite system.

Moreover, we also study transfer entropy within our setup. Transfer entropy is an informational theoretical measure of how the dynamics of a process depends on another process [41], which is an important concept in the analysis of time series [42]. Ito and Sagawa [43] have recently obtained a fluctuation relation for very general dynamics involving the entropy variation of the external medium due to a subsystem and transfer entropy. Here, we obtain a similar fluctuation relation for a bipartite system. This fluctuation relation implies that the entropy decrease of the external medium due to subsystem x is bounded by the transfer entropy from x to y, where y can be interpreted as a Maxwell's demon.

Hence, we find three different bounds for the entropy decrease of the external medium due to x: the entropy reduction of x, the transfer entropy from x to y and the entropy increase of the external medium due to y. We show that the entropy reduction of x is always the best bound. Furthermore, by studying a particular four state model we observe that the transfer entropy can be larger than the entropy increase of the medium due to y.

The paper is organized as follows. In the next section, we define bipartite systems and the thermodynamic entropy production. Moreover, we explain in which sense a subsystem can be interpreted as a Maxwell's demon. We define the transfer entropy and obtain an analytical upper bound for it in section 3. In section 4 we prove an integral fluctuation relation involving the transfer entropy and show that the transfer entropy from x to y is larger than the entropy reduction of x due to y. Our results are illustrated with a simple four state system in section 5, where we summarize and compare the different inequalities obtained in this paper. We conclude in section 6.

2. Bipartite systems and thermodynamic entropy production

2.1. Basic definitions and inequalities

We restrict to a class of Markov processes which we call bipartite [39, 40]. The states are labeled by the pair of variables (α, i), where α ∈{1, ..., Ωx} and i ∈{1, ..., Ωy}. The transition rates from (α, i) to $(\beta ,j)$, are given by

Equation (1)

The central feature of the network of states is that when a jump occurs only one of the variables changes. Examples of a bipartite systems, inter alia, are stochastic models for cellular sensing [39, 44, 45], where one variable could represent the activity of a receptor and the other the concentration of some phosphorylated internal protein. We also denote a state of the system at time t by z(t) = (x(t), y(t)), where x(t) ∈{1, ..., Ωx} and y(t) ∈{1, ..., Ωy}. Hence, the subsystem x is related to the variable denoted by Greek letters and the subsystem y is related to the Roman letters.

The rate of entropy increase of the external medium [46] is then divided into two parts, one caused by jumps in the x variable and the other by jumps in the y variable. More precisely, we define

Equation (2)

and

Equation (3)

where $P_i^{\alpha }$ is the stationary probability distribution. The total entropy production, which fulfills the second law of thermodynamics, is then given by [46]

Equation (4)

The rate σx (σy) can be interpreted as the rate of increase of the entropy of the external medium due to the dynamics of the x (y) subsystem. It is important to notice that σx is not a coarse grained entropy rate [47]–[49]: knowing only the x time series is not sufficient to calculate σx.

The rate of change of the Shannon entropy of the system is known to be zero in the stationary state; this can be written as [46]

Equation (5)

Considering the term originating due to the x jumps we define

Equation (6)

We interpret this quantity as the rate at which the entropy of the subsystem x is reduced due to its coupling with y. This can be understood in the following way. If the state of the subsystem y is i, then the stationary probability of state α given i is $P(\alpha |i)= P_i^\alpha /\sum _\beta P_i^\beta $. Therefore, the rate of change of the Shannon entropy of the subsystem x for y = i is just $\sum _{\alpha }P_i^\alpha \sum _{\beta \neq \alpha } w^{\alpha \beta }_i\ln ({P(\alpha |i)}/{P(\beta |i)})$. Summing over all possible y we obtain (6). In this view it is as if the subsystem x's transition rates $w_i^{\alpha \beta }$ and probabilities P(α|i) depend on time due to the y jumps. In appendix A we consider a functional of the stochastic trajectory that when averaged gives hx. With this functional, this interpretation of hx becomes even more clear. Similarly, for the subsystem y we have

Equation (7)

From (5) it follows that

Equation (8)

i.e., the entropy reduction of subsystem x equals the entropy increase of subsystem y.

Besides the second law inequality for the full system (4), we also have for the total entropy production caused by transitions in the subsystem x

Equation (9)

This inequality has been considered explicitly in [50] and is a direct consequence of the log sum inequality. In appendix A we prove a more general integral fluctuation relation which implies (9). This fluctuation relation is similar to the fluctuation relation for the house-keeping entropy obtained by considering the dual dynamics [46, 51]. The same inequality is valid for the subsystem y,

Equation (10)

2.2. Subsystem y as a Maxwell's demon

Let us now consider a case where σx is negative, i.e., the entropy of the external medium decreases at rate − σx due to the subsystem x dynamics. From relations (8)–(10) we obtain the following inequalities:

Equation (11)

The second inequality can be interpreted in the following way. If we consider the subsystem y as a Maxwell's demon, then the rate of entropy reduction of the external medium − σx is bounded by the rate hx at which the entropy of the subsystem x is reduced due to its coupling to the Maxwell's demon. Furthermore, the first inequality contains an integrated description with the rate at which entropy increases in the Maxwell's demon − hy = hx being bounded by the rate of entropy increase in the external medium due to the y dynamics σy.

For a more specific interpretation involving the first law we assume that the transition rates take the following local detailed balance form:

Equation (12)

where $E_i^\alpha $ is the internal energy of the state (α, i) and $\omega _i^{\alpha \beta }$ is the work extracted from the system in the jump $(\alpha ,i)\to (\beta ,j)$. We set kBT = 1 throughout the paper. In the stationary state the rate of internal energy change due to x jumps is given by

Equation (13)

The rate of extracted work is

Equation (14)

From the relation σx = −epsilonxωout and the first law we identify σx as the dissipated heat due to the x jumps. Likewise σy is the dissipated heat due to y jumps. For the special case epsilonx = 0, we have − σx = ωout: the second inequality in (11) implies that the rate of extracted work is bounded by hx. The first inequality in (11) means that the rate of entropy decrease hx is bounded by the heat σy that is dissipated by the demon.

Let us make a comparison with the model introduced by Mandal and Jarzynski [34], where a tape composed of bits interacts with a system connected to a heat bath. By increasing the Shannon entropy of the tape the system can deliver work to a work reservoir. In the above interpretation, this delivered work corresponds to − σx and the tape is analogous to the subsystem y: the subsystem x can deliver work by increasing the entropy of the subsystem y. In this sense we can see the subsystem y as an information or entropy reservoir (see [37, 38] for definitions of an information reservoir).

Summarizing the above discussion, bipartite systems provide a particularly transparent description of Maxwell's demon, with the full thermodynamic cost being easily accessible through the standard second law inequality (4). We proceed by defining transfer entropy, which, as we will show in section 4, also provides a bound for − σx.

3. Shannon entropy rate and transfer entropy

We first consider a discrete time Markov chain with time spacing τ and transition probabilities corresponding to the transition rates (1), i.e.,

Equation (15)

We denote the full state of the system at time by zn = (xn, yn), where xn ∈{1, ..., Ωx} and yn ∈{1, ..., Ωy}. For the case where xn−1 = α, $x_{n}=\beta $, yn−1 = i and yn = j, we represent the transition probability $W_{ij}^{\alpha \beta }$ by W(xn, yn|xn−1, yn−1). Furthermore, the stochastic trajectory of the full process is written as $z_0^n= (z_0,z_1,\ldots ,z_n)$. Whereas $z_0^n$ is Markovian, the stochastic trajectories of the two coarse grained processes $x_0^n$ and $y_0^n$ are in general non-Markovian.

The Shannon entropy rate is a measure of how much the Shannon entropy of a stochastic trajectory increases as we increase the length of the trajectory n. For a generic process $a_0^n$ (where a = x, y, z) it is defined as

Equation (16)

where $P[a_0^n]$ is the probability of the trajectory $a_0^n$. In particular, since the full process is Markovian, its Shannon entropy rate is given by [52]

Equation (17)

The equality in the second line is convenient for the subsequent discussion where we will take the limit τ → 0. In general, a similar formula for the rates Hx and Hy in terms of the stationary distribution is not known and these Shannon entropy rates have to be calculated numerically [39, 40], [53]–[55].

A closely related quantity is the conditional Shannon entropy, which is defined as

Equation (18)

In the limit of n we have $H(a_n|a_0^{n-1})\to H_a$. Moreover, the conditional Shannon entropy decreases for increasing n: knowledge of a longer past decreases randomness [52]. Therefore, the conditional Shannon entropies $H(x_n|x_0^{n-1})$ and $H(y_n|y_0^{n-1})$, which can be calculated in terms of the stationary probability distribution, provide upper bounds on the Shannon entropy rates Hx and Hy, respectively. More precisely, it can be shown that for any finite n and up to order τ, the conditional Shannon entropies are given by [40]

Equation (19)

and

Equation (20)

where

Equation (21)

and

Equation (22)

with $P_i= \sum _{\beta }P_i^\beta $ and $P^\alpha = \sum _{j}P_j^\alpha $.

The transfer entropy from x to y is defined as [41]

Equation (23)

where $H(y_n|x_0^{n-1},y_0^{n-1})\equiv -{\tau }^{-1}\sum _{x_0^{n-1},y_0^n} P[x_0^{n-1},y_0^{n}]\ln P[y_n|x_0^{n-1},y_0^{n-1}]$. It is the reduction on the conditional Shannon entropy of the y process generated by knowing the x process. In other words, it measures the dependence of y on x or the flow of information from x to y. In the same way, the transfer entropy from y to x is written as

Equation (24)

The transfer entropy is in general not symmetric, i.e., $T^n_{x\to y }\neq T^n_{y\to x}$.

The conditional Shannon entropy $H(x_n|x_0^{n-1},y_0^{n-1})$ can be written as

Equation (25)

where the first equality comes from the fact that the full process is Markovian and in the second equality we have performed the substitutions xn−1α, yn−1i and $x_n\to \beta $. Analogously, we obtain

Equation (26)

In this paper we are interested in the transfer entropy in the continuous time limit τ → 0, which is defined as

Equation (27)

where we used relation (26) in the second equality. The conditional Shannon entropies diverge as lnτ for τ → 0 but the transfer entropy is well behaved in this limit. More clearly, from formula (20), the Shannon entropy rate $H_y= \lim _{n\to \infty } H(y_n|y_0^{n-1})$ diverges as $-\sum _{i,\alpha }P^\alpha _i\sum _{i\neq j} w_{ij}^{\alpha }\ln \tau $ in the limit τ → 0, canceling the lnτ term in (27).

Moreover, as the conditional Shannon entropy (20) for finite n bounds the Shannon entropy rate Hy from above, from equations (20) and (26) we obtain an analytical upper bound on $\mathcal {T}_{x\to y}$, given by

Equation (28)

This result is similar to the analytical upper bound on the rate of mutual information (see appendix B) we obtained in [39, 40]. Similarly, for the transfer entropy from y to x, which is defined as $T^n_{y\to x }\equiv H(x_n|x_0^{n-1})-H(x_n|x_0^{n-1},y_0^{n-1})$, we obtain

Equation (29)

Furthermore, if there is a clear timescale separation, i.e., if the x process is much faster than y, $w^{\alpha \beta }_i\gg w^{\alpha }_{ij}$, then $\mathcal {T}_{x\to y}\to \overline {\mathcal {T}}_{x\to y}$ [40]. Similarly, if the y process is much faster $\mathcal {T}_{y\to x}\to \overline {\mathcal {T}}_{y\to x}$.

While considering the discrete time case and taking the limit τ → 0 is convenient to calculate the upper bound (28), it is more efficient to consider continuous time trajectories with their waiting times to obtain the transfer entropy numerically. This issue and the relation between the transfer entropy and the rate of mutual information are discussed in appendix B.

4. Inequalities for the transfer entropy

4.1. Integral fluctuation relation

We consider a generic functional of the random variables $x^n_{n-1}$ and $y_0^{n}$, which is written as $\hat {F}[x_n,y_n;x_{n-1},y_0^{n-1}]$. The average of the functional is denoted by angular brackets, i.e.,

Equation (30)

where we used the Markov property $P[x_{n-1}^n,y_0^{n}]= W[x_n,y_n|x_{n-1},y_{n-1}]P[x_{n-1},y_0^{n-1}]$. Note that W denotes a transition probability where the time index n is irrelevant. In particular, we define the functionals

Equation (31)

and

Equation (32)

We can prove the following integral fluctuation relation:

Equation (33)

This relation does not depend on the transition probabilities having the form (15), therefore, it is valid also for systems that are not bipartite. Using Jensen's inequality we then obtain

Equation (34)

Finally, it is straightforward to show that

Equation (35)

Furthermore, as we show in appendix C,

Equation (36)

Therefore, inequality (34) implies

Equation (37)

A closely related fluctuation relation for causal networks has recently been obtained by Ito and Sagawa [43]. To obtain a fluctuation relation similar to (33) using the framework from [43], the stochastic trajectory of a bipartite system should be viewed as a causal network with two connected rows, corresponding to the x and y processes.

Comparing our autonomous system, where there are no explicit measurements or feedback, with standard feedback driven systems, the inequality (37) is analogous to the second law inequality for feedback driven systems as derived in [10]. As pointed out in [17], the quantity that bounds the extracted work in feedback driven systems is precisely the transfer entropy from the system to the controller performing the measurements, which is equivalent to $\mathcal {T}_{x\to y}$.

4.2. Comparison between hx and $\mathcal {T}_{x\to y}$

We would like to compare the bounds − σxhx and $-\sigma _x\le \mathcal {T}_{x\to y}$ in order to assess which one is stronger. By considering the average $\langle \hat {\sigma }_x+ \hat {T}_{x\to y}\rangle $, we obtain the following inequality:

Equation (38)

where we used the log sum inequality for the sum over $y_0^{n-2}$ and the definition

Equation (39)

By noting that $\langle \hat {h}_{x}\rangle /\tau = h_x+\textrm {O}(\tau )$, we obtain that the inequality (38) in the limit τ → 0 implies

Equation (40)

Therefore, hx provides a better bound on − σx than the transfer entropy $\mathcal {T}_{x\to y}$.

The above inequality can also be seen in a different way. In general, we do not know an analytical expression for the transfer entropy in terms of the stationary distribution. With the upper bound (28) and the inequality (40) we find

Equation (41)

Summarizing inequalities (11) and (41), we obtain

Equation (42)

Knowing now that hx is the best bound on − σx we would also like to investigate whether there is a unique relation between σy and the transfer entropy $\mathcal {T}_{x\to y}$. The following example will demonstrate that there is no such inequality.

5. Four state system

To illustrate our results we consider the simplest bipartite system which is a four state model. The transition rates are defined in figure 1. The parameters γx and γy set the timescales of the x and y transitions, respectively. We consider the case where qyqx ≤ 1/2 so that the probability current runs in the clockwise direction. In this case, σy and − σx are both positive.

Figure 1.

Figure 1. Four state system.

Standard image High-resolution image

We can interpret the model of figure 1 as follows. We consider two coupled proteins x and y that each can be in an inactive or active state, represented by 1 and 2, respectively. A chemical reaction, with chemical potential difference Δμx ≥ 0, drives the transitions of the x protein favoring the states (1,2) and (2,1), where the proteins are in different configurations. Local detailed balance is then written as

Equation (43)

Another chemical reaction drives the y transitions, also favoring anti-alignment of the proteins, implying in the local detailed balance relation

Equation (44)

The condition qyqx ≤ 1/2 reads ΔμyΔμx ≥ 0. In this case the chemical reaction driving the y transitions feeds work into the system at a rate σy and the system does work against the chemical reaction driving the x transitions at a rate − σx.

Explicitly, the stationary current is given by

Equation (45)

with the stationary probabilities $2P_1^1=2P^2_2=(\gamma _xq_x+\gamma _yq_y)/(\gamma _x+\gamma _y)$ and $2P_1^2=2P^1_2=1-2P_1^1$. Moreover, the rate of extracted work is given by

Equation (46)

and the rate of energy input is

Equation (47)

The rate of entropy reduction of subsystem x due to its coupling to y is

Equation (48)

The upper bound of the transfer entropy reads

Equation (49)

where $r_y\equiv 2P_1^2q_y+2P_1^1(1-q_y)$. An analytical expression for the transfer entropy $\mathcal {T}_{x\to y}$ is not known but we can determine it numerically.

Figure 2.

Figure 2. The rates of extracted work − σx and the bounds hx, σy, $\mathcal {T}_{x\to y}$, $\overline {\mathcal {T}}_{x\to y}$ as a function of qy for qx = 0.3, γx = 1, and γy = 5 (left panel), γy = 1 (right panel) for the network shown in figure 1. The transfer entropy $\mathcal {T}_{x\to y}$ is calculated using the numerical method from [40], as explained in appendix B.

Standard image High-resolution image

In figure 2 we compare the three different bounds on the extracted work − σx. Besides the illustration of inequalities (42), we can also see that $\mathcal {T}_{x\to y}$ approaches $\overline {\mathcal {T}}_{x\to y}$ as the y process becomes slower, as discussed in section 3. The main result we obtain from these plots is the crossing of the transfer entropy $\mathcal {T}_{x\to y}$ and σy, with the input σy being smaller near equilibrium, qx = qy, and larger in the far from equilibrium limit, qy → 0.

6. Conclusion

We have studied a series of second law like inequalities valid for bipartite systems. Besides the standard entropy production (4), the entropy production of a subsystem (9) and the inequalities involving transfer entropy (37) and (40) have been analyzed. Moreover, inspired by the fluctuation relation recently obtained by Ito and Sagawa [43] we have obtained the fluctuation relation (33), which in the continuous time limit leads to the inequality involving transfer entropy. From the summary of the inequalities (42) we have obtained that hx, the rate of entropy reduction of x due to the coupling with y, provides the best bound on − σx. As a particularly interesting interpretation, we have shown that a bipartite system provides a transparent realization of Maxwell's demon, with an integrated description of the subsystem and demon being easily accessible through the standard entropy production.

By analyzing a simple four state model we have shown that the transfer entropy $\mathcal {T}_{x\to y}$ can be larger than the entropy rate proportional to the heat dissipated by the demon σy. While the crossing between $\mathcal {T}_{x\to y}$ and σy has been obtained for a specific model we conjecture it to be more general because it depends on two general properties: $\mathcal {T}_{x\to y}$ not being zero in equilibrium, where σy = 0, and $\mathcal {T}_{x\to y}$ being finite when the y transition rate goes to zero, where σy diverges. Furthermore, as the transfer entropy is generally not zero in equilibrium it should be useless as a bound on − σx near equilibrium, e.g., in the linear response regime.

It is interesting to compare this work with [37], where a simplified version of the Mandal and Jarzynski model [35] for a tape interacting with a thermodynamic system was analyzed. In [37] the entropy (or information) reservoir is a tape composed of a sequence of bits, while here it is the y subsystem. Moreover, a series of inequalities similar to (42) have been obtained in [37], with the Shannon entropy difference of the tape providing the best bound on the extracted work, as is the case of hx here. Likewise, the mutual information between the tape and the system crosses the full input of work to reset the tape (see also [28]), corresponding to the crossing of $\mathcal {T}_{x\to y}$ and σy here.

Summarizing, the series of inequalities studied here, and the methods to calculate quantities like the rate of mutual information and transfer entropy that we have developed in [40], form a solid theoretical framework for bipartite systems, which constitute an important class of Markov processes. Among possible applications of our results, the investigation of bipartite models for cellular sensing is an interesting direction for future work.

Acknowledgment

Support by the ESF through the network EPSD is gratefully acknowledged.

Appendix A.: Fluctuation relation for the total entropy of the subsystem

The full Markovian stochastic trajectory from time 0 to T is denoted by $z(t)_0^T=(z_0,\tau _0;z_1,\tau _1;\ldots ;z_N,\tau _N)$, where the waiting times fulfill τ0 + τ1 + ⋯ + τN = T and N is the number of jumps in the trajectory. The probability density of a trajectory is written as

Equation (A.1)

where P(z0) is the initial distribution, wznzn+1 denotes the transition rate from zn to zn+1 defined in (1) and λzn ≡∑ zwznz is the escape rate.

In order to obtain the fluctuation relation leading to (6) we consider the modified transition rates

Equation (A.2)

and denote the path probability (A.1) obtained with these modified transition rates by $P^{\dagger }[z(t)_0^T]$, where the initial probability is also P(z0). The escape rates for u are written as ψzn ≡∑ zuznz. Furthermore, we define the functionals

Equation (A.3)

Equation (A.4)

where δyn, yn+1 is the Kronecker delta function. In the limit T we have 〈ΔHx〉/Thx and 〈ΔSx〉/Tσx, where the angular brackets here denote an integral over all stochastic paths (note that this is different from section 4).

The usual ratio of path probabilities is then given by

Equation (A.5)

where the functional $\Lambda _x[z(t)_0^T]\equiv \sum _{n=0}^{N}(\lambda _{z_n}-\psi _{z_n})\tau _n$ comes from the fact that the escape rates of w and u are different. Using standard methods [46], relation (A.5) implies

Equation (A.6)

From Jensen's inequality and

Equation (A.7)

we obtain the second law for the subsystem x (9).

Let us make the following remarks. One could consider an x transition from α to $\beta $ as dependent on time due to changes in the variable i in the transition rates $w_{i}^{\alpha \beta }$. Within this view, the rate u corresponds to a sort of 'adjoint' dynamics and relation (A.5) is similar to the ratio of probabilities involving the forward adjoint trajectory in the fluctuation relation for the house-keeping entropy derived in [51]. Furthermore, denoting by Ny the number of jumps for which the variable y changes and considering the interval between two y jumps [ny, ny + 1] we write

Equation (A.8)

where xi is the x state at the time of the jump ny, xf is the x state at the time of the jump ny + 1 and j is the y state in the time interval between the jumps ny and ny + 1. The functional (A.3) can then be written as

Equation (A.9)

In this form it becomes clear that the rate hx = 〈ΔHx〉/T in the large T limit is the rate of entropy reduction of subsystem x due to the subsystem y dynamics.

Appendix B.: Transfer entropy in continuous time

Using the notation of appendix A, the continuous time Shannon entropy rate is given by [56]

Equation (B.1)

If we compare this formula with (17) we see that the continuous time entropy rate does not show the lnτ divergence.

The coarse grained trajectories are written as $x(t)_0^T=(x_0,\tau _0^x;x_1,\tau _1^x;\ldots ;x_{N_x},\tau _{N_x}^x)$ and $y(t)_0^T=(y_0,\tau _0^y;y_1,\tau _1^y;\ldots ;y_{N_y},\tau _{N_y}^y)$, where $\tau _0^x+\tau _1^x+\cdots +\tau _{N_x}^x=\tau _0^y+\tau _1^y+ \cdots +\tau _{N_y}^y=T$ and Nx (Ny) is the number of jumps where the x (y) variable changes. Due to the bipartite nature of the transition rates N = Nx + Ny. The continuous time Shannon entropy rate of the y process is defined as

Equation (B.2)

In contrast to limτ→0Hy, this continuous time Shannon entropy rate does not have the divergent term proportional to lnτ. Hence, the transfer entropy (27) can be written as

Equation (B.3)

Therefore, to calculate the transfer entropy we just have to obtain the non-Markovian entropy rate $\mathcal {H}_y$; this can be achieved by using a numerical method to estimate the Shannon entropy for non-Markovian continuous time processes developed in [40].

We can apply the same procedure to the transfer entropy from y to x, which is written as

Equation (B.4)

Finally, we would like to point out the relation

Equation (B.5)

between the rate of mutual information $\mathcal {I}$ and the transfer entropy. The rate of mutual information measures how correlated the x and y processes are, without any specific direction. Obviously, the lower bound on the transfer entropy (40) can be extended to a lower bound on the rate of mutual information, i.e., $\mathcal {I}\ge |h_x|$.

Appendix C.: Proof of relation (36)

We start by rewriting (32) in the from

Equation (C.1)

We are interested in the limit n and for n large enough we can substitute $H(x_{n-1}|y_0^{n-1})$ by $H(x_{n}|y_0^{n})$, leading to

Equation (C.2)

Hence, from the definition (27), in order to prove (36) we have to show that

Equation (C.3)

Comparing this with (25), which reads

Equation (C.4)

it is then analogous to demonstrate that

Equation (C.5)

The term inside the logarithm can be written as

Equation (C.6)

In the following we consider three cases.

First, we consider ynyn−1, which implies xn = xn−1. Equation (C.6) takes the form

Equation (C.7)

where in the last equality we used $W[x_{n-1},\tilde {y}_{n}|\tilde {x}_{n-1},y_{n-1}]=1+\textrm {O}(\tau )$ for $\tilde {x}_{n-1}= x_{n-1}$ and $\tilde {y}_{n}= y_{n-1}$, and $W[x_{n-1},\tilde {y}_{n}|\tilde {x}_{n-1},y_{n-1}]=\textrm {O}(\tau )$ otherwise.

Second, we take xnxn−1, implying yn = yn−1. The term (C.6) is now written as

Equation (C.8)

where we used $\sum _{\tilde {x}_n}W[\tilde {x}_n,y_{n-1}|x_{n-1},y_{n-1}]=1+\textrm {O}(\tau )$.

Third, we consider xn = xn−1 and yn = yn−1. It is convenient to define L through the equality W[x, y|x', y'] = δx, x'δy, y' + τL[x, y|x', y'], where L is the stochastic matrix corresponding to the transition rates (1). The term (C.6) now becomes

Equation (C.9)

Finally, the quantity (C.5) can be separated into three contributions: one for ynyn−1, the second for xnxn−1 and the third for xn = xn−1 and yn = yn−1. Considering equations (C.7) and (C.8), we see that the first two contributions give O(τ2), leading to

Equation (C.10)

Using equation (C.9) we obtain

Equation (C.11)

where, from ∑ x, yL[x, y|x', y'] = 0 for all x', y', the term in the third line is zero and the terms in the second and fourth lines cancel. This concludes the proof of (C.5), which implies (36).

Please wait… references are loading.
10.1088/1742-5468/2014/02/P02016