Next Article in Journal
Adaptive Synchronization Strategy between Two Autonomous Dissipative Chaotic Systems Using Fractional-Order Mittag–Leffler Stability
Next Article in Special Issue
Exclusion and Underdetermined Qualia
Previous Article in Journal
Influence of Chronic Obstructive Pulmonary Disease and Moderate-To-Severe Sleep Apnoea in Overnight Cardiac Autonomic Modulation: Time, Frequency and Non-Linear Analyses
Previous Article in Special Issue
Measuring Integrated Information: Comparison of Candidate Measures in Theory and Simulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Integrated Information as a Measure of Cognitive Processes in Coupled Genetic Repressilators

1
Department of Mathematics, University College London, London WC1E 6BT, UK
2
The Alan Turing Institute, London NW1 2DB, UK
3
Institute for Women’s Health, University College London, London WC1E 6BT, UK
4
Department of Applied Mathematics and Laboratory of Systems Biology of Aging, Lobachevsky State University of Nizhniy Novgorod, 603022 Nizhniy Novgorod, Russia
5
Department of Pediatrics, Faculty of Pediatrics, Sechenov University, 119146 Moscow, Russia
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(4), 382; https://doi.org/10.3390/e21040382
Submission received: 20 November 2018 / Revised: 1 April 2019 / Accepted: 2 April 2019 / Published: 10 April 2019
(This article belongs to the Special Issue Integrated Information Theory)

Abstract

:
Intercellular communication and its coordination allow cells to exhibit multistability as a form of adaptation. This conveys information processing from intracellular signaling networks enabling self-organization between other cells, typically involving mechanisms associated with cognitive systems. How information is integrated in a functional manner and its relationship with the different cell fates is still unclear. In parallel, drawn originally from studies on neuroscience, integrated information proposes an approach to quantify the balance between integration and differentiation in the causal dynamics among the elements in any interacting system. In this work, such an approach is considered to study the dynamical complexity in a genetic network of repressilators coupled by quorum sensing. Several attractors under different conditions are identified and related to proposed measures of integrated information to have an insight into the collective interaction and functional differentiation in cells. This research particularly accounts for the open question about the coding and information transmission in genetic systems.

1. Introduction

Information-processing in living systems spans different degrees of biological complexity. From neural networks, where we find integration of information from different segregated modules in the nervous system, to the gene regulation and cell differentiation, whose structural mechanisms allow integration of signal transduction processes for survival and adaptation. Growing evidence shows that this degree of integration of stimuli is a required step to exhibit consciousness in the human brain where the information one receives is segregated and stored and later integrated and displayed.
These underlying mechanisms of information-processing in complex vertebrates can also be observed in simpler organisms such as unicellular organisms. As reviewed by [1], bacteria hold the ability for receiving, processing, and encoding information from the environment. This is done through inputs from different biochemical entities sensed from several signaling pathways. Later, this information is encoded for decision-making. This process can involve causal dependencies such as allocating previous information for calibration of responses or behavior, and even anticipation of stimuli. Thus, it is not surprise that several researchers have stressed the role of unicellular organisms as part of an evolutionary process for the development of the human brain [2,3,4]. Multiple microbiologists suggest that bacteria exhibit cognitive capacities that draw parallels in functionality as well as in the molecular mechanisms, ecology and evolution [1].
In cells, there is a functional integration of information channels coming from a vast repertoire of signaling pathways in which memory, learning, and decision-making take place under a changing environment. At the intracellular level, cells process information through gene regulatory networks, in which each gene influences the behavior of the other. The mechanisms found here serve to transduce the information sensed from the surroundings and the individual demands, including population density, individual cellular stress, stress induced from neighboring cells and the memory of previous individual and multicellular states [5,6,7]. At multicellular level, there is a recursive communication with the surrounding through signaling architectures, which include sensory mechanisms such as cell population density to brain-like electrical signaling. Such an interplay among the different levels of complexity allows processes such as cell differentiation, metabolism and DNA repair [8].
Still, an agreed definition of cognition and its boundaries across living systems remains elusive, and typically the term is weighted against the human case. To date, some authors describe cognition as a set of mechanisms such as information acquisition and coding from the environment [9] that allow biological entities act in a specific way to satisfy a viability constraint [10], or as a complex of sensory and information-processing derived by the structural and causal bidirectional interaction between organism and environment [1,5,11]. These interactions allow coding and processing of information to satisfy the requirements in living systems and provide the ability of acquiring, retaining, and using information from the environment to sensing, learning, and decision-making for adaptation of behavior and physiology for survival and growth. Despite these insights, the biological mechanisms at the cellular level remain unclear. Further investigation is needed to have a better understanding of its capabilities and identify its link with higher organisms.
The question that concerns this discussion is to what extent does the whole regulatory system in genetics have a balance between integration (coherence among the components of the system) and segregation (components dynamically independent) or to what extent is the whole system generating more information than the sum of its parts, allowing emergent behavior related to correlated spatial and temporal structures [12,13,14]. As seen, this integration provides the needed structure for more complex organisms [15,16,17,18,19].
The integrated information theory (IIT) provides the mathematical framework using information-theoretic principles in which it is attempted to measure the degree of interconnection of the parts of a system in terms of information exchange [20,21,22,23], where the information typically arises from causal interactions. Originally IIT is related to the conscious experience and its different mathematical descriptions have been adopted also as quantitative measure for complex dynamics [14,21,24,25,26,27].
To the best of our knowledge there is a lack of literature following this approach in genetic systems. Existing studies using information theory include the application of methods such as entropy and mutual information for network reconstruction [28,29,30]. These measures are also applied for research focused on the architecture and function of gene networks, where the network functionality is computed through the channel capacity and rate of signal transmission between regulatory signals and effector proteins [31]. In signaling systems this includes the determination of the amount of information transfer and reliability of signal transduction [32]. Some other applications include complex identification and resilience of network in protein-protein interactions [33] as well as reverse engineering and structure-dynamics identification in metabolic networks [34].
The proposed study based on IIT measures is of interest as it provides a practical application of sophisticated methods to quantify integration of information, which is fundamental for cell communication. Beyond neural systems, signal transduction holds mechanisms that integrate stimuli to achieve functionality in cells. Using computational modelling based on gene regulation we could account for mechanisms of intracellular and extracellular organization including multistability, stochastic gene expression, oscillations, pattern formation, among others. This is also of interest to understand emergent phenomena and optimization in the process of information exchange, and it could be also extended to design principles to more standard synthetic construction processes.
We address this question by studying an ensemble of existing synthetic repressilators. The repressilator network is an example of an engineered genetic system with predictive properties through modeling of its transcriptional regulation [35]. As a single cell, it is inferred that the protein concentrations oscillate over time. This has been observed experimentally with spontaneous oscillations with a period less than the division cycle. However, in realistic living systems cells do not act in isolation, but through cooperative interactions using signaling networks [36]. Thus, the intercellular communication enables the system to exhibit a rich repertoire of dynamical solutions reflecting phenotypic states, which may account for adaptation and differentiation [5,37]. In previous research, to understand cell-cell interaction the repressilator was coupled with other genetic oscillators through signaling molecules (acylated homoserine lactone) largely dependent on the population density. It was found that increasing the coupling among cells with different periods allowed them to reach phase-locking synchronization. The same system could be rewired using a different signaling pathway [38] in such a way that coupling exhibits for increasing coupling phase-repulsive ending with oscillation death. At the same time, multistable regimes appear with the possibility of different cluster distributions and emergence of chaos [38,39,40,41]. All these dynamical regimes could be interpreted as diversification of phenotypic states. This implies the coexistence of different communities in which the oscillators exhibit similar behavior, whose grouping ability is influenced by the degree of chaos.
In this paper, we compute integrated information (II) using the repressilator circuit coupled via quorum sensing. We use three different measures of II, namely, decoder-based integrated information [13], stochastic interaction [12,14] and whole-minus-sum integrated information [42,43]. This is particularly insightful as there are many candidate measures and still it is missing a consensus about their applicability and relationship in dynamical systems. Thus, we show the different levels of complexity that appear in terms of the signaling architecture and cell coupling. First, in the phase-attractive coupling scheme we calculate II in the transition to synchronization as a function of the cell density in a noisy population of genetic oscillators. It is found that the considered measures of II can detect order-disorder transitions exhibiting maximal susceptibility at self-organized criticality. In the case of phase-repulsive coupling, II is calculated for cells that exhibit different oscillatory clusters that change its distribution over time in terms of the degree of chaos. We explore the properties of standard measures of II and compare their behavior across these dynamical regimes. Our results give insights to understanding their applicability and to which extent each one reflects some degree of dynamical complexity.

1.1. Coupled Repressilators

Introduced from the seminal work [44], the repressilator is an artificial genetic network formed of three genes interacting through a negative feedback loop. This is, the activity of each gene is repressed in a cyclic manner: gene tetR encodes the TetR protein, which in turn, represses the transcription of protein CI produced from gene cl. Finally, CI binds to laCI to inhibit gene LacI. As mentioned previously, as a single cell, the repressilator network can produce oscillations, but here we are interested in including the coupling to understand the dynamics of multicellular systems. This can be modelled using a quorum sensing module using bacterium vibrio fischeri composed of two proteins LuxI and LuxR as suggested initially in [17,35]. Protein LuxI synthesizes an autoinducer (AI) that encodes the coupling among the neighboring cells. AI molecules diffuse throughout the extracellular space, and forms the complex LuxR-AI that binds to lacI, which in turn inhibits the expression of gene tetR. Under this general mechanism, different architectures are possible in terms of the gene from the repressilator circuit that regulates the concentration of LuxI. This enables the multicellular system to exhibit different dynamics that would not be observed in an isolated cell. In this work we concentrate on two coupling schemes, in particular, the phase-attractive and phase-repulsive coupling. Both descriptions are illustrated in Figure 1.

1.2. Phase-Attractive Coupled Repressilators

Consider a multicellular system in which each cell contains a repressilator network. Following the description from above, the repressilator is coupled through a quorum sensing mechanism in which AI molecules diffuse from the cellular membrane to other cells inducing the activity of a second plasmid lacI. In this way, the mRNA dynamics can be described by the following set of equations
d a i d t = a i + α 1 + C i n d b i d t = b i + α 1 + A i n d c i d t = c i + α 1 + B i n + κ S i 1 + S i
where a i , b i and c i denote the mRNA concentrations in cell i = 1 , . . . , N for genes tetR, cI, and laCI, respectively. Each mRNA transcripts proteins with concentration levels denoted by A i , B i and C i . The concentration of the AI in cell i is denoted by S i . Here mRNA and protein concentrations have been rescaled by their degradation rate (assumed equal for the three genes). The cooperativity number is denoted by the Hill coefficient n, and the maximum transcription rate is denoted by α . Finally, κ describes the maximum transcription amount of lacI by saturated concentration of AI.
The protein concentration in cell i evolves according to
d A i d t = β a ( a i A i ) d B i d t = β b ( b i B i ) d C i d t = β c ( c i C i )
where β a , b , c denotes the ratio between mRNA and protein lifetimes for each respective gene, tetR, cl, and lacl, respectively.
Observe that the architecture in the multicellular system is provided by the coupling scheme. To get phase synchronization we assume that the production of LuxR is governed by the expression of lacl. Thus, there is a reinforcement in the concentration of LacI as we increase the production of AI molecules.
Using this condition, the evolution for the AI is expressed by
d S i d t = k s 0 S i + k s 1 A i η ( S i S e )
here we assume a first order degradation, synthesis and intercellular diffusion encoded by the parameters, k s 0 , k s 1 and η , respectively. The diffusion rate of AI through the cell membrane η can be expressed as η = δ / V c , where V c is the cell volume and δ encodes the membrane permeability and surface area. The amount S e represents the extracellular concentration of AI, which in the quasi-state regime the mean field approximation can be written as S e = Q S ¯ , where S ¯ = j = 1 N S j / N denotes the average AI concentration over all cells in the system. The coupling Q is linearly dependent of the cell density δ N / V e x t in the system through the expression Q = δ N / V e x t / ( k s e + δ N / V e x t ) , where N is the number of cells, V e x t the extracellular volume and k s e the AI degradation rate.
Figure 2a–c shows the transition to the phase-locking state as we increase the cell density through the coupling parameter Q.

1.3. Phase-Repulsive Coupled Repressilators

If we rewire the quorum sensing pathway by setting the gene luxI repressed by the activity of protein TetR we obtain phase-repulsive coupling. This is a consequence of the competition between the induced negative feedback created by tetR and lacI and the repressilator circuit. In addition, the repression of CI favors the production of luxI. Then, by including these observations (3) can be modified as follows
d S i d t = k s 0 S i + k s 1 B i η ( S i S e )
In systems where N 2 it has been identified several attractors that include phase-shifted oscillations, clustering and chaos (Figure 2d–f, respectively) for increasing cell density. Thus, by defining the architecture for cell signaling there are implications in the emergence of cooperative behavior with different phase relations. If we relate this behavior with actual cellular states, understanding the different solutions is of crucial importance to get an insight into cell differentiation between different fates, its variability and robustness. In this work, we follow the discussion provided by [39,40] by tuning the cell density through the coupling Q. In particular, we focus on a dynamical regime that encompasses clustering, i.e., synchronized oscillation in subgroups, followed by chaos between two torus bifurcations.

1.4. Information-Theoretic Measures

We explore the spatio-temporal interactions of genetic oscillators that communicate through quorum sensing. The concentrations of mRNAs and proteins oscillate with a specified period and amplitude exhibiting different configurations in terms of the signaling architecture and biological parameters. In this work we choose the protein B to encode the dynamic state of each cell (due to the model symmetry this has minor effects in the results). In turn, oscillations from individual cells are binarized using as a threshold the mean of its instantaneous amplitude from the Hilbert transform. In other words, the absolute value of the analytical signal L H [ B ] defined as
L H [ B ] = B ( t ) + i H B ( t ) = L B ( t ) e i θ ( t )
where
H B ( t ) = 1 π P . V . B ( τ ) t τ d τ
is the Hilbert transform (P.V. denotes the Cauchy principal value). Thus, we can write
X ( t ) = 1 , B ( t ) L B ( t ) 0 , otherwise
where X ( t ) denotes the binarized state of B ( t ) .
This is useful to the general picture of gene activation in regulatory networks, in which every gene switches between a high or low state in terms of the interplay of activation and inhibition of the chemical signals of the other genes in the network. Therefore, this procedure naturally can be applied to other kind of gene dynamics in which a transition between states is considered. This approach for binarization has been followed for recent work in neuroscience [45]. In this paper, we integrate numerically the set of equations that defines the time evolution of the system and then by binarizing each channel from the solution we obtain a dataset where at each time step the state consists of a bit array of length equal to the number of channels.
Formally, let X to be a multidimensional stationary time series composed of binary information channels X i , where i = 1 , . . . , N with M observations (time steps) each. Then each state x is a bit array of N bits. Consider Y, the lagged signal of process by τ units of time. The current state of X is denoted by X t and Y by Y t = X t + τ . The time-delayed mutual information (TDMI) is defined as
I ( X Y ) = H ( X ) + H ( Y ) H ( X Y )
where X Y denotes a joint event, expressed as the concatenation of two-bit arrays X and Y. The Shannon entropy H ( X ) = x p ( x ) log p ( x ) quantifies the average number of bits needed to encode a state of the system [46]. Observe that H is time-independent because of the stationarity assumption. TDMI provides a measure about causal correlations of a many-body system. This is given in terms of the reduction of uncertainty of state at t + τ given the evolution up to t steps, i.e., it provides the uncertainty reduction about the future state by knowing the present states (or vice versa). In this way, TDMI quantifies the predictability in the evolution of the entire system. This complexity alone has proven to be useful in as a flexible tool, from identifying nonlinear correlations and phase transitions to network inference in genetic networks [30]. TDMI implies that we decompose the system X in two signals, which represent the evolution of same process, but are separated by a time lag. In this way, spatio-temporal interactions X Y are lost, and we measure the degree of causal correlations between them. We are also interested in considering the interaction between the components of the system. Then we now partition the system into subsystems containing a defined number of channels in each one, and measure how tightly interconnected the subsystems are in terms of information exchange. In our case using binary time series the full set of N bits can be now divided into non-overlapping non-empty subsets of bits. Such a splitting is commonly a bipartition Π = A B such that X = X A X B , due to computational tractability. Other partition scheme is the atomic partition, in which every subsystem contains one channel only. This latter one provides an upper bound for the measure of integration in a system, because we neglect the interactions of every X i , for i = 1 , . . . , N . Thus, this scheme provides the maximum information loss.
IIT has proposed different versions to measure integration of information, in which the fundamental idea is to quantify the loss of information by assuming the whole system partitioned in its components [13,20,21,22]. Different complexities have been suggested as potential measures of integrated information expressed as the distance between the probability distribution of the actual system and the product of probability distributions associated with its parts, provided some theoretical constraints [14]. Several measures have been derived following this approach although the empirical calculation is not plausible in some cases due to theoretical conditions and computational demands. In this paper, we examine three different measures which allow practical computations from experimental data using empirical distribution probabilities.
The first measure we use is the whole-minus-sum integrated information introduced by Barrett and Seth Φ W M S [42,47], defined as
Φ W M S ( X Y ) : = Φ W M S ( X Y ; MIB )
where Φ W M S denotes the effective information
Φ W M S ( X Y ; AB ) : = I ( X Y ) ( I ( X A Y A ) + I ( X B Y B ) )
MIB AB is the bipartition with the minimum information loss quantified by the mutual information between groups of channels.
This measure can be interpreted as the amount of integration of the components of a system through the creation of information when the system acts as a whole as compared with the information generated by its parts. However, it has been criticized since it does not satisfy the positivity requirement of IIT which makes it difficult for interpretation. Still, we find it plausible for discussion with more sophisticated approaches since it determines, from an information-theoretical point of view, the degree of synergetic or redundant influences of the whole system in its evolution. Thus, giving a practical insight to the global integration of causal interactions.
Another measure is the decoder-based integrated information Φ * proposed in [13]. As in (10), this quantity aims to measure global integrations through the difference of mutual information between the actual system and a partitioned one, where the parts are assumed independent, and satisfies the theoretical postulates of IIT. Thus, Φ * quantifies the extent of information loss when connections in the system are neglected through a mismatched probability distribution. This can be written as follows,
Φ * ( X Y ; MIB ) : = I ( X Y ) I * ( X Y ; MIB ) I * ( X Y ; MIB ) = max β I ˜ ( X Y ; β )
where I ˜ denotes the mismatched decoding through the expression
I ˜ ( X Y ; β , MIB ) = X p ( X ) log Y p ( Y ) q ( X | Y ) β    + X , Y p ( X Y ) log q ( X | Y ) β
where β is a parameter obtained from a gradient descent approach, and q ( X | Y ) denotes the mismatched decoding probability distribution defined for bipartitions AB as
q ( X | Y ) = p ( X A | Y A ) p ( X B | Y B )
Finally, the stochastic interaction Φ ˜ introduced in [12] is defined as
Φ ˜ ( X Y ; AB ) = H ( Y A | X A ) + H ( Y B | X B ) H ( Y | X )
It quantifies the global stochastic interaction in the system dynamics measured through the total uncertainty reduction of each of its parts during a transition to future states, where it is assumed that each part at each time step has knowledge of its own state and its complementary. It can also be interpreted as the extent of how tightly all parts of a system are in its dynamics in terms of information exchange more than the sum of its parts.
It can be shown that (14) can be written as
Φ ˜ ( X Y ; AB ) = Φ W M S ( X Y ; AB ) + I ( Y A Y B )
In this way, the stochastic interaction encodes the spatial correlation I ( Y A Y B ) measured through the mutual information between halves of the system and the net synergy.
For this work, we use the protein concentration of a chosen gene of the repressilator circuit from each cell to encode the global dynamics in the network. Thus, for each time step, the global state will be defined as a binary segment of length N, where each digit represents the binarized state from an individual cell using the binarization scheme previously explained (7). Each state has been assigned with a probability of occurrence, and 2 N is the maximum number of possible states for time step. Finally, for the integrated information measures were computed using the ’Practical PHI toolbox’ and we apply the Queyranne algorithm [13,14,48,49] for the MIB search.

1.5. Synchronization Properties

The coupled repressilators with quorum sensing exhibit several oscillator configurations: from in-phase behavior, splay-states, chaos, and clustering. We consider the generalized order parameters [50,51,52]
R k ( t ) = e i k θ j ( t ) j , j = 1 , . . . , N , k N
where θ j ( t ) is obtained from (5). Thus, it is possible to quantify the degree of k -cluster synchronization and asymmetries. In addition, the fluctuations in the order parameter can be computed through its variance
σ 2 ( R k ) = var t R k ( t )

2. Results

We show the plots of the measures of integrated information described above from which we contrast their differences for the specific genetic repressilator architectures. For each coupling scheme we show the qualitative trends for each measure in terms of the cell density encoded in the coupling and system size and compare the results with the underlying dynamics assessed by other standard approaches. The analysis will be split according to the different types of coupling.
In each case, the time series are generated from Equations (1) and (2), where S e is determined from (3) (phase-attractive coupling) or (4) (phase-repulsive coupling). A fourth order Runge-Kutta method with randomized initial conditions is used for integration. We select a step size of 0.05 and 1.5 × 10 6 steps from which we remove d 5 × 10 5 time steps to avoid transient effects. Probabilities are computed from the frequency counted empirically from where we get all the information-theoretic measures in bits.
The results presented in the following sections consider the bipartition scheme unless it is stated otherwise. Due to explanatory purposes when we discuss the integrated information as a function of the system size we will use the atomic partition. The reason comes from the intractability of computing in a reasonable time any candidate measure of integrated information for larger values of N. However, from the simulations in this paper we find that the qualitative trends are the same for intermediate sizes N, being just slightly higher than the bipartition scheme. This is because the atomic partition allows to maximize the informational loss as the system is split completely [13,53]. In these terms we can regard this measurement as an upper bound for the integrated information using any other partition scheme. Moreover, cell populations appear as interacting elements in large scale. Thus, it is worth understanding the influence of the size in the balance of segregation and integration of the system dynamics, which can be computed through the assumption of a split system in all its parts.

2.1. Phase Synchronization

In realistic conditions the population of repressilators exhibits important diversity in its oscillatory behavior as a consequence of cell diversity. Following [35], the lifetime ratio β plays a key role in defining the period length. Thus, we encode this diversity within the community by assuming β non-uniformly distributed. In particular, we choose a normal distribution with mean 1.0 and standard deviation Δ β . What we observe is a set of oscillations with a broad distribution of uncoupled phases. For increasing coupling there is a transition to phase synchronization, such that the larger is the variation in β the larger Q we need to reach the phase-locking state. In Figure 3c we show as complexity measure the variance of the order parameter R 1 defined in (17) averaged over 10 3 oscillators for different Q and Δ β 2 = 0.005 . The system has negligible fluctuations for small coupling which are maintained constant as intercell effects are not strong enough. As cell density increases, there is a partial phase-locking which increases suddenly to complete synchrony. At the same time, the fluctuations in R 1 span abruptly from zero to its maximum value around the transition regime at Q 0.68 . This range of transition agrees with the García-Ojalvo’s order parameter reported with the original model.

2.1.1. Integration and Synchronization

Next, we take N = 4 cells to exemplify to which extent the synchronization process is captured by the measures of integrated information. In Figure 3a we see the measures drawn from Equations (9)–(15) using τ = 300 time steps, corresponding to approximately the mean period of oscillation of the cell population. As observed, the measures Φ W M S and Φ * monotonically increase from zero reaching its peak as the transition point is approached and drop sharply with similar behavior after the critical point. Then, the whole-minus-sum measure Φ W M S reaches negative values while the decoder-based integration drops to zero. The stochastic interaction Φ ˜ follows the same trend as the other Φ measures before the transition regime but increases as the system synchronizes.
Thus, both Φ W M S and Φ * indicate that at the transition point to phase synchronization, the system is responsive to its own states more than the aggregation of the individual responses. After reaching synchronization there is high integration and segregation is negligible. As a consequence, these measures show a drop to lower values. This is not the case for Φ ˜ , which seems to increase as the system integrates more. It is worth understanding the influence of the spatial segregation and integration in the dynamics, to relate it with our previous results. We quantify the mean mutual information between all possible halves of the system for increasing coupling denoted as I A B = I ( X A X B ) , Equation (8) (Figure 3a). As expected, for small values of coupling, there is small correlation which is enhanced during the phase transition with a step-like trend. Next, it reaches a plateau in the strong coupling regime. Thus, we see that Φ ˜ is dominated by the behavior of the spatial correlation in the process X t , Φ W M S interprets the high correlation of the synchronized system as redundancy, and Φ * vanishes due to high synchronization between cells.
The trends described above depend on the lag τ under consideration. In particular, we observe two main cases: for τ where the TDMI is maximum, the measures Φ W M S and Φ * are maximized at criticality, while for τ where the TDMI is minimum, the is a general tendency of Φ W M S and Φ * to decrease towards the super-critical regime. Meanwhile, Φ ˜ keeps the same behavior already described for any lag.
Figure 4 shows the transition depicted by Φ W M S , Φ * and Φ ˜ in terms of Q and τ for Δ β 2 = 0.005 . When considering the relationship with the mutual information I ( X t , X t + τ ) we observe that the integrated information measures are maximized when the TDMI is minimized, i.e., when X t and X t + τ have the least redundancy. This is the situation where X t and X t + τ are out-of-phase. In a similar way, the integrated information measures are minimized when X t and X t + τ are in-phase. This holds true for all values of Q for Φ W M S and Φ ˜ .
In the case of Φ * , the aforementioned feature is satisfied before synchronization, but after the phase transition it vanishes for all τ . Thus, from the perspective of Φ W M S and Φ * the whole system has more information about its current state given its past than its individual components when X t and X t + τ are sufficiently independent but still preserving some degree of correlation.

2.1.2. Integration and Scaling Behavior

In general, the qualitative trends shown in Figure 4 hold regardless of the system size. However, it is worth discussing the influence of the TDMI and spatial correlation in the measures of integrated information. In Figure 5a,b we show the whole-minus-sum integration and the normalized stochastic interaction across different sizes N and lags τ . As discussed previously, it is observed that regardless of the system size the integration measures are symmetric with respect to the TDMI. In particular, when the TMDI is minimized, Φ W M S grows with a concave downward trend for increasing N while for lags τ in which the TDMI is maximized, Φ W M S becomes increasingly negative. In contrast, Φ ˜ keeps a linear growth independent of τ for large N with fixed coupling. This means that if X t + τ encodes the most possible information of X t , the integrated information will be maximized, reaching a plateau for increasing system size. Meanwhile, if X t and X t + τ are highly correlated, the components of the system will hold more information about their own state than the system.
Next, we study how the different measures change for increasing Q using the first lag that minimizes the TDMI, τ = 60 to maximize the amount of influence between X t and X t + τ . Figure 5d shows different complexities, including the integrated information measures and the total correlation defined as C ( X ) = i = 1 N H ( X i ) H ( X ) for an intermediate size N = 15 (observe that C ( X ) replaces I ( X A X B ) in Equation (14) when we consider the atomic partition). As it is shown, for coupling Q before the transition to synchronization Φ W M S and Φ * remain steady followed by a drop to a low value of integrated information just right after the cutoff point for synchronization. Thus, at the transition point, both measures are maximally sensitive and decrease in the super-critical regime where the system loses segregation.
Interestingly, the decoder-based integration has a negligible integrated information when compared with the whole-minus-sum integration despite showing a similar trend. The dependency between them is not entirely clear, but one can infer that the mismatched decoding is assuming that as the system grows, the patterns in the sub-critical region are noise driven due to the uncoupling between cells (high segregation) followed by a transition to a high correlated regime (high integration). This is out of the scope of this paper and further investigation is needed.
Meanwhile, Φ ˜ remains steady for Q spanning from zero to just before criticality, then it dips around the transition point, and stabilizes and remains constant at a higher value of integrated information with respect to the unsynchronized state. In comparison with Φ W M S and Φ * , this measure is dominated by C for large N, which scales linearly, Figure 5c. On the other hand, Φ ˜ dips because of the contribution from Φ W M S is reduced while the total correlation C increases due to the phase synchronization.
From a biological perspective these results suggest that the coupling at the transition point to synchronization would be favored to keep a balance between integration and differentiation of the regulatory dynamics in a coherent way. High integration leads the system to exhibit split dynamics while some degree of segregation provides evolutionary advantage to the whole system to act to internal and external demands.

2.2. Phase-Repulsive Coupling

2.2.1. Irregular Self-Oscillations

In Figure 6a we show the numerical simulations of the bifurcation plot for N = 4 cells for the model (Section 1.3). It is observed regular oscillations which are completely out-of-phase up to around Q = 0.49 followed by the stable formation of subgroups of synchronized cells (clustering). As we reach the first torus bifurcation T R 1 around Q * = 0.58 there is emergence of chaos reflected in oscillations of amplitude with increasing irregularity [40]. This behavior is maintained just before the second bifurcation point T R 2 around Q = 1.13 in which regular oscillations are recovered. This behavior is observed across different number of cells for N 4 for this analysis.
As a proof of concept, we characterize the cluster distributions through the 1,2-cluster synchronization (16) and relate it with the measures of integrated information. From previous results about clustering of coupled repressilators [40], clusters are defined in terms of the absolute difference of the concentration levels of the signaling molecule S i ( t ) . Here, the cluster ordering is measured from R 1 , 2 using the phase of protein oscillations B i ( t ) . However, the main feature about the cluster dynamics is equivalent to that described by the [40]. This is the enhanced switching from one cluster distribution to another for increasing degree of chaos, as we describe below.
Figure 6b shows R 1 , 2 order parameters for increasing Q. Each data point indicates the degree of 1,2-cluster synchronization averaged in time. As discussed previously in [41], clusters can be either stable and therefore not changing its composition along the time, or transient, which implies the formation of an initial cluster that is later decomposed after some time steps in another arrangement and so forth. The full history of formation and decomposition of clusters is uniquely determined by the initial state. In addition, the lifetime of one cluster before switching to another arrangement will depend on its degree of stability. Consequently, we get multiple branches for increasing Q with each branch corresponding to a basin of attraction. Hence, provided some initial conditions the oscillators settle down to different cluster distributions with a defined degree of order.
Now, following R 2 (Figure 6b) we observe that one group displays fully synchronous behavior, corresponding to a 2:2 cluster as N = 4 , while other keeps its phase unsynchronized. The lower branch slowly rises at Q * followed by the appearance of two new groups at Q = 0.64 . Meanwhile, there is a sudden decrease of the upper branch. Later, these clusters are broken in multiple branches between Q = 0.67 and Q = 0.75 . Up to this point we see that the stability of the arrangement between oscillators is continuously reduced for increasing Q as a response of the enhanced formation and decomposition of clusters in some time span, so the temporal average synchronization is reduced for phase-locked 2:2 clusters. This trend is maximized at Q = 0.75 since right after this value we observe a sequence of points which defines a very disperse and irregular branch just before Q = 0.95 . This region corresponds to the fully developed chaos, in which there is a continuous formation and decompositions between cluster groups for an initial distribution of oscillators. As chaos increases the lifetime of such clusters is severely shortened [40] followed by the switch to new distribution throughout the time. This grouping ability losses stability to become more sensitive to initial conditions implying an irregular shape. Around Q = 0.95 there is a peak to complete 2-cluster phase-locking, which suggests a narrow window in which chaos reverts to its periodic motion as a 2:2 cluster. Finally, there appear again multiple branches leading to the transition towards the initial regular oscillations. On the other hand, order parameter R 1 provides the degree of nonuniform distribution of oscillators in different clusters. As in the case of R 2 , we see the emergence of branches leading to a unique group with high asymmetry as chaos is increased. In contrast, lower values of R 1 are attained in the stable region outside of both torus bifurcation points reflecting larger symmetry in the cluster oscillations.
Finally, if we run simulations for larger system sizes the cluster distributions clearly increase their allowed arrangements as the system size grows. Consequently, the degree of cluster synchrony and asymmetry can follow a very disperse trend instead of well-defined branches even in the stable region, although we point out that the behavior for strong chaos keeps very similar. This fact is reflected in the measures of integrated information as discussed below.

2.2.2. Chaos and Integration

Next, we relate the previous description with the trends followed by the integrated information measures. In Figure 7 we show Φ W M S , Φ ˜ and Φ * for increasing coupling Q and lag τ for N = 4 .
Compared to Figure 6b one can observe that the integration measures appear to track the broken symmetry due to the formation of new branches for increasing Q. First, in the clustering regime Φ W M S and Φ ˜ hold a periodic trend for increasing τ as a consequence of the regular oscillations of the cluster formations. Within this regime oscillations are highly correlated, so that the stochastic interaction is greater in comparison with the other measures. Meanwhile, Φ * represents this regime with null integrated information, as the degree of complexity is minimum during synchronization.
Next, during the onset of chaos, all measures detect the irregular oscillations following a similar trend when we vary τ . However, they differ as the coupling Q increases (Figure 8). Both Φ W M S and Φ * start with a lower value which is increased non-uniformly along Q with the appearance of weak chaos. In turn, there is a drop as the chaotic behavior is enhanced. Meanwhile, Φ ˜ goes from high values of integration due to the stable cluster formation and decreases slightly at Q * , followed by a second drop as the spatial interaction collapses to its minimum value.
During the strong chaotic regime all complexity measures decay to zero with an oscillatory trend in the limit τ , since X t and X t + τ (and the individual system components X i , t and X i , t + τ ) are no longer correlated, so the TDMI vanishes. Meanwhile, for the stochastic interaction, the remaining interaction is the spatial correlation encoded in C, which seems to be decreased within this regime as shown. Nevertheless, Φ ˜ peaks around Q = 0.95 while Φ W M S becomes more negative and Φ * vanishes. This suggests a periodic window given that the system becomes highly correlated in the view of Φ ˜ and Φ W M S .
Finally, just before Q = 1.1 the maximum amplitudes of the bifurcation plot seem to group in three different regions as observed in Figure 6a, implying an increasing stability ending with the recovery of the periodic behavior after T R 2 . As a result, Φ W M S and Φ * increase qualitatively similar during the formation of such groups followed by a drop after the torus bifurcation. In particular, Φ W M S reaches negative values while Φ * keeps negligible due to the regular clustering synchronizations. In contrast, Φ ˜ increases progressively reaching high values of integrated information at the torus bifurcation point and stabilizes at a similar value attained just before T R 1 .

2.2.3. Temporal Mixing Increases Integrated Information

As suggested in the past section, the system size plays an important role in defining the degree of dynamical complexity. Again, we make use of the atomic partition to describe the integrated information. Figure 8 shows Φ W M S , Φ * and Φ ˜ for increasing Q at different N. In particular, we choose a lag τ = 220 time steps, as it keeps close to the first minimum of the TDMI for the range of Q considered, so that it reduces as much as possible the redundancy between X t and X t + τ , although both processes are not entirely independent.
Just before T R 1 we see that for increasing N the measure Φ W M S has a highly disperse arrangement with multiple values of integrated information. This is a result of having several stable cluster distributions as the system grows, each with a certain degree of complexity. Next, after Q * , and before Q = 0.75 , the value Φ W M S is no longer spread and grows smoothly with N. In contrast, for fully developed chaos, Φ W M S starts with small values increasing with N and defines a peak between Q = 0.8 and Q = 0.95 . This is precisely the range in which the chaotic behavior is maximized according to [40].
As discussed in the section about phase-attractive coupling, the increasing of Φ W M S and Φ * during chaos could be associated with the patterns accessible to the system. The formation and decomposition of clusters favor the accessible states encoded as binary arrays that scale with the system size, but keep upper bounded by the TDMI. Therefore, we also find a saturation of complexity as the system grows as evidenced in Figure 9a. Nevertheless, observe that in all the cases Φ W M S attributes higher integrated information to the system than Φ * , which means that the information loss due to the mismatch decoding I * from Φ * is greater. Still, both measures reflect the same properties in the system. This suggests that the multistability provided by the switching of cellular fates increases the balance of segregation and integrations of the causal dynamics of whole system in comparison with the one of its single components leading to the emergence of integrated information. Moreover, this balance has a limiting amount provided by the asymptotic trend of TDMI as the system becomes larger. Next, the whole-minus-sum integration Φ W M S starts decreasing as the oscillation amplitudes in the bifurcation plot are less spread. In particular, we observe that Φ W M S drops during the formation of clustered amplitudes in three groups (Figure 6a) followed by a second drop to negative values after T R 2 .
On the other hand, observe that the stochastic interaction Φ ˜ decreases as the number of branches increases. In these conditions the cluster lifetime is shortened implying a decrease of spatial correlations. This effect is quantified by the mutual information between halves I ( X A , X B ) . As observed the qualitative trend is clearly similar to Φ ˜ , so the stochastic information is just reflecting spatial correlation (integration) but not segregation due to the formation of patterns. In addition, as shown in Figure 9a the contribution of the spatial correlation becomes stronger for larger N since C scales linearly after N = 50 cells. Then, from the perspective of the stochastic interaction this rapid switching in the system could avoid synergistic influences on its future states more than the independent oscillators.

3. Conclusions

In this work, we assess the degree of integrated information using whole-minus-sum integration, decoder-based integration and stochastic interaction using a model of genetic oscillators coupled by quorum sensing. This work includes two coupling architectures leading to different dynamical regimes representing cellular fates. The phase-attractive coupling allows phase synchronization for increasing cell density, and the phase-repulsive coupling favors cluster formation with increased switching and asymmetry with the degree of chaos.
Our findings suggest that these measures can proxy the transition between dynamical regimes encoded as binary patterns. The whole-minus-sum integration and decoder-based integration are observed to drop in the transition to phase synchronization of cells. In particular, whole-minus-sum integration becomes negative for highly correlated causal interactions between the current and past states. Meanwhile, decoder-based integration vanishes when the system becomes more synchronized. This is because high integration with no segregation implies a low value of integrated information. Thus, these measures have maximal susceptibility near criticality. Also, both measures are increased when there is a balance between the segregation and integration of oscillations due to the transient stability of cellular fates i.e., frequent switching of synchronized clusters, a typical mechanism for decision-making genetic regulatory networks. Both observations suggest that in terms of integration of information, it is evolutionary stable for cells to maintain at criticality prior full synchronization to allow available responses from the cells to intrinsic and extrinsic demands and keeping some degree of integration. In a similar way, the diversity and switching between different attractors not just provide an adaptive advantage to the system but according to our results also maximizes the integration of independent cell dynamics in a coherent way.
Meanwhile, the stochastic interaction, is mainly driven by the spatial correlation and it is not sensitive to the diversity of available states as the other measures considered in this paper. During the transition to synchronization it dips around the critical point due to a combined effect of the decrease of segregation, but it is later enhanced as the system synchronizes favoring full integration. Also, in the chaotic regime, it seems that lower cluster lifetime between switching to new distributions decreases its value, while a larger one favors it. As an overall, for this analysis this measure did not provide better insight than the mutual information between subsystems could.
In all cases, the integrated information measures are symmetric with respect to the TDMI of the whole system. This implies that the balance of segregation and integration of the causal dynamics is favored when the current state of the system shares the less possible information with the past states, but still with some degree of correlation.
Therefore, this suggests that cells, as a system of interacting genetic networks, show enough complexity to exhibit cause-effect power as a whole above and beyond its parts, a required step to develop consciousness according to IIT. However, as we believe, consciousness has not been evolutionarily developed on the genetic level, probably due to the slow communication speed based on diffusion. Further research is required to have a better understanding of the evolutionary mechanisms for higher cognitive processes.

Author Contributions

Conceptualization, L.A. and A.Z.; Software and Investigation, L.A.; Writing—Original Draft Preparation, L.A.; Writing—Review and Editing, L.A. and A.Z.; Supervision, A.Z.

Funding

A.Z. acknowledges support from Russia (RU) under Grant No. 074-02-2018-330. L.A. acknowledges CONACYT for his doctoral scholarship to conclude this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lyon, P. The cognitive cell: Bacterial behavior reconsidered. Front. Microbiol. 2015, 6, 264. [Google Scholar] [CrossRef] [PubMed]
  2. Allman, J.M. Evolving Brains; Scientific American Library: New York, NY, USA, 1999. [Google Scholar]
  3. Greenspan, R.J. An Introduction to Nervous Systems; Cold Spring Harbor Laboratory Press: Cold Spring Harbor, NY, USA, 2007. [Google Scholar]
  4. Shapiro, J.A. Thinking about bacterial populations as multicellular organisms. Annu. Rev. Microbiol. 1998, 52, 81–104. [Google Scholar] [CrossRef] [PubMed]
  5. Koseska, A.; Bastiaens, P.I. Cell signaling as a cognitive process. EMBO J. 2017, 36, 568–582. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Crick, F. Central Dogma of Molecular Biology. Nature 1970, 227, 561–563. [Google Scholar] [CrossRef]
  7. Joyce, G.F. Bit by bit: The Darwinian basis of life. PLoS Biol. 2012, 10, e1001323. [Google Scholar] [CrossRef] [PubMed]
  8. Ben-Jacob, E.; Lu, M.; Schultz, D.; Onuchic, J.N. The physics of bacterial decision making. Front. Cell. Infect. Microbiol. 2014, 4, 154. [Google Scholar] [CrossRef]
  9. Shettleworth, S.J. Where is the comparison in comparative cognition? Alternative research programs. Psychol. Sci. 1993, 4, 179–184. [Google Scholar] [CrossRef]
  10. Bourgine, P.; Stewart, J. Autopoiesis and cognition. Artif. Life 2004, 10, 327–345. [Google Scholar] [CrossRef]
  11. Maturana, H.R.; Varela, F.J. Autopoiesis and Cognition: The Realization of the Living; D. Reidel Pub. Co.: Dordrecht, The Netherlands, 1980. [Google Scholar]
  12. Ay, N. Information geometry on complexity and stochastic interaction. Entropy 2015, 17, 2432–2458. [Google Scholar] [CrossRef]
  13. Oizumi, M.; Amari, S.I.; Yanagawa, T.; Fujii, N.; Tsuchiya, N. Measuring integrated information from the decoding perspective. PLoS Comput. Biol. 2016, 12, e1004654. [Google Scholar] [CrossRef]
  14. Oizumi, M.; Tsuchiya, N.; Amari, S. Unified framework for information integration based on information geometry. Proc. Natl. Acad. Sci. USA 2016, 20, 14817–14822. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, T.S.; Chen, L.N.; Aihara, K. Molecular Communication through Stochastic Synchronization Induced by Extracellular Fluctuations. Phys. Rev. Lett. 2005, 95, 178103. [Google Scholar] [CrossRef] [PubMed]
  16. Zhou, T.S.; Zhang, J.J.; Yuan, Z.J.; Xu, A.L. External Stimuli Mediate Collective Rhythms: Artificial Control Strategies. PLoS ONE 2007, 2, e231. [Google Scholar] [CrossRef]
  17. McMillen, D.; Kopell, N.; Hasty, J.; Collins, J. Synchronizing genetic relaxation oscillators by intercell signaling. Proc. Natl. Acad. Sci. USA 2002, 99, 679–684. [Google Scholar] [CrossRef] [Green Version]
  18. Pikovsky, A.; Rosenblum, M.; Kurths, J. Synchronization: A Universal Concept in Nonlinear Sciences; Cambridge University Press: New York, NY, USA, 2001. [Google Scholar]
  19. Kauffman, S.A. The sciences of complexity and origins of order. In The Principles of Organization in Organisms; Mittenthal, J.E., Baskin, A.B., Eds.; Addison-Wesley: New York, NY, USA, 1992. [Google Scholar]
  20. Tononi, G. Consciousness as integrated information: A provisional manifesto. Biol. Bull. 2008, 215, 216–242. [Google Scholar] [CrossRef]
  21. Balduzzi, D.; Tononi, G. Integrated information in discrete dynamical systems: Motivation and theoretical framework. PLoS Comput. Biol. 2008, 4, e1000091. [Google Scholar] [CrossRef] [PubMed]
  22. Oizumi, M.; Albantakis, L.; Tononi, G. From the Phenomenology to the Mechanisms of Consciousness: Integrated Information Theory 3.0. PLoS Comput. Biol. 2014, 10, e1003588. [Google Scholar] [CrossRef] [PubMed]
  23. Tegmark, M. Improved measures of integrated information. PLoS Comput. Biol. 2016, 12, e1005123. [Google Scholar] [CrossRef] [PubMed]
  24. Kim, H.; Hudetz, A.G.; Lee, J.; Mashour, G.A.; Lee, U.; ReCCognition Study Group. Estimating the Integrated Information Measure Phi from High-Density Electroencephalography during States of Consciousness in Humans. Front. Hum. Neurosci. 2018, 16, 12–42. [Google Scholar]
  25. Tononi, G.; Sporns, O.; Edelman, G.M. A measure for brain complexity: Relating functional segregation and integration in the nervous system. Proc. Natl. Acad. Sci. USA 1994, 91, 5033–5037. [Google Scholar] [CrossRef] [PubMed]
  26. Mediano, P.A.; Farah, J.C.; Shanahan, M. Integrated information and metastability in systems of coupled oscillators. arXiv, 2016; arXiv:1606.08313. [Google Scholar]
  27. Kanwal, M.S.; Grochow, J.A.; Ay, N. Comparing information-theoretic measures of complexity in Boltzmann machines. Entropy 2017, 19, 16. [Google Scholar] [CrossRef]
  28. Zhao, W.; Serpedin, E.; Dougherty, E.R. Inferring gene regulatory networks from time series data using the minimum description length principle. Bioinformatics 2006, 22, 2129–2135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Teschendorff, A.E.; Sollich, P.; Kuehn, R. Signaling entropy: A novel network-theoretical framework for systems analysis and interpretation of functional. Methods 2014, 67, 282–293. [Google Scholar] [CrossRef] [PubMed]
  30. Meyer, P.E.; Konton, K.; Lafitte, F.; Bontempi, G. Information-theoretic inference of large transcriptional regulatory networks. EURASIP J. Bioinform. Syst. Biol. 2007, 2007, 9879. [Google Scholar] [CrossRef] [PubMed]
  31. Tkacik, G.; Walczak, A.M. Information transmission in genetic regulatory networks: A review. J. Phys. Condens. Matter 2011, 23, 153102. [Google Scholar] [CrossRef] [PubMed]
  32. Waltermann, C.; Klipp, E. Information theory based approaches to cellular signaling. Biochim. Biophys. Acta (BBA)-Gen. Subj. 2011, 1810, 924–932. [Google Scholar] [CrossRef] [PubMed]
  33. Manke, T.; Demetrius, L.; Vingron, M. Lethality and entropy of protein interaction networks. Genome Inf. Ser. 2005, 16, 159–163. [Google Scholar]
  34. Wagner, A. From bit to it: How a complex metabolic network transforms information into living matter. BMC Syst. Biol. 2007, 1, 33. [Google Scholar] [CrossRef] [PubMed]
  35. García-Ojalvo, J.; Elowitz, M.B.; Strogatz, S.H. Modeling a synthetic multicellular clock: Repressilators coupled by quorum sensing. Proc. Natl. Acad. Sci. USA 2004, 101, 10955–10960. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Taga, M.E.; Bassler, B.L. Chemical communication among bacteria. Proc. Natl. Acad. Sci. USA 2003, 100, 14549–14554. [Google Scholar] [CrossRef] [Green Version]
  37. Kaneko, K.; Yomo, T. Cell division, differentiation and dynamic clustering. Phys. D 1994, 75, 89–102. [Google Scholar] [CrossRef] [Green Version]
  38. Ullner, E.; Koseska, A.; Kurths, J.; Volkov, E.; Kantz, H.; García-Ojalvo, J. Multistability of synthetic genetic networks with repressive cell-to-cell communication. Phys. Rev. E 2008, 78, 031904. [Google Scholar] [CrossRef] [PubMed]
  39. Ullner, E.; Zaikin, A.; Volkov, E.I.; García-Ojalvo, J. Multistability and clustering in a population of cellular genetic oscillators via phase repulsive cell-to-cell communication. Phys. Rev. Lett. 2007, 99, 148103. [Google Scholar] [CrossRef] [PubMed]
  40. Koseska, A.; Ullner, E.; Volkov, E.; Kurths, J.; García-Ojalvo, J. Cooperative differentiation through clustering in multicellular populations. J. Theor. Biol. 2010, 263, 189–202. [Google Scholar] [CrossRef]
  41. Koseska, A.; Volkov, E.; Zaikin, A.; Kurths, J. Inherent multistability in arrays of autoinducer coupled genetic oscillators. Phys. Rev. E 2007, 75, 031916. [Google Scholar] [CrossRef] [PubMed]
  42. Barrett, A.B.; Seth, A.K. Practical measures of integrated information for time series data. PLoS Comput. Biol. 2011, 7, e1001052. [Google Scholar] [CrossRef] [PubMed]
  43. Seth, A.K.; Barrett, A.B.; Barnett, L. Causal density and integrated information as measures of conscious level. Philos. Trans. R. Soc. Lond. A Math. Phys. Eng. Sci. 2011, 369, 3748–3767. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Elowitz, M.; Leibler, S. A synthetic oscillatory network of transcriptional regulators. Nature 2000, 403, 335–338. [Google Scholar] [CrossRef]
  45. Schartner, M.; Seth, A.; Noirhomme, Q.; Boly, M.; Bruno, M.A.; Laureys, S.; BarretT, A. Complexity of Multi-Dimensional Spontaneous EEG Decreases during Propofol Induced General Anaesthesia. PLoS ONE 2015, 10, e0133532. [Google Scholar] [CrossRef] [PubMed]
  46. Cover, T.M.; Thomas, J.A. Elements of Information Theory; John Wiley and Sons: Hoboken, NJ, USA, 2012. [Google Scholar]
  47. Griffith, V.; Koch, C. Quantifying synergistic mutual information. In Guided Self-Organization: Inception; Prokopenko, M., Ed.; Emergence, Complexity and Computation; Springer: Berlin/Heidelberg, Germany, 2014; Volume 9, pp. 159–190. [Google Scholar]
  48. Hidaka, S.; Oizumi, M. Fast and exact search for the partition with minimal information loss. PLoS ONE 2018, 13, e0201126. [Google Scholar] [CrossRef]
  49. Kitazono, J.; Kanai, R.; Oizumi, M. Efficient algorithms for searching the minimum information partition in integrated information theory. Entropy 2018, 20, 173. [Google Scholar] [CrossRef]
  50. Skardal, P.S.; Ott, E.; Restrepo, J.G. Cluster synchrony in systems of coupled phase oscillators with higher-order coupling. Phys. Rev. E 2011, 84, 036208. [Google Scholar] [CrossRef]
  51. Nordenfelt, A. An entropy based clustering order parameter for finite ensembles of oscillators. Chaos 2015, 25, 113110. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Osipov, G.; Kurths, J.; Zhou, C. Synchronization in Oscillatory Networks; Springer: Berlin, Germany, 2007. [Google Scholar]
  53. Toker, D.; Sommer, F. Moving Past the Minimum Information Partition: How to Quickly and Accurately Calculate Integrated Information. arXiv, 2016; arXiv:1605.01096. [Google Scholar]
Figure 1. Diagram of the repressilator circuit including the quorum sensing mechanism. Left: the phase-attractive coupling. Right: the phase-repulsive coupling.
Figure 1. Diagram of the repressilator circuit including the quorum sensing mechanism. Left: the phase-attractive coupling. Right: the phase-repulsive coupling.
Entropy 21 00382 g001
Figure 2. Time series of protein levels B i for different coupling Q. (ac) phase-attractive coupling with N = 100 and κ = 20 . Cell variability is in terms of lifetime ratios β a , b , c N ( 1 , 0.003 ) ; (df) phase-repulsive coupling with N = 4 cells, κ = 25 , β a = 0.85 , β b , c = 0.1 ; (a,d) Q = 0.4 , (b,e) Q = 0.57 and (c,f) 0.8 . Other parameters are α = 216 , n = 2.6 , η = 2.0 , k s 0 = 1 and k s 1 = 0.01 . (ac) describe the transition to synchronization by increasing the coupling Q, (df) show the transition from stable clusters of regular oscillations to a chaotic behavior. Each color line represents the protein dynamics of each cell in the network.
Figure 2. Time series of protein levels B i for different coupling Q. (ac) phase-attractive coupling with N = 100 and κ = 20 . Cell variability is in terms of lifetime ratios β a , b , c N ( 1 , 0.003 ) ; (df) phase-repulsive coupling with N = 4 cells, κ = 25 , β a = 0.85 , β b , c = 0.1 ; (a,d) Q = 0.4 , (b,e) Q = 0.57 and (c,f) 0.8 . Other parameters are α = 216 , n = 2.6 , η = 2.0 , k s 0 = 1 and k s 1 = 0.01 . (ac) describe the transition to synchronization by increasing the coupling Q, (df) show the transition from stable clusters of regular oscillations to a chaotic behavior. Each color line represents the protein dynamics of each cell in the network.
Entropy 21 00382 g002
Figure 3. (a) Mutual information I A B and integrated information measures Φ W M S , Φ * and Φ ˜ for increasing coupling Q. (b) Mutual information I ( X t , X t + τ ) and Φ * for increasing τ and Q = 0.5 . (c) Order parameter R 1 and its variance for increasing Q. Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Figure 3. (a) Mutual information I A B and integrated information measures Φ W M S , Φ * and Φ ˜ for increasing coupling Q. (b) Mutual information I ( X t , X t + τ ) and Φ * for increasing τ and Q = 0.5 . (c) Order parameter R 1 and its variance for increasing Q. Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Entropy 21 00382 g003
Figure 4. Measures of integrated information for increasing Q and τ . Lifetime ratio β N ( 1 , 0.005 ) . Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Figure 4. Measures of integrated information for increasing Q and τ . Lifetime ratio β N ( 1 , 0.005 ) . Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Entropy 21 00382 g004
Figure 5. (a) Whole-minus-sum integration Φ W M S and (b) stochastic interaction Φ ˜ for increasing τ for different system sizes N. (c) Mutual information I ( X t , X t + τ ) and total correlation C ( X t ) for increasing size N. (d) Integration measures Φ W M S , Φ * and Φ ˜ for increasing coupling Q. For comparison we include C ( X t ) . For (ac) we use Q = 0.5 and (c,d) we use τ = 60 units with β N ( 1 , 0.005 ) . Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Figure 5. (a) Whole-minus-sum integration Φ W M S and (b) stochastic interaction Φ ˜ for increasing τ for different system sizes N. (c) Mutual information I ( X t , X t + τ ) and total correlation C ( X t ) for increasing size N. (d) Integration measures Φ W M S , Φ * and Φ ˜ for increasing coupling Q. For comparison we include C ( X t ) . For (ac) we use Q = 0.5 and (c,d) we use τ = 60 units with β N ( 1 , 0.005 ) . Other parameters are the same as in Figure 2 for the phase-attractive coupling.
Entropy 21 00382 g005
Figure 6. (a) Numerical bifurcation plot for the full amplitude oscillations and (b) order parameter R 1 , 2 for different values of coupling Q.
Figure 6. (a) Numerical bifurcation plot for the full amplitude oscillations and (b) order parameter R 1 , 2 for different values of coupling Q.
Entropy 21 00382 g006
Figure 7. Whole-minus-sum integration Φ W M S , stochastic interaction Φ ˜ and decoder-based integration Φ * for different values of coupling Q and lag τ for N = 4 cells. Other parameters are the same as in Figure 2 for the phase-repulsive coupling.
Figure 7. Whole-minus-sum integration Φ W M S , stochastic interaction Φ ˜ and decoder-based integration Φ * for different values of coupling Q and lag τ for N = 4 cells. Other parameters are the same as in Figure 2 for the phase-repulsive coupling.
Entropy 21 00382 g007
Figure 8. (a) whole-minus-sum integration Φ W M S , (b) decoder-based integration Φ * and (c) stochastic interaction Φ ˜ for different values of coupling Q and different number of cells N. We use τ = 220 time steps. Other parameters are the same as in Figure 2 for the case of phase-repulsive coupling.
Figure 8. (a) whole-minus-sum integration Φ W M S , (b) decoder-based integration Φ * and (c) stochastic interaction Φ ˜ for different values of coupling Q and different number of cells N. We use τ = 220 time steps. Other parameters are the same as in Figure 2 for the case of phase-repulsive coupling.
Entropy 21 00382 g008
Figure 9. (a) Total correlation C ( X t ) and TDMI I ( X t , X t + τ ) for increasing system size N and Q = 0.7 (weak chaotic regime) (b) Mean mutual information between all possible halves of the system I ( X A , X B ) for increasing coupling Q. We use τ = 220 time steps. Other parameters are the same as in Figure 2 for the case of phase-repulsive coupling.
Figure 9. (a) Total correlation C ( X t ) and TDMI I ( X t , X t + τ ) for increasing system size N and Q = 0.7 (weak chaotic regime) (b) Mean mutual information between all possible halves of the system I ( X A , X B ) for increasing coupling Q. We use τ = 220 time steps. Other parameters are the same as in Figure 2 for the case of phase-repulsive coupling.
Entropy 21 00382 g009

Share and Cite

MDPI and ACS Style

Abrego, L.; Zaikin, A. Integrated Information as a Measure of Cognitive Processes in Coupled Genetic Repressilators. Entropy 2019, 21, 382. https://doi.org/10.3390/e21040382

AMA Style

Abrego L, Zaikin A. Integrated Information as a Measure of Cognitive Processes in Coupled Genetic Repressilators. Entropy. 2019; 21(4):382. https://doi.org/10.3390/e21040382

Chicago/Turabian Style

Abrego, Luis, and Alexey Zaikin. 2019. "Integrated Information as a Measure of Cognitive Processes in Coupled Genetic Repressilators" Entropy 21, no. 4: 382. https://doi.org/10.3390/e21040382

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop