The apparent dichotomy between information-processing and dynamical approaches to complexity science forces researchers to choose between two diverging sets of tools and explanations, creating conflict and often hindering scientific progress. Nonetheless, given the shared theoretical goals between both approaches, it is reasonable to conjecture the existence of underlying common signatures that capture interesting behavior in both dynamical and information-processing systems. Here, we argue that a pragmatic use of integrated information theory (IIT), originally conceived in theoretical neuroscience, can provide a potential unifying framework to study complexity in general multivariate systems. By leveraging metrics put forward by the integrated information decomposition framework, our results reveal that integrated information can effectively capture surprisingly heterogeneous signatures of complexity—including metastability and criticality in networks of coupled oscillators as well as distributed computation and emergent stable particles in cellular automata—without relying on idiosyncratic, ad hoc criteria. These results show how an agnostic use of IIT can provide important steps toward bridging the gap between informational and dynamical approaches to complex systems.

Originally conceived within theoretical neuroscience, integrated information theory (IIT) has been rarely used in other fields—such as complex systems or non-linear dynamics—despite the great value it has to offer. In this article, we inspect the basics of IIT, dissociating it from its contentious claims about the nature of consciousness. Relieved of this philosophical burden, IIT presents itself as an appealing formal framework to study complexity in biological or artificial systems, applicable in a wide range of domains. To illustrate this, we present an exploration of integrated information in complex systems and relate it to other notions of complexity commonly used in systems such as coupled oscillators and cellular automata. Through these applications, we advocate for IIT as a valuable framework capable of revealing common threads between diverging branches of complexity science.

Most theories about complexity are rooted in either information theory or dynamical systems perspectives—two disciplines with very different aims and toolkits. The former, built after the work of Turing and Shannon, focuses mainly on discrete systems and considers complexity in terms of information processing, universal computation, distributed computation, and coherent emergent structures.1 The latter, following the tradition started by Poincaré and Birkhoff, focuses on continuous systems and studies their behavior using attractors, phase transitions, chaos, and metastability.2 

This methodological divide has contraposed how various communities of researchers think about complex systems, even to the extent of triggering some longstanding disagreements. This is particularly evident in the field of cognitive neuroscience, where proponents of computational approaches claim that the brain works similarly to a Turing machine,3–6 while opponents believe that cognitive processes are essentially continuous and rate-dependent.7–9 A related debate has taken place in the artificial intelligence community between symbolic and connectionist paradigms for the design of intelligent systems.10 Modern stances on these problems have promoted hybrid approaches,11 bypassing ontological arguments toward epistemological perspectives where both information and dynamics represent equally valid methods for enquiry.12 

Interestingly, bridging the gap between the information-processing and dynamical systems literature has proven scientifically fruitful. Examples of this are Wolfram’s categorization of cellular automata in terms of their attractors, which provided insights into the possible types of distributed computation enabled by these systems according to dynamical properties of their trajectories,13 and Langton’s intuition that computation takes place in a phase transition “at the edge of chaos.”14 This rich point of contact, in turn, suggests that what informational and dynamical approaches deem as interesting might have a common source, beyond the apparent dissimilarities introduced by heterogeneous tools and disciplinary boundaries.

In this article, we propose Integrated Information Theory (IIT),15–17 together with its recent extension Integrated Information Decomposition (ΦID),18 as a conceptual framework that can help bridge the gap between information-processing and dynamical systems approaches. At its inception, IIT was a theoretical effort that attempts to explain the origins and fundamental nature of consciousness.19,20 The boldness of IIT’s claims has not gone unnoticed, and they have caused a heated debate in the neuroscience community.21–23 Unfortunately, its audacious claims about consciousness have kept many scientists away from IIT, thereby preventing some of its valuable theoretical insights from reaching other areas of knowledge.

We advocate for the adoption of a pragmatic IIT, which can be used to analyze and understand complex systems without the philosophical burden of its origins in consciousness science. Consequently, the goal of this paper is to dissociate IIT’s claims as a theory of consciousness from its formal contributions and put the latter to use in the context of complexity science. For this purpose, we demonstrate that integrated information—as calculated in ΦID—peaks sharply for oscillatory dynamical systems that exhibit criticality and metastability and also characterizes cellular automata that display distributed computation via persistent emergent structures. These findings illustrate the remarkable flexibility of integrated information measures in discovering features of interest in a wide range of scenarios, without relying on domain-specific considerations. Overall, this work reveals how a grounded, demystified interpretation of IIT can allow us to identify features that are transversal across complex information-processing and dynamical systems.

The rest of this article is structured as follows. Section II presents the core formal ideas of IIT and ΦID in a simple manner and puts forward a sound and demystified way of interpreting its key quantities. Our main results are presented in Secs. III and IV: the former shows that integrated information can capture metastability and criticality in dynamical systems, and the latter that integrated information is a distinctive feature of distributed computation. Finally, Sec. V discusses the significance of these results and summarizes our main conclusions.

IIT constitutes one of the first attempts to formalize what makes a system “more than the sum of its parts,” building on intuitive notions of synergy and emergence that have been at the core of complexity science since its origins.24,25 IIT proposes integrated information for that role, informally defining it as information that is contained in the interactions between the parts of a system and not within the parts themselves. The core element of IIT is Φ, a scalar measure that accounts for the amount of integrated information present in a given system.

While faithful to its original aims, throughout its life, IIT has undergone multiple revisions. Of all of them, we will focus on the theory as introduced by Balduzzi and Tononi in 2008.16 While more recent accounts of the theory exist,17 these place a much stronger emphasis on its goals as a theory of consciousness, at the expense of a departure from standard information-theoretic practice and more convoluted algorithms—which have hindered its reach and made the theory applicable only in small discrete systems.

This section provides a succinct description of the mathematical formulas behind IIT 2.0,16 following Barrett and Seth’s26 concept of empirical integrated information. The overall analysis procedure is represented schematically in Fig. 1.

FIG. 1.

Graphical illustration of an integrated information analysis in complex systems. Integrated information theory (IIT) can be used to analyze a wide range of complex systems from dynamical systems to cellular automata and empirical data (left). Due to the generality of information-theoretic measures, it can be applied to either real- or discrete-valued data. After a suitable statistical model has been estimated, the system is partitioned following Eq. (2), and effective information φ is computed for each partition (middle). Finally, the partition with the “cruelest cut” (more formally, the minimum information partition) is selected, and the final value of integrated information Φ is computed (right).

FIG. 1.

Graphical illustration of an integrated information analysis in complex systems. Integrated information theory (IIT) can be used to analyze a wide range of complex systems from dynamical systems to cellular automata and empirical data (left). Due to the generality of information-theoretic measures, it can be applied to either real- or discrete-valued data. After a suitable statistical model has been estimated, the system is partitioned following Eq. (2), and effective information φ is computed for each partition (middle). Finally, the partition with the “cruelest cut” (more formally, the minimum information partition) is selected, and the final value of integrated information Φ is computed (right).

Close modal

The building block of integrated information is a measure of effective information (i.e., excess of predictive information) typically denoted as φ.16 Effective information quantifies how much better a system X is at predicting its own future after a time τ when it is considered a whole compared to when it is considered the sum of two subsystems M1 and M2 [so that X=(M1,M2)]. In other words, φ evaluates how much predictive information is generated by the system over and above the predictive information generated by the two subsystems alone. For a given bipartition B={M1,M2}, the effective information of the system X beyond B is calculated as

(1)

where I is Shannon’s mutual information. We refer to τ as the integration timescale.27 

The core idea behind the computation of Φ is to (i) exhaustively search all possible partitions of the system, (ii) calculate φ for each of them, and (iii) select the partition with lowest φ (under some considerations, see below), termed the minimum information bipartition (MIB). Then, the integrated information of the system is defined as the effective information beyond its MIB. Concretely, the integrated information Φ associated with the system X over the integration scale τ is given by

(2a)
(2b)
(2c)

where K is a normalization factor introduced to avoid biasing Φ to excessively unbalanced bipartitions. Defined this way, Φ can be understood as the minimum information loss incurred by considering the whole system as two separate subsystems.

An important drawback of Φ is that it can take negative values, which hinders its interpretation as a measure of system-wide integration.28 Recently, Mediano et al.18 showed that Φ quantifies not only information shared across and transferred between the parts, but also includes a negative component measuring redundancy—i.e., when the parts contain the same predictive information. Therefore, Φ measures a balance between information transfer and redundancy such that Φ<0 when the system is redundancy-dominated.

A principled way to address this limitation is to refine Φ by disentangling the different information phenomena that drive it. This can be achieved via the ΦID framework,18 which provides a taxonomy of “modes” of information dynamics in multivariate dynamical systems. Using ΦID, one can define a revised version of φ, denoted as φR,18 which removes the negative redundancy component in Φ by simply adding it back in. Work is ongoing on understanding different possible redundancy functions. Here, we compute φR via the minimum mutual information (MMI)29 redundancy function, which yields

(3)

Using φR, we can define ΦR analogously through Eq. (2).

Note that this revised measure of integrated information not only has better theoretical properties than the traditional φ, as discussed by Mediano et al.,18 but also has been observed to be superior in practical neuroimaging analyses.30 

Conceptually, there are several ways to interpret Φ, which highlight different aspects of the systems under study. One interpretation—particularly relevant for complexity science –is based on the theory of information dynamics, which decomposes information processing in complex systems in terms of storage, transfer, and modification.31–34 From this perspective and based on earlier results,18 ΦR can be seen as capturing a combination of information modification across multiple parts of the system, information transfer from one part to another, and storage in coherent structures that span across more than one system variable.

Alternatively, a more quantitative and mathematically rigorous way of interpreting ΦR is in terms of prediction bounds. The conditional entropy (a complement of the mutual information35) provides an upper bound on the optimal prediction performance36,37 such that a system with low conditional entropy can be predicted accurately. Therefore, mutual information acts as a bound too: the higher the mutual information between two variables, the better one can be predicted from the other. Thus, ΦR measures to what extent the full state of the system enables better predictions than the states of the parts separately.

Note that other interpretations of Φ-related quantities exist (most notably through information geometry and statistical inference38,39), although they do not apply as cleanly to Eqs. (1) and (3).

This section explores the usage of integrated information to study dynamical systems, exploring the relationship between Φ, metastability, and phase transitions. For this, we focus on systems of coupled oscillators, which are ubiquitous in both natural and engineered environments, making them of considerable scientific interest.2 Typical studies of oscillatory systems—such as the classic work of Kuramoto40—examine the conditions under which the system stabilizes on states of either full synchronization or desynchronization, although these two extremes are by no means representative of all real-world synchronization phenomena. Many systems of interest, including the human brain, exhibit synchronous rhythmic activity on multiple spatial and temporal scales but never settle into a stable state, entering so-called chimera-like states41 of high partial synchronization only temporarily. A system of coupled oscillators that continually moves between highly heterogeneous states of synchronization is said to be metastable.42 

In 2010, Shanahan42 showed that a modular network of phase-lagged Kuramoto oscillators can exhibit metastable chimera states. Variants of this model have since been used to replicate the statistics of the brain under a variety of conditions, including wakeful rest43 and anesthesia.44 In the following, we study these metastable oscillatory systems through the lens of integrated information theory.

We examine a community-structured network of coupled Kuramoto oscillators (shown in Fig. 2), building on the work of Shanahan.42 The network is composed of N communities of m oscillators each, with every oscillator being coupled to all other oscillators in its community and to each oscillator in the rest of the network with probability q. The state of the ith oscillator is determined by its phase θi, the evolution of which is governed by

(4)

where ω is the natural frequency of the oscillators, α is a global phase lag, Kij are the connectivity coefficients, and κ is a normalization constant. To reflect the community structure, the coupling between two oscillators i,j is defined as

(5)

with a>b. The system is tuned by modifying the value of the phase lag, parameterized by β=π/2α. We note that the system is fully deterministic; i.e., there is no noise injected in the dynamical equations.

FIG. 2.

A community-structured network of coupled oscillators exhibits metastable chimera-like behavior. (a) We study a network formed by 8 communities of 32 Kuramoto oscillators each, with strong intra-community coupling and weak inter-community coupling and dynamics parameterized by a phase lag β. (b) Time series of synchrony values Rc for each community (c=1,,8) show that the behavior of the system changes drastically with β. For low β, all communities remain desynchronized; for high β, all communities synchronize; and it is for intermediate β that the system behaves as a metastable chimera, with different communities entering intermittent periods of internal synchrony.

FIG. 2.

A community-structured network of coupled oscillators exhibits metastable chimera-like behavior. (a) We study a network formed by 8 communities of 32 Kuramoto oscillators each, with strong intra-community coupling and weak inter-community coupling and dynamics parameterized by a phase lag β. (b) Time series of synchrony values Rc for each community (c=1,,8) show that the behavior of the system changes drastically with β. For low β, all communities remain desynchronized; for high β, all communities synchronize; and it is for intermediate β that the system behaves as a metastable chimera, with different communities entering intermittent periods of internal synchrony.

Close modal

To assess the dynamical properties of the oscillators, we consider their instantaneous synchronizationR and metastability indexλ. The instantaneous synchronization at time t of community c{1,,N}, comprising oscillators Ic, quantifies their dispersion in the θ-space,

(6)

Building on this notion, the metastability of each community is defined as the variance of synchrony over time, and the overall metastability is its average across communities,

(7a)
(7b)

Communities that are either hypersynchronized or completely desynchronized are both characterized by small values of λc, whereas only communities whose elements fluctuate in and out of synchrony have a high λc. Put simply, a system of oscillators exhibits metastability to the extent that its elements fluctuate between states of high and low synchrony. In addition to metastability, we also consider the global synchrony of a network defined as the spatiotemporal average of the instantaneous synchrony,

(8)

For tractability, we calculate ΦR with respect to the coalition configuration of the system, defined for each community c and time t as

where γ is the coalition threshold. This representation provides N interdependent binary time series Xt:=(Xt1,,XtN), which indicates the set of communities that are more internally synchronized. The Shannon entropy of Xt is referred to as the coalition entropyHc and quantifies the diversity of synchronization patterns across communities.

We simulated a network composed of N=8 communities of m=32 oscillators each. The probability of connections across communities was set to q=1/8, with connection strengths of a=0.6 within communities and b=0.4 across. The natural frequency used was ω=1 and the normalization constant κ=64. We ran 1500 simulations with values of β distributed uniformly at random in the range [0,2π) using a 4th-order Runge–Kutta algorithm, using a step size of 0.05 for numerical integration. Each simulation was run for 5×106 time steps, discarding the first 104 to avoid transient effects and applying a thinning factor of 5. For the results presented here, we used γ=0.8, and we confirmed that results were qualitatively stable for a wide range of threshold values. All information-theoretic measures are reported in bits.

1. Metastability and ΦR at the phase transition

We first study the system from a purely dynamical perspective, and, replicating previous findings,42 we find two well differentiated dynamical regimes: one of hypersynchronization and one of complete desynchronization, with strong metastability appearing in the narrow transition bands between them (Fig. 3). Interestingly, it is in this transition region where the oscillators operate in a critical regime poised between order and disorder and where complex phenomena appear. As the system moves from desynchronization to full synchronization, there is a sharp increase in metastability, followed by a smoother decrease as the system becomes hypersynchronized.

FIG. 3.

Phase transition between desynchronization and hypersynchronization in coupled oscillators. Global synchrony ξ and metastability λ for different phase lags β in the critical region. Rapid increase of metastability marks the onset of the phase transition.

FIG. 3.

Phase transition between desynchronization and hypersynchronization in coupled oscillators. Global synchrony ξ and metastability λ for different phase lags β in the critical region. Rapid increase of metastability marks the onset of the phase transition.

Close modal

Importantly, ΦR was found to exhibit a similar behavior to λ: it is zero for desynchronized systems, peaks in the transition region, and shrinks again in the fully ordered regimes (Fig. 4). This shows that networks in metastable regimes are the only ones that exhibit integrated information. When comparing ΦR with the coalition entropy, results show both peaks at the same point, although the peak in ΦR is much narrower than the peaks in λ and Hc. Hence, while some values of β do give rise to non-trivial dynamics, it is only at the center of the critical region that these dynamics give rise to integrated information.

FIG. 4.

Integrated information ΦR peaks in the phase transition of coupled oscillators. Within the broad region between order and disorder in which Hc rises, there is a narrower band in which complex spatiotemporal patterns generate high ΦR. Calculations use τ=100.

FIG. 4.

Integrated information ΦR peaks in the phase transition of coupled oscillators. Within the broad region between order and disorder in which Hc rises, there is a narrower band in which complex spatiotemporal patterns generate high ΦR. Calculations use τ=100.

Close modal

These results imply that ΦR is sensitive to more subtle dynamic patterns than the other measures considered and is in that sense more discriminating. In effect, a certain degree of internal variability is necessary to establish integrated information, but not all configurations with high internal variability lead to a high ΦR. Also, ΦR accounts for spatial and temporal patterns in a way that the other metrics do not.45 

2. Integrated information at multiple timescales

As a further analysis, we can investigate the behavior of ΦR at multiple timescales by varying the integration timescale parameter τ (see Fig. 2 for a visual guide of different τ values). Figure 5 shows ΦR for several values of τ and compares it with standard time-delayed mutual information (TDMI) I(Xtτ;Xt). Note that this analysis cannot be carried out with Hc or other measures of complexity that are not sensitive to temporal order—i.e., that are functions of p(Xt) and not p(Xt|Xtτ).

FIG. 5.

Redundancy dominates at shorter timescales, synergy at longer ones. (a) Integrated information ΦR grows with increasing integration timescale τ, indicating a shift from redundancy- to synergy-dominated dynamics. (b) Time-delayed mutual information I(Xtτ;Xt) decreases with higher τ, indicating an overall loss of predictability for longer time horizons.

FIG. 5.

Redundancy dominates at shorter timescales, synergy at longer ones. (a) Integrated information ΦR grows with increasing integration timescale τ, indicating a shift from redundancy- to synergy-dominated dynamics. (b) Time-delayed mutual information I(Xtτ;Xt) decreases with higher τ, indicating an overall loss of predictability for longer time horizons.

Close modal

Results show that ΦR and TDMI exhibit opposite trends with respect to changes in τ: TDMI decreases with higher τ, while ΦR increases. At short timescales, the system is highly predictable—thus the high TDMI—but this short-term evolution does not imply much system-wide interaction—thus the low ΦR. Together, the high TDMI and low ΦR suggest that at short timescales, the system is redundancy-dominated: the system contains information about its future, but this information can be obtained from the parts separately. Conversely, for prediction at longer timescales, TDMI decreases but ΦR increases, indicating that while the system is overall less predictable, this prediction is enabled by the information contained in the interaction between the parts.

3. Robustness of ΦR against measurement noise

Finally, we study the impact of measurement noise on ΦR, wherein the system runs unchanged, but our recording of it is imperfect. For this, we run the (deterministic) simulation as before and generate the sequence of coalition configurations and then emulate the effect of uncorrelated measurement noise by flipping each bit in the time series with probability p, yielding a corrupted time series X^t. Finally, ΦR is recalculated on the corrupted time series (Fig. 6). To quantify the impact of noise, we studied the ratio between the corrupted and the original time series,

(9)

To avoid instabilities as ΦR[X;τ]0, we calculate η only in the region within 0.05 rad of the center of the peak shown in Fig. 4, where ΦR[X;τ] is large. The inset of Fig. 6 shows the mean and standard deviation of η at different noise levels.

FIG. 6.

Integrated information is highly sensitive to measurement noise. Integrated information ΦR for different levels of measurement noise p. The inset shows the mean and variance of the ratio η between ΦR of the corrupted and the original time series (blue) and an exponential fit η=exp(p/), with 0.04 (red).

FIG. 6.

Integrated information is highly sensitive to measurement noise. Integrated information ΦR for different levels of measurement noise p. The inset shows the mean and variance of the ratio η between ΦR of the corrupted and the original time series (blue) and an exponential fit η=exp(p/), with 0.04 (red).

Close modal

Results show that ΦR decays exponentially with p (Fig. 6, upper panel), reflecting a gradual loss of the precise spatiotemporal patterns that are characteristic of the system. In particular, ΦR was found to be highly sensitive to noise and to undergo a rapid decline, as a measurement noise of 5% can wipe out 70% of the observed integrated information of the system. While the distortion has a stronger effect on time series with greater ΦR, it preserves the dominant peak for all values of p.

Overall, in this section, we have shown that a network of Kuramoto oscillators presents a sharp, clear peak of integrated information around its phase transition that coincides with a strong increase in metastability. Furthermore, we have found that ΦR is informative, as it can reveal information about timescales of interaction between system components. Finally, ΦR was also found to be sensitive, as it vanishes quickly if the specific spatiotemporal patterns of the system under study are disrupted. This, in turn, suggests that it is highly unlikely to observe significant values of ΦR due to artifacts induced by (uncorrelated observational) noise.

In Sec. III, we related integrated information to dynamical complexity by linking ΦR with criticality and metastability in coupled oscillators. We now move on to cellular automata (CA), a well-known class of systems widely used in the study of distributed computation.31,46 Our aim here is to relate IIT to distributed computation in two ways: at a global scale, ΦR is higher for complex, class IV13 automata and at a local scale ΦR is higher for emergent coherent structures, such as blinkers, gliders, and collisions.

A CA is a multi-agent system in which every agent has a finite set of possible states and evolves in discrete time steps following a set of simple rules based on its own and other agents’ states. CA have been often used in studies of self-organization,1,47 and some of them are capable of universal computation.13 In a CA, agents (or cells) are arranged in a one-dimensional cyclic array (or tape). The state of each cell at a given time step has a finite number of possible states, which is determined via a boolean function (or rule), which uses as arguments the state of itself and its immediate neighbors at the previous time step. The same boolean function dictates the evolution of all agents in the system, inducing a spatial translational symmetry. Each CA, irrespective of its number of agents, is usually denoted by its rule.1 

For all the results presented below, we follow the simulation parameters used by Lizier in his study of local information dynamics in CA:31 we initialize a tape of length 104 with i.i.d. random variables, discard the first 100 steps of simulation, and run 600 more steps that are used to estimate the probability distributions used in all information-theoretic measures.

Our first analysis focuses on elementary cellular automata (ECA), a specific subclass of CA. In ECA, each cell has two possible states (usually denoted as white or black). ECA are traditionally denoted by their rule number, between 0 and 255, and grouped in four complexity classes:13 Class I rules have attractors consisting of single absorbing states; Class II rules evolve toward periodic orbits with relatively short cycle lengths; and Class III and IV rules have attractors with length of the order of the size of their phase space, with the latter being characterized by the presence of highly structured patterns and persistent structures.

As a first experiment, we calculate the average integrated information of each ECA, separating each automaton by complexity class (Fig. 7). For this, we followed the classifications defined in Wolfram’s original article13 as well as other clear-cut rules and excluded border cases, which did not neatly fit into a single category.

FIG. 7.

Integrated information grows monotonically with Wolfram class number. (a) Examples of each complexity class (ECA rules 32, 56, 75, and 54, respectively), showing noticeable differences in behavior. Notice the presence of localized particles in the class IV rule. (b) Correspondingly, ΦR is the highest for the more complex classes IV and III and lower (and often negative) for the simpler behaviors in classes I and II (error bars correspond to standard deviation across rules; each rule was simulated multiple times to obtain an accurate estimation of ΦR).

FIG. 7.

Integrated information grows monotonically with Wolfram class number. (a) Examples of each complexity class (ECA rules 32, 56, 75, and 54, respectively), showing noticeable differences in behavior. Notice the presence of localized particles in the class IV rule. (b) Correspondingly, ΦR is the highest for the more complex classes IV and III and lower (and often negative) for the simpler behaviors in classes I and II (error bars correspond to standard deviation across rules; each rule was simulated multiple times to obtain an accurate estimation of ΦR).

Close modal

Results show that ΦR correlates strongly with complexity as discussed by Wolfram: automata of higher classes have consistently higher ΦR than automata of lower classes, and the difference between classes I, II and III, IV is stark.

It is worth noting the small difference between classes III and IV. This is likely related to the blurriness of the line separating both classes—visually, it is hard to judge whether structures are “coherent enough” to support distributed computation, and formally, the problem of determining whether a particular rule belongs to class III or IV is considered undecidable.48,49 Based on this, we may tentatively suggest that the capacity to integrate information is a necessary, but not sufficient, condition for universal computation.

In his seminal 1990 article, Langton14 took a step beyond Wolfram’s classification and argued that the complexity and universality observed in ECA may reflect a broader phenomenon called computation at the edge of chaos. In this view, computation is made possible by indefinitely long transient states, a manifestation of critical slowing-down,50 that form the particle-like structures seen in class IV rules.

Langton’s argument starts by defining a parameter λ, which represents the fraction of neighborhoods in a CA’s rule table that map to a non-quiescent state (i.e., a non-white color). Then, by initializing one automaton with an empty rule table and progressively filling it with non-quiescent states, one can observe a transition point with exponentially long, particle-like transients [Fig. 8(a)]. Here, we repeat Langton’s experiments using a 6-color, range-2 CA, and compute its average Φ as its rule table gets populated and λ increases.

FIG. 8.

Integrated information peaks at the edge of chaos. (a) Sample runs from a random 6-color, range-2 cellular automaton with different λ values, starting from a blank tape with 20 randomized cells in the middle. The parameter λ corresponds to the fraction of non-blank cells in an automaton’s update rule. (b) Integrated information ΦR peaks at an intermediate level of λ. (c) When plotted against Δλ, the distance from a transition event, all runs align on a similar ΦR profile.

FIG. 8.

Integrated information peaks at the edge of chaos. (a) Sample runs from a random 6-color, range-2 cellular automaton with different λ values, starting from a blank tape with 20 randomized cells in the middle. The parameter λ corresponds to the fraction of non-blank cells in an automaton’s update rule. (b) Integrated information ΦR peaks at an intermediate level of λ. (c) When plotted against Δλ, the distance from a transition event, all runs align on a similar ΦR profile.

Close modal

In agreement with Langton’s argument, we found that integrated information has largest values for intermediate values of λ, coinciding with the automata’s transition to a chaotic regime [Fig. 8(b)]. Interestingly, this shows that rules with high ΦR are the ones at the critical region—where computation is possible.

Another unusual feature of Fig. 8(b) is that there is a region where complex, high-ΦR automata coexist with simpler ones. This phenomenon was reported by Langton14 already: different automata experience a “transition to chaos” at different values of λ. This motivated a further analysis of measures of complexity as a function of Δλ, the distance from the transition event for that particular automaton. As expected, when aligned by their critical λ value and plotted against Δλ [Fig. 8(c)], all curves align onto a consistent, seemingly universal, picture of integrated information across the λ range.

For completeness, it is worth mentioning why at the right side of Figs. 8(b) and 8(c), ΦR does not vanish for high λ (as one could expect, given that the single-cell autocorrelation does14). This is essentially due to the determinism and locality of the automaton’s rule: given a spatially extended set of cells, it is always possible to predict the middle ones with perfect certainty. At the same time, cutting the system with a bipartition will reduce the spatial extent of this predictable region so that the predictability of the whole is greater than the predictability of the parts, and thus, Φ>0.

In the experiments above, we have shown that more complex automata integrate more information. However, this is not enough to make a case for ΦR as a marker of distributed computation—it may just be the case that medium-λ CA have higher ΦR due to general properties of their rule tables or for some other reasons. In this section, we address this possible counter-argument by showing that the increase in ΦR is due to the emerging particles and, therefore, can be directly associated with distributed computation.

To show this, we run large simulations of ECA rules 54 and 110 and evaluate several local information measures in small fragments of both (Fig. 9). Specifically, we compute Lizier’s measures of storage (excess entropy, ek) and transfer (transfer entropy, tk), as well as local integrated information, using the binary time series of local neighborhoods of ECA cells as input (details in the  Appendix). Note that this pointwise Φ has not been ΦID-revised, as the development of pointwise ΦID metrics is still an open problem.51 A dedicated treatment of pointwise ΦID is part of our ongoing work and will be presented in a separate publication.

FIG. 9.

Local integrated information detects coherent structures in cellular automata. Local information measures in cellular automata, applied to simulations of rule 54 (top) and 110 (bottom). Local excess entropy is high for static particles, local transfer entropy for moving particles, and local integrated information Φ for both. For all these measures, blue and red indicate positive and negative values, respectively.

FIG. 9.

Local integrated information detects coherent structures in cellular automata. Local information measures in cellular automata, applied to simulations of rule 54 (top) and 110 (bottom). Local excess entropy is high for static particles, local transfer entropy for moving particles, and local integrated information Φ for both. For all these measures, blue and red indicate positive and negative values, respectively.

Close modal

As expected, TE is high in gliders (moving particles), while excess entropy is high in blinkers (static particles), confirming Lizier’s results that these structures perform information transfer and storage in CA.31 More interestingly for our purposes is that Φ is high in all of them—gliders, blinkers, and the collisions between them.

When studied at a local scale in space and time, we see that information integration encompasses the three categories—storage, transfer, and modification—and that Φ can detect all of them without having been explicitly designed to do so. This reinforces our claim that Φ (and, in turn, its revised version ΦR) is a generic marker of emergent dynamics and is connected with known measures of information processing. This relationship can be understood in a more mathematically rigorous manner via ΦID,18 which shows that Φ is decomposable in terms of information transfer and synergy.

Many theoretical efforts to understand complexity have their roots in information-theoretic or dynamical systems perspectives. On the one hand, information-theoretic approaches focus on discrete systems displaying coordinated activity capable of universal computation, e.g., via particles in Wolfram’s class IV automata.13 On the other hand, dynamical approaches focus on continuous coupled dynamical systems near a phase transition, displaying a stream of transient states of partial coherence that balance robustness and adaptability.52,53 The results presented in this paper reveal how integrated information is effective at characterizing complexity in both information-processing and dynamical systems and hence can be regarded as a common signature of complex behavior across a wide range of systems. Overall, this paper shows how a grounded, demystified understanding of IIT combined with the improved metrics provided by ΦID can be used as a first step toward a much needed point of contact between these diverging branches of complexity science, helping to bridge the gap between information-processing and dynamical approaches.

It is important to remark that the relationship between integrated information, metastability, criticality, and distributed computation is not an identity and that their agreement is an important finding conveyed by Φ.

Metastability, in the case of oscillator networks, is a community-local quantity—that is, it corresponds to an average over quantities (λc) that depend on the temporal diversity seen within each community, independently of the rest. In stark contrast, Φ relies on the irreducible interactions between communities. Interestingly, our finding reveals a close relationship between the two, insofar as internal variability enables the system to visit a larger repertoire of states in which system-wide interaction can take place.

Criticality is, in general, enabled by a precise balance of two opposite forces (typically characterized as order and chaos in physics) that enables peculiar and fascinating phenomena, such as scale-freeness, extreme sensitivity to perturbations, and universality classes.54,55 In contrast, one of the core ideas in IIT is that integration and differentiation are not opposite forces but can actually coexist together.56,57 Therefore, integrated information is not maximized by optimal trade-offs, but by mechanisms that incentivize direct and synergistic information transfer.18 

Finally, distributed computation is mainly based on intuitive but ultimately informal notions developed after Wolfram’s and Langton’s work. Our results establish a strong relationship between the capability of information-processing complex systems to integrate information and their ability to perform highly non-trivial information processing via emergent coherent structures, such as blinkers, gliders, and collisions. The fact that higher integrated information was carried by coherent emerging structures is consistent with recent accounts of causally emergent dynamics,58 which further supports the case of Φ being an effective quantitative indicator of distributed emergent computation with numerous potential future applications.

The conceptual link between complexity and integrated information is by no means a novel idea: in fact, in the early days of IIT, integrated information and complexity were closely intertwined59—as was its close relation with the theory of complex networks.60 Unfortunately, as the theory turned more convoluted and less applicable, this link lost relevance and IIT drifted away from standard information-theoretic methods.

Nonetheless, recent research has again brought complexity science to the fore within the IIT community. In some cases, information-theoretic measures originally conceived as measures of complexity have been re-purposed within IIT.61 In others, new measures of complexity are inspired by previous work on IIT.39 In contrast with its consciousness-focused sibling, advancements in this more pragmatic IIT (cf. Sec II) have been enabled by a simplification and unification of the underlying principles of the theory—for example, in terms of information geometry62 or information decomposition.18 

Given its origins in theoretical neuroscience, it is no surprise that most of IIT’s relevance to complex systems has been mostly on models of neural dynamics. In this context, the combination of integration and segregation captured by Φ has been linked to other dynamical features of interest (such as metastability and neural avalanches) in a variety of settings, including, e.g., models of whole-brain activity63 and spiking neural networks.64 

Relatedly, there has been recent work linking Φ with phase transitions in thermodynamic systems.65 Together with recent results linking information processing and stochastic thermodynamics,66 this exciting avenue of research opens the door for a purely physical interpretation of Φ in general physical systems.

Finally, note that for our main analyses, we estimate ΦR using the simple yet effective MMI redundancy function.29 However, there are multiple possible redundancy functions,67,68,74 and while they tend to yield qualitatively similar results in practice,30,69 they have been shown to differ in some specific scenarios,68 and therefore, it is important to elucidate their similarities and differences in practical analyses. There is ongoing investigation into the different properties of each redundancy function, and future work will explore the specific advantages of diverse redundancy functions when used to calculate ΦR.

This paper puts forward a pragmatic argument that metrics of integrated information—such as the ones provided by IIT and ΦID—can allow us to investigate unexplored commonalities between informational and dynamical complexity, pointing out a promising avenue to reconcile their divide and benefit subsidiary disciplines, such as cognitive and computational neuroscience. Note that Φ is by no means the only quantity that peaks with the system’s complexity—in cellular automata, one could use the autocorrelation of a single cell and in coupled oscillators the variance of Kuramoto’s order parameter. However, the feature that makes Φ unique is that it is applicable across the board and yields the desired results in different kinds of systems without requiring idiosyncratic, ad hoc measures. This unification, analogous to the transversal role that Fisher information plays in phase transitions over arbitrary order parameters,70 posits integrated information as a key quantity for the study of complex dynamics, whose relevance has just started to be uncovered.

P.A.M.M. and D.B. are funded by the Wellcome Trust (Grant No. 210920/Z/18/Z). F.E.R. is supported by the Ad Astra Chandaria Foundation. A.B.B. is grateful to the Dr. Mortimer and Theresa Sackler Foundation, which supports the Sackler Centre for Consciousness Science.

The authors declare no conflict of interest.

P.A.M.M. and F.E.R. contributed equally to this work.

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

A natural way to extend IIT 2.0 for complex systems analysis is to consider local versions of Φ, which can be built via the framework introduced by Lizier.32 Local (or pointwise) information measures are able to identify coherent, emergent structures known as particles, which have been shown to be the basis of the distributed information processing that takes place in systems such as cellular automata.31,33,34

One of the most basic pointwise information metrics is the local mutual information, which is defined as

(A1)

so that E[i(X;Y)]=I(X;Y) is the usual mutual information. By evaluating i on every x,y pair, one can determine which particular combinations of symbols play a predominant role for the observed interdependency between X and Y. (More specifically, the local mutual information captures specific deviations between the joint distribution and the product of the marginals.) Building on these ideas, Lizier proposed a taxonomy of distributed information processing as composed of storage, transfer, and modification.31 For this, consider a bivariate stochastic process (Xt,Yt) with tZ and introduce the shorthand notation Xt(k)=(Xtk,,Xt) and Xt(k+)=(Xt,,Xt+k1) for the corresponding past and future embedding vectors of length k. In this context, storage within the subprocess Xt is identified with its excess entropyEk=I(Xt(k);Xt+1(k+))71 and transfer from Yt to Xt+1 with the transfer entropyTEk=I(Xt+1;Xt(k)|Yt(k)).72 Interestingly, both quantities have corresponding local versions,

(A2)
(A3)

such that, as expected, E[ek]=Ek and E[tk]=TEk. Note that to measure transfer in either direction for the results in Fig. 9, we compute the local TE from a cell to its left and right neighbors and take the maximum of the two.

These ideas can be used to extend the standard formulation of integrated information measures in two ways. First, by using embedding vectors, the IIT metrics are applicable to non-Markovian systems.73 Second, by formulating pointwise measures, one can capture spatiotemporal variations in Φ. Mathematically, we reformulate Eq. (2) introducing these modifications as

(A4)

and apply the same partition scheme described in Sec. II A to obtain an “embedded” integrated information, Φk. Then, the equation above can be readily made into a local measure by replacing mutual information with its local counterpart,

(A5)

such that, as expected, φk[X;τ,B]=E[ϕk[xt;τ,B]].

1.
S.
Wolfram
,
A New Kind of Science
(
Wolfram Media
,
2002
), p.
1197
.
2.
A.
Pikovsky
,
M.
Rosenblum
, and
J.
Kurths
,
Synchronization: A Universal Concept in Nonlinear Sciences
(
Cambridge University Press
,
2001
), p.
432
.
3.
J. A.
Fodor
,
The Language of Thought
(
Harvard University Press
,
1975
), Vol. 5.
4.
Z.
Pylyshyn
and
W.
Turnbiull
, “
Computation and cognition: Toward a foundation for cognitive science
,”
Can. Psychol.
27
,
85
87
(
1986
).
5.
M.
Rescorla
, “From Ockham to Turing—And back again,” in Philosophical Explorations of the Legacy of Alan Turing (Springer, 2017), pp. 279–304.
6.
M.
Milkowski
,
Explaining the Computational Mind
(
MIT Press
,
2013
).
7.
T.
Van Gelder
, “
What might cognition be, if not computation?
,”
J. Philos.
92
,
345
381
(
1995
).
8.
T.
Van Gelder
, “
The dynamical hypothesis in cognitive science
,”
Behav. Brain Sci.
21
,
615
628
(
1998
).
9.
L. B.
Smith
, “
Cognition as a dynamic system: Principles from embodiment
,”
Dev. Rev.
25
,
278
298
(
2005
).
10.
M. L.
Minsky
, “
Logical versus analogical or symbolic versus connectionist or neat versus scruffy
,”
AI Mag.
12
,
34
51
(
1991
).
11.
M.
Garnelo
and
M.
Shanahan
, “
Reconciling deep learning with symbolic artificial intelligence: Representing objects and relations
,”
Curr. Opin. Behav. Sci.
29
,
17
23
(
2019
).
12.
R. D.
Beer
and
P. L.
Williams
, “
Information processing and dynamics in minimally cognitive agents
,”
Cogn. Sci.
39
,
1
38
(
2015
).
13.
S.
Wolfram
, “
Universality and complexity in cellular automata
,”
Physica D
10
,
1
35
(
1984
).
14.
C. G.
Langton
, “
Computation at the edge of chaos: Phase transitions and emergent computation
,”
Physica D
42
,
12
37
(
1990
).
15.
G.
Tononi
and
O.
Sporns
, “
Measuring information integration
,”
BMC Neurosci.
4
,
31
(
2003
).
16.
D.
Balduzzi
and
G.
Tononi
, “
Integrated information in discrete dynamical dystems: Motivation and theoretical framework
,”
PLoS Comput. Biol.
4
,
e1000091
(
2008
).
17.
M.
Oizumi
,
L.
Albantakis
, and
G.
Tononi
, “
From the phenomenology to the mechanisms of consciousness: Integrated Information Theory 3.0
,”
PLoS Comput. Biol.
10
,
e1003588
(
2014
).
18.
P.
Mediano
,
F.
Rosas
,
R. L.
Carhart-Harris
,
A. K.
Seth
, and
A. B.
Barrett
, “Beyond integrated information: A taxonomy of information dynamics phenomena,” arXiv:1909.02297 (2019).
19.
G.
Tononi
and
C.
Koch
, “
Consciousness: Here, there and everywhere?
,”
Philos. Trans. R. Soc. B
370
,
20140167
(
2015
).
20.
G.
Tononi
,
M.
Boly
,
M.
Massimini
, and
C.
Koch
, “
Integrated information theory: From consciousness to its physical substrate
,”
Nat. Rev. Neurosci.
17
,
450
(
2016
).
21.
A. B.
Barrett
and
P.
Mediano
, “
The Phi measure of integrated information is not well-defined for general physical systems
,”
J. Conscious. Stud.
26
,
11
20
(
2019
).
22.
M. A.
Cerullo
, “
The problem with Phi: A critique of integrated information theory
,”
PLoS Comput. Biol.
11
,
e1004286
(
2015
).
23.
G.
Mindt
, “
The problem with the ‘information’ in integrated information theory
,”
J. Conscious. Stud.
24
,
130
154
(
2017
).
24.
P. W.
Anderson
, “
More is different
,”
Science
177
,
393
396
(
1972
).
25.
M. M.
Waldrop
,
Complexity: The Emerging Science at the Edge of Order and Chaos
(
Simon and Schuster
,
1993
).
26.
A. B.
Barrett
and
A. K.
Seth
, “
Practical measures of integrated information for time-series data
,”
PLoS Comput. Biol.
7
,
e1001052
(
2011
).
27.
Note that the quantites in Eqs. (1)–(3) rely only on pairwise temporal correlations. This is sufficient to capture a statistical structure in Markovian systems, but not in non-Markovian ones, which can have zero pairwise correlation even in the presence of complex dynamics.73 A possible measure of integrated information for non-Markovian systems is described in the  Appendix and applied in Sec. IV C.
28.
P.
Mediano
,
A. K.
Seth
, and
A. B.
Barrett
, “
Measuring integrated information: Comparison of candidate measures in theory and simulation
,”
Entropy
21
,
17
(
2019
).
29.
A. B.
Barrett
, “
Exploration of synergistic and redundant information sharing in static and dynamical Gaussian systems
,”
Phys. Rev. E
91
,
052802
(
2015
).
30.
A. I.
Luppi
,
P. A.
Mediano
,
F. E.
Rosas
,
J.
Allanson
,
J. D.
Pickard
,
R. L.
Carhart-Harris
,
G. B.
Williams
,
M. M.
Craig
,
P.
Finoia
,
A. M.
Owen
et al., “A synergistic workspace for human consciousness revealed by integrated information decomposition,” bioRxiv (2020).
31.
J. T.
Lizier
, “The local information dynamics of distributed computation in complex systems,” Springer theses (Springer, Berlin, 2010).
32.
J. T.
Lizier
,
M.
Prokopenko
, and
A. Y.
Zomaya
, “Information transfer by particles in cellular automata,” in Australian Conference on Artificial Life (Springer, 2007), pp. 49–60.
33.
J. T.
Lizier
,
M.
Prokopenko
, and
A. Y.
Zomaya
, “
Local measures of information storage in complex distributed computation
,”
Inf. Sci.
208
,
39
54
(
2012
).
34.
J. T.
Lizier
,
M.
Prokopenko
, and
A. Y.
Zomaya
, “
Information modification and particle collisions in distributed computation
,”
Chaos
20
,
037109
(
2010
).
35.
R. G.
James
,
C. J.
Ellison
, and
J. P.
Crutchfield
, “
Anatomy of a bit: Information in a time series observation
,”
Chaos
21
,
037109
(
2011
).
36.
M.
Hellman
and
J.
Raviv
, “
Probability of error, equivocation, and the Chernoff bound
,”
IEEE Trans. Inf. Theory
16
,
368
372
(
1970
).
37.
M.
Feder
and
N.
Merhav
, “
Relations between entropy and error probability
,”
IEEE Trans. Inf. Theory
40
,
259
266
(
1994
).
38.
D.
Cohen
,
S.
Sasai
,
N.
Tsuchiya
, and
M.
Oizumi
, “
A general spectral decomposition of causal influences applied to integrated information
,”
J. Neurosci. Methods
330
,
108443
(
2020
).
39.
C.
Langer
and
N.
Ay
, “
Complexity as causal information integration
,”
Entropy
22
,
1107
(
2020
).
40.
Y.
Kuramoto
,
Chemical Oscillations, Waves and Turbulence
(
Dover Publications
,
1984
), p.
164
.
41.
M. J.
Panaggio
, and
D. M.
Abrams
, “
Chimera states: Coexistence of coherence and incoherence in networks of coupled oscillators
,”
Nonlinearity
28
,
67
87
(
2015
).
42.
M.
Shanahan
, “
Metastable chimera states in community-structured oscillator networks
,”
Chaos
20
,
013108
(
2010
).
43.
J.
Cabral
,
E.
Hugues
,
O.
Sporns
, and
G.
Deco
, “
Role of local network oscillations in resting-state functional connectivity
,”
NeuroImage
57
,
130
139
(
2011
).
44.
M.
Schartner
,
A. K.
Seth
,
Q.
Noirhomme
,
M.
Boly
,
M.-A.
Bruno
,
S.
Laureys
, and
A. B.
Barrett
, “
Complexity of multi-dimensional spontaneous EEG decreases during propofol induced general anaesthesia
,”
PLoS One
10
,
e0133532
(
2015
).
45.
This can be verified by performing a random time-shuffle on the time series, which leaves λ and Hc unaltered as they do not explicitly depend on time correlations but makes ΦR shrink to zero.
46.
M.
Mitchell
, “Computation in cellular automata: A selected review,” in Non-Standard Computation, edited by T. Gramß, S. Bornholdt, M. Groß, M. Mitchell, and T. Pellizzari (Wiley-VCH, 1996), pp. 95–140.
47.
F.
Rosas
,
P. A.
Mediano
,
M.
Ugarte
, and
H. J.
Jensen
, “
An information-theoretic approach to self-organisation: Emergence of complex interdependencies in coupled dynamical systems
,”
Entropy
20
,
793
(
2018
).
48.
K.
Culik II
and
S.
Yu
, “
Undecidability of CA classification schemes
,”
Complex Syst.
2
,
177
190
(
1988
).
49.
G. J.
Martinez
,
J. C.
Seck-Tuoh-Mora
, and
H.
Zenil
, “
Computation and universality: Class IV versus class III cellular automata
,”
J. Cell. Autom.
7
,
393
430
(
2013
).
50.
M.
Scheffer
,
J.
Bascompte
,
W. A.
Brock
,
V.
Brovkin
,
S. R.
Carpenter
,
V.
Dakos
,
H.
Held
,
E. H.
Van Nes
,
M.
Rietkerk
, and
G.
Sugihara
, “
Early-warning signals for critical transitions
,”
Nature
461
,
53
(
2009
).
51.
Indeed, although some pointwise redundancy functions exist,74 the relatively much simpler redundancy in Eq. (3) is not naturally translated into a pointwise setting, and thus, we restrict the pointwise analyses to standard integrated information only.
52.
J. S.
Kelso
,
Dynamic Patterns: The Self-organization of Brain and Behavior
(
MIT Press
,
1995
).
53.
M. I.
Rabinovich
,
R.
Huerta
,
P.
Varona
, and
V. S.
Afraimovich
, “
Transient cognitive dynamics, metastability, and decision making
,”
PLoS Comput. Biol.
4
,
e1000072
(
2008
).
54.
J. M.
Beggs
and
N.
Timme
, “
Being critical of criticality in the brain
,”
Front. Physiol.
3
,
163
(
2012
).
55.
G.
Pruessner
,
Self-Organised Criticality: Theory, Models and Characterisation
(
Cambridge University Press
,
2012
).
56.
G.
Tononi
,
O.
Sporns
, and
G. M.
Edelman
, “
A measure for brain complexity: Relating functional segregation and integration in the nervous system
,”
Proc. Natl. Acad. Sci. U.S.A.
91
,
5033
5037
(
1994
).
57.
F. E.
Rosas
,
P. A.
Mediano
,
M.
Gastpar
, and
H. J.
Jensen
, “
Quantifying high-order interdependencies via multivariate extensions of the mutual information
,”
Phys. Rev. E
100
,
032305
(
2019
).
58.
F. E.
Rosas
,
P. A.
Mediano
,
H. J.
Jensen
,
A. K.
Seth
,
A. B.
Barrett
,
R. L.
Carhart-Harris
, and
D.
Bor
, “
Reconciling emergences: An information-theoretic approach to identify causal emergence in multivariate data
,”
PLoS Comput. Biol.
16
,
e1008289
(
2020
).
59.
G.
Tononi
,
G. M.
Edelman
, and
O.
Sporns
, “
Complexity and coherency: Integrating information in the brain
,”
Trends Cogn. Sci.
2
,
474
484
(
1998
).
60.
O.
Sporns
,
G.
Tononi
, and
G. M.
Edelman
, “
Connectivity and complexity: The relationship between neuroanatomy and brain dynamics
,”
Neural Netw.
13
,
909
922
(
2000
).
61.
N.
Ay
, “
Information geometry on complexity and stochastic interaction
,”
Entropy
17
,
2432
2458
(
2015
).
62.
M.
Oizumi
,
N.
Tsuchiya
, and
S.-I.
Amari
, “
Unified framework for information integration based on information geometry
,”
Proc. Natl. Acad. Sci. U.S.A.
113
,
14817
14822
(
2016
).
63.
E.
Tagliazucchi
, “
The signatures of conscious access and its phenomenology are consistent with large-scale brain communication at criticality
,”
Conscious. Cogn.
55
,
136
147
(
2017
).
64.
P.
Mediano
and
M.
Shanahan
, “Balanced information storage and transfer in modular spiking neural networks,” arXiv:1708.04392 (2017).
65.
M.
Aguilera
, “
Scaling behaviour and critical phase transitions in integrated information theory
,”
Entropy
21
,
1198
(
2019
).
66.
J. M.
Parrondo
,
J. M.
Horowitz
, and
T.
Sagawa
, “
Thermodynamics of information
,”
Nat. Phys.
11
,
131
139
(
2015
).
67.
R. A. A.
Ince
, “
Measuring multivariate redundant information with pointwise common change in surprisal
,”
Entropy
19
,
318
(
2017
).
68.
R.
James
,
J.
Emenheiser
, and
J.
Crutchfield
, “
Unique information via dependency constraints
,”
J. Phys. A: Math. Theor.
52
,
014002
(
2018
).
69.
T. M.
Tax
,
P. A. M.
Mediano
, and
M.
Shanahan
, “
The partial information decomposition of generative neural network models
,”
Entropy
19
,
474
(
2017
).
70.
M.
Prokopenko
,
J. T.
Lizier
,
O.
Obst
, and
X. R.
Wang
, “
Relating Fisher information to order parameters
,”
Phys. Rev. E
84
,
041116
(
2011
).
71.
J. P.
Crutchfield
and
D. P.
Feldman
, “
Regularities unseen, randomness observed: Levels of entropy convergence
,”
Chaos
13
,
25
54
(
2003
).
72.
T.
Schreiber
, “
Measuring information transfer
,”
Phys. Rev. Lett.
85
,
461
464
(
2000
).
73.
P. M.
Riechers
and
J. P.
Crutchfield
, “
Fraudulent white noise: Flat power spectra belie arbitrarily complex processes
,”
Phys. Rev. Res.
3
,
013170
(
2021
).
74.
C.
Finn
and
J. T.
Lizier
, “
Pointwise partial information decomposition using the specificity and ambiguity lattices
,”
Entropy
20
,
297
(
2018
).