Part I: Foundations

Integration Measures

Introduction
0:00 / 0:00

Integration Measures

Let’s define precise measures of integration that will play a central role in the phenomenological analysis.

The first is transfer entropy, which captures directed causal influence between components. The transfer entropy from process XX to process YY measures the information that XX provides about the future of YY beyond what YY’s own past provides:

TEXY=I(Xt;Yt+1Y1:t)\text{TE}_{X \to Y} = \MI(X_t; Y_{t+1} | Y_{1:t})

The deepest measure is integrated information (Φ\Phi). Following IIT, the integrated information of a system in state s\state is the extent to which the system’s causal structure exceeds the sum of its parts:

Φ(s)=minpartitions PD[p(st+1st)pPp(st+1pstp)]\Phi(\state) = \min_{\text{partitions } P} D\left[ p(\state_{t+1} | \state_t) | \prod_{p \in P} p(\state^p_{t+1} | \state^p_t) \right]

where the minimum is over all bipartitions of the system, and DD is an appropriate divergence (typically Earth Mover’s distance in IIT 4.0).

In practice, computing Φ\Phi exactly is intractable. Three proxies make it operational:

  1. Transfer entropy density—average transfer entropy across all directed pairs:
    TEˉ=1n(n1)ijTEij\bar{\text{TE}} = \frac{1}{n(n-1)} \sum_{i \neq j} \text{TE}_{i \to j}
  2. Partition prediction loss—the cost of factoring the model:
    ΔP=Lpred[partitioned model]Lpred[full model]\Delta_P = \mathcal{L}_{\text{pred}}[\text{partitioned model}] - \mathcal{L}_{\text{pred}}[\text{full model}]
  3. Synergy—the information that components provide jointly beyond their individual contributions:
    Syn(X1,,XkY)=I(X1,,Xk;Y)iI(Xi;YXi)\text{Syn}(X_1, …, X_k \to Y) = \MI(X_1, …, X_k; Y) - \sum_i \MI(X_i; Y | X_{-i})

A complementary measure captures the system’s representational breadth rather than its causal coupling. The effective rank of a system with state covariance matrix CC measures how many dimensions it actually uses:

reff=(trC)2tr(C2)=(iλi)2iλi2\effrank = \frac{(\tr C)^2}{\tr(C^2)} = \frac{\left(\sum_i \lambda_i\right)^2}{\sum_i \lambda_i^2}

where λi\lambda_i are the eigenvalues of CC. This is bounded by 1reffrank(C)1 \leq \effrank \leq \rank(C), with reff=1\effrank = 1 when all variance is in one dimension (maximally concentrated) and reff=rank(C)\effrank = \rank(C) when variance is uniformly distributed across all active dimensions.