Skip to content

This framework asks: Does information flow directionally from one circuit component to another across the computation?

Transfer entropy (TE) quantifies the directed, time-asymmetric information flow from a source process to a target process. In transformer circuits, “time” corresponds to layer depth: information flows from earlier layers to later layers through the residual stream. TE measures how much knowing the activation of an earlier component reduces uncertainty about a later component, beyond what the later component’s own history provides.

This is crucial for circuit discovery because it distinguishes genuine information transmission from mere correlation. Two heads may be correlated because they both read from the same residual stream position, but TE identifies which one actually informs the other.

SourceYearKey contribution
Schreiber, “Measuring Information Transfer”2000Original transfer entropy definition
Cover & Thomas, Elements of Information Theory2006Directed information and causal conditioning
Barnett et al., “Granger Causality and Transfer Entropy Are Equivalent for Gaussian Variables”2009Equivalence to Granger causality under Gaussianity
Bossomaier et al., An Introduction to Transfer Entropy2016Comprehensive treatment with estimation methods

Transfer entropy from process ( X ) to process ( Y ) is defined as:

[ T_{X \to Y} = I(Y_t; X_{t-1} \mid Y_{t-1}) ]

In the circuit context, let ( X_\ell ) be a component’s output at layer ( \ell ) and ( Y_{\ell+k} ) a downstream component at layer ( \ell+k ). The transfer entropy becomes:

[ T_{X \to Y} = H(Y_{\ell+k} \mid Y_{\ell+k-1}) - H(Y_{\ell+k} \mid Y_{\ell+k-1}, X_\ell) ]

This is positive only if knowing ( X_\ell ) reduces uncertainty about ( Y_{\ell+k} ) beyond what the target’s own earlier state provides. Asymmetry (( T_{X \to Y} \neq T_{Y \to X} )) reveals directed information flow.

Observational causal sensitivity estimation captures directed influence by measuring how perturbations to one component’s activation propagate to downstream components — a finite-difference analogue of transfer entropy.

What it establishes: Directed information flow between circuit components across layers. What it does not establish: Whether the transferred information is task-relevant (high TE could reflect noise propagation).

Usage:

uv run python 07_ocse.py --tasks ioi sva

| Pattern | What it means | |---|---|---| | High ( T_{X \to Y} ), low ( T_{Y \to X} ) | Genuine directed flow from X to Y | | High TE in both directions | Shared input or confounded relationship | | TE peaks at specific layer gaps | Information is transmitted with characteristic depth | | Zero TE despite high MI | Components share info via a common cause, not direct transmission |

Transfer entropy is the information-theoretic analogue of C07 (Granger Causality) — they are equivalent for Gaussian processes. Where TE finds directed flow, C08 (OCSE) validates it via observational perturbation, and C09 (NOTEARS) attempts to recover the full DAG structure. The causal pillar then tests whether these directed flows are necessary via intervention.