Skip to content

This framework asks: How many dimensions carry meaningful variance — is the representation concentrated or distributed?

The participation ratio (PR) is a single scalar that summarizes spectral concentration. Unlike threshold-based measures (e.g., “dimensions for 90% variance”), PR provides a smooth, threshold-free estimate of effective dimensionality. Originally from condensed matter physics (localization of wavefunctions), PR elegantly captures whether a representation is dominated by a few directions (low PR) or spreads evenly across many (high PR).

For circuit analysis, PR distinguishes heads that perform rank-1-like operations (low PR, interpretable as single-feature detectors) from heads that implement genuinely high-dimensional transformations (high PR, likely performing complex combinatorial operations).

SourceYearKey contribution
Bell & Dean, “Atomic vibrations in vitreous silica”1970Original participation ratio in physics
Gao et al., “A theory of multineuron dimensionality, dynamics and measurement”2017PR for neural population dimensionality
Litwin-Kumar et al., “Optimal Degrees of Synaptic Connectivity”2017PR in biological neural circuits
Recanatesi et al., “Dimensionality compression and expansion in deep neural networks”2019PR dynamics through network layers

Given the eigenvalue spectrum ( \lambda_1, \ldots, \lambda_d ) of the activation covariance matrix, the participation ratio is:

[ \text{PR} = \frac{\left(\sum_{i=1}^d \lambda_i\right)^2}{\sum_{i=1}^d \lambda_i^2} ]

PR ranges from 1 (all variance in one dimension) to ( d ) (uniform across all dimensions). It can be normalized:

[ \text{PR}_{\text{norm}} = \frac{\text{PR}}{d} \in [1/d, ; 1] ]

The inverse participation ratio (IPR = 1/PR) measures localization — how concentrated the variance is. For a rank-( k ) representation with equal eigenvalues, PR = ( k ) exactly. For exponentially decaying spectra, PR is dominated by the few largest eigenvalues.

Per-Head Participation Ratio (participation_ratio.py)

Section titled “Per-Head Participation Ratio (participation_ratio.py)”

Computes PR for each circuit head’s output activations across a corpus, reporting both raw PR and normalized PR.

What it establishes: The effective dimensionality of each head’s operation — low PR means rank-deficient, high PR means full-rank. What it does not establish: What the active dimensions encode (combine with E01/E02 for interpretation).

Usage:

uv run python participation_ratio.py --tasks ioi sva

Tracks how PR evolves across layers, revealing dimensionality compression/expansion dynamics.

What it establishes: Where the network compresses (decreasing PR) or expands (increasing PR) its representational capacity. What it does not establish: Whether compression is lossy or lossless.

Usage:

uv run python participation_ratio.py --tasks ioi sva --dynamics
PatternWhat it means
PR ~ 1-3Near rank-1; head implements a simple projection or feature detector
PR ~ 10-30 (for d=64)Moderate dimensionality; structured but non-trivial operation
PR ~ dFull-rank; head uses all available dimensions
PR decreases through layersProgressive compression toward a decision subspace
PR varies across heads at same layerFunctional specialization — some heads simple, others complex