L-PACT Framework Challenges Brain-Language Model Alignment Claims
A new study has introduced L-PACT, a framework designed to evaluate how well language models align with brain activity, arguing that just looking at prediction scores isn't enough. It examines various types of evidence, including predictive, relational, mechanism-stripping, and reliability-bounded evidence, using different neural datasets and model representations. The framework compares actual model features to nuisance baselines and rigorous controls, checks if model-to-brain patterns mirror brain-to-brain connections, and adjusts scores after mechanism stripping. The analysis includes 414 predictive-control rows, 2,304 relational profile rows, 4,320 mechanism-stripping rows, 420 brain-brain ceiling rows, and 146 integrated decisions. These results challenge the common belief that high prediction scores mean models are accurately reflecting brain-related language processing.
Key facts
- L-PACT is a source-audited framework for evaluating brain-language model alignment.
- The framework assesses predictive, relational, mechanism-stripping, and reliability-bounded evidence.
- Real model features are compared with nuisance baselines and severe controls.
- Model-to-brain profiles are tested against brain-to-brain patterns.
- Mechanism stripping is used to recompute held-out scores.
- Evidence is normalized against brain-brain ceilings.
- The analysis contains 414 predictive-control rows, 2304 relational profile rows, 4320 mechanism-stripping rows, 420 brain-brain ceiling rows, and 146 integrated decisions.
- The study argues prediction scores alone are not enough to claim alignment.
Entities
—