Transformers' Graph Connectivity Learning Depends on Graph Structure
A new study investigates whether transformer-based Large Language Models (LLMs) can learn to infer transitive relations from training examples, a capability essential for reasoning tasks such as causal inference. The task is equivalent to determining connectivity in directed graphs: if A connects to B and B connects to C, then A connects to C. Prior work examined in-context learning of transitivity, but this research focuses on learning from training data and the impact of scaling. The findings reveal that transformers can learn connectivity in some graph structures but not others, highlighting limitations in their reasoning abilities.
Key facts
- The study is published on arXiv with ID 2509.22343.
- It examines transformer-based LLMs' ability to infer transitive relations.
- Transitive inference is equivalent to connectivity in directed graphs.
- Prior research focused on in-context examples, not training-based learning.
- The study explores how scaling affects this capability.
- Results show transformers succeed on some graph structures but fail on others.
Entities
Institutions
- arXiv