DiffMAS: Training Framework for Latent Communication in Multi-Agent LLM Systems
A new training framework called DiffMAS has been introduced by researchers, which integrates latent communication as a learnable aspect in multi-agent systems utilizing large language models. In contrast to traditional methods that depend on static text-based protocols, DiffMAS leverages internal representations like key-value caches for communication between agents. This framework enables parameter-efficient supervised training across multi-agent latent trajectories, facilitating the joint learning of information encoding and interpretation during interactions. Experiments conducted on mathematical reasoning, scientific question answering, code generation, and commonsense benchmarks demonstrate notable enhancements in reasoning accuracy and decoding. This study fills a void in current multi-agent LLM research, which mainly concentrates on agent roles and orchestration while treating communication as a fixed interface.
Key facts
- DiffMAS treats latent communication as a learnable component in multi-agent LLM systems.
- It uses internal representations like key-value caches instead of text-based protocols.
- The framework performs parameter-efficient supervised training over multi-agent latent trajectories.
- Agents jointly learn how information should be encoded and interpreted across interactions.
- Experiments cover mathematical reasoning, scientific QA, code generation, and commonsense benchmarks.
- DiffMAS consistently improves reasoning accuracy and decoding.
- Current multi-agent LLM research focuses on agent roles and orchestration, not communication.
- The work is published on arXiv with ID 2604.21794.
Entities
Institutions
- arXiv