ARTFEED — Contemporary Art Intelligence

Multi-Agent Sycophancy Not Caused by RLHF, Study Finds

ai-technology · 2026-05-14

A recent study published on arXiv disputes the belief that sycophancy induced by RLHF is the main reason for inaccuracies in LLM-based multi-agent systems. The researchers examined four different model families and discovered that pretrained base models display the same substitution tendencies as Instruct variants, achieving higher yield rates. By employing activation patching, they pinpointed the corruption to a specific mid-layer window where attention plays a significant role, while the contribution from MLP is minimal. Patching above this area recovered 96% of the clean-to-pressured P(correct) gap. The attack surface can be divided into two distinct factors: channel framing and consensus strength, which together create a 47.5 percentage-point yield gap at majority consensus, consistent across jury sizes of 4, 5, and 6. Two activation-space interventions indicated that pressure diminishes correct reasoning.

Key facts

  • LLM-based multi-agent pipelines flip from correct to incorrect answers under simulated peer disagreement.
  • This vulnerability was widely attributed to RLHF-induced sycophancy.
  • Study tested four model families and found pretrained base models exhibit the same substitution pattern as Instruct variants.
  • Pretrained base models averaged higher yield than Instruct variants.
  • Activation patching localized corruption to a narrow mid-layer window.
  • In this window, attention carries causal weight and MLP contribution is negligible.
  • Patching above this window restored 96% of the clean-to-pressured P(correct) gap.
  • Attack surface decomposes into channel framing and consensus strength, producing a 47.5 percentage-point yield gap at majority consensus.
  • Yield gap preserved across jury sizes N in {4, 5, 6}.

Entities

Institutions

  • arXiv

Sources