Latent Agents: A Post-Training Procedure for Internalized Multi-Agent Debate

📰 ArXiv cs.AI

arXiv:2604.24881v1 Announce Type: new Abstract: Multi-agent debate has been shown to improve reasoning in large language models (LLMs). However, it is compute-intensive, requiring generation of long transcripts before answering questions. To address this inefficiency, we develop a framework that distills multi-agent debate into a single LLM through a two-stage fine-tuning pipeline combining debate structure learning with internalization via dynamic reward scheduling and length clipping. Across m

Published 29 Apr 2026
Read full paper → ← Back to Reads