Latent Agents: A Post-Training Procedure for Internalized Multi-Agent Debate
📰 ArXiv cs.AI
arXiv:2604.24881v1 Announce Type: new Abstract: Multi-agent debate has been shown to improve reasoning in large language models (LLMs). However, it is compute-intensive, requiring generation of long transcripts before answering questions. To address this inefficiency, we develop a framework that distills multi-agent debate into a single LLM through a two-stage fine-tuning pipeline combining debate structure learning with internalization via dynamic reward scheduling and length clipping. Across m
DeepCamp AI