When Reasoning Models Hurt Behavioral Simulation: A Solver-Sampler Mismatch in Multi-Agent LLM Negotiation

📰 ArXiv cs.AI

Stronger reasoning in LLMs can hurt behavioral simulation in multi-agent negotiation, learn how to identify and address solver-sampler mismatch

advanced Published 16 Apr 2026
Action Steps
  1. Identify the objective of your simulation: is it to solve a strategic problem or to sample plausible boundedly rational behavior?
  2. Assess the potential for solver-sampler mismatch in your LLM negotiation model
  3. Evaluate the trade-off between reasoning strength and simulation fidelity in your model
  4. Consider using techniques such as regularization or noise injection to mitigate over-optimization
  5. Test and validate your model's performance in simulations with varying levels of complexity and uncertainty
Who Needs to Know This

Researchers and developers working on multi-agent systems and LLM negotiation can benefit from understanding the limitations of reasoning models in behavioral simulation, to improve the fidelity of their simulations

Key Insight

💡 Reasoning-enhanced models can become better solvers and worse simulators, highlighting the need for a nuanced approach to modeling boundedly rational behavior

Share This
🚨 Stronger reasoning in LLMs can hurt behavioral simulation in multi-agent negotiation! 🤖 Learn how to identify and address solver-sampler mismatch #LLMs #MultiAgentSystems
Read full paper → ← Back to Reads