CARO: Chain-of-Analogy Reasoning Optimization for Robust Content Moderation

📰 ArXiv cs.AI

arXiv:2604.10504v1 Announce Type: new Abstract: Current large language models (LLMs), even those explicitly trained for reasoning, often struggle with ambiguous content moderation cases due to misleading "decision shortcuts" embedded in context. Inspired by cognitive psychology insights into expert moderation, we introduce \caro (Chain-of-Analogy Reasoning Optimization), a novel two-stage training framework to induce robust analogical reasoning in LLMs. First, \caro bootstraps analogical reasoni

Published 14 Apr 2026
Read full paper → ← Back to Reads