FlexGuard: Continuous Risk Scoring for Strictness-Adaptive LLM Content Moderation
📰 ArXiv cs.AI
arXiv:2602.23636v3 Announce Type: replace-cross Abstract: Ensuring the safety of LLM-generated content is essential for real-world deployment. Most existing guardrail models formulate moderation as a fixed binary classification task, implicitly assuming a fixed definition of harmfulness. In practice, enforcement strictness - how conservatively harmfulness is defined and enforced - varies across platforms and evolves over time, making binary moderators brittle under shifting requirements. We firs
DeepCamp AI