The Salami Slicing Threat: Exploiting Cumulative Risks in LLM Systems
📰 ArXiv cs.AI
arXiv:2604.11309v1 Announce Type: cross Abstract: Large Language Models (LLMs) face prominent security risks from jailbreaking, a practice that manipulates models to bypass built-in security constraints and generate unethical or unsafe content. Among various jailbreak techniques, multi-turn jailbreak attacks are more covert and persistent than single-turn counterparts, exposing critical vulnerabilities of LLMs. However, existing multi-turn jailbreak methods suffer from two fundamental limitation
DeepCamp AI