Semantic Denial of Service in LLM-controlled robots

📰 ArXiv cs.AI

arXiv:2604.24790v1 Announce Type: cross Abstract: Safety-oriented instruction-following is supposed to keep LLM-controlled robots safe. We show it also creates an availability attack surface. By injecting short safety-plausible phrases (1-5 tokens) into a robots audio channel, an adversary can trigger the models safety reasoning to halt or disrupt execution without jailbreaking the model or overriding its policy. In the embodied setting, this is a semantic denial-of-service attack: the agent sto

Published 29 Apr 2026
Read full paper → ← Back to Reads