Do LLMs Know Tool Irrelevance? Demystifying Structural Alignment Bias in Tool Invocations

📰 ArXiv cs.AI

arXiv:2604.11322v1 Announce Type: cross Abstract: Large language models (LLMs) have demonstrated impressive capabilities in utilizing external tools. In practice, however, LLMs are often exposed to tools that are irrelevant to the user's query, in which case the desired behavior is to refrain from invocations. In this work, we identify a widespread yet overlooked mechanistic flaw in tool refusal, which we term structural alignment bias: Even when a tool fails to serve the user's goal, LLMs still

Published 14 Apr 2026
Read full paper → ← Back to Reads