Detecting Hallucinations in SpeechLLMs at Inference Time Using Attention Maps

📰 ArXiv cs.AI

arXiv:2604.19565v1 Announce Type: cross Abstract: Hallucinations in Speech Large Language Models (SpeechLLMs) pose significant risks, yet existing detection methods typically rely on gold-standard outputs that are costly or impractical to obtain. Moreover, hallucination detection methods developed for text-based LLMs do not directly capture audio-specific signals. We investigate four attention-derived metrics: AUDIORATIO, AUDIOCONSISTENCY, AUDIOENTROPY, and TEXTENTROPY, designed to capture patho

Published 22 Apr 2026
Read full paper → ← Back to Reads