TokUR: Token-Level Uncertainty Estimation for Large Language Model Reasoning

📰 ArXiv cs.AI

arXiv:2505.11737v4 Announce Type: replace-cross Abstract: While Large Language Models (LLMs) have demonstrated impressive capabilities, their output quality remains inconsistent across various application scenarios, making it difficult to identify trustworthy responses, especially in complex tasks requiring multi-step reasoning. In this paper, we propose a Token-level Uncertainty estimation framework for Reasoning (TokUR) that enables LLMs to self-assess and self-improve their responses in mathe

Published 14 Apr 2026
Read full paper → ← Back to Reads