Large Language Models Explore by Latent Distilling

📰 ArXiv cs.AI

arXiv:2604.24927v1 Announce Type: cross Abstract: Generating diverse responses is crucial for test-time scaling of large language models (LLMs), yet standard stochastic sampling mostly yields surface-level lexical variation, limiting semantic exploration. In this paper, we propose Exploratory Sampling (ESamp), a decoding approach that explicitly encourages semantic diversity during generation. ESamp is motivated by the well-known observation that neural networks tend to make lower-error predicti

Published 29 Apr 2026
Read full paper → ← Back to Reads