Comparative reversal learning reveals rigid adaptation in LLMs under non-stationary uncertainty

📰 ArXiv cs.AI

arXiv:2604.04182v1 Announce Type: new Abstract: Non-stationary environments require agents to revise previously learned action values when contingencies change. We treat large language models (LLMs) as sequential decision policies in a two-option probabilistic reversal-learning task with three latent states and switch events triggered by either a performance criterion or timeout. We compare a deterministic fixed transition cycle to a stochastic random schedule that increases volatility, and eval

Published 7 Apr 2026
Read full paper → ← Back to News