Generalization at the Edge of Stability

📰 ArXiv cs.AI

arXiv:2604.19740v1 Announce Type: cross Abstract: Training modern neural networks often relies on large learning rates, operating at the edge of stability, where the optimization dynamics exhibit oscillatory and chaotic behavior. Empirically, this regime often yields improved generalization performance, yet the underlying mechanism remains poorly understood. In this work, we represent stochastic optimizers as random dynamical systems, which often converge to a fractal attractor set (rather than

Published 22 Apr 2026
Read full paper → ← Back to Reads