ArXiv TLDR

Stability and Generalization in Looped Transformers

🐦 Tweet
2604.15259

Asher Labovich

cs.LGcs.AI

TLDR

This paper introduces a fixed-point framework to analyze stability and generalization in looped transformers, showing recall with outer normalization is key.

Key contributions

  • Introduces a fixed-point framework for analyzing stability (reachability, input-dependence, geometry) in looped transformers.
  • Theoretically proves looped networks without recall cannot achieve strong input-dependence or stable fixed points.
  • Shows recall combined with outer normalization ensures reachable, smooth, and backprop-stable fixed points.
  • Empirically validates framework predictions across chess, sudoku, and prefix-sums tasks, introducing internal recall.

Why it matters

Looped transformers offer compute scaling but struggle with generalization. This paper provides a crucial theoretical framework and empirical evidence for architectural choices that enable stable and generalizable extrapolation. It offers clear guidance for designing more robust looped transformer models.

Original Abstract

Looped transformers promise test-time compute scaling by spending more iterations on harder problems, but it remains unclear which architectural choices let them extrapolate to harder problems at test time rather than memorize training-specific solutions. We introduce a fixed-point based framework for analyzing looped architectures along three axes of stability -- reachability, input-dependence, and geometry -- and use it to characterize when fixed-point iteration yields meaningful predictions. Theoretically, we prove that looped networks without recall have countable fixed points and cannot achieve strong input-dependence at any spectral regime, while recall combined with outer normalization reliably produces a regime in which fixed points are simultaneously reachable, locally smooth in the input, and supported by stable backpropagation. Empirically, we train single-layer looped transformers on chess, sudoku, and prefix-sums and find that downstream performance tracks the framework's predictions across tasks and architectural configurations. We additionally introduce internal recall, a novel recall placement variant, and show that it becomes competitive with -- and on sudoku, substantially better than -- standard recall placement once outer normalization is applied.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.