Topological Signatures of Grokking
Yifan Tang, Qiquan Wang, Inés García-Redondo, Anthea Monod
TLDR
Persistent homology identifies a clear topological signature of grokking, showing a sharp increase in H1 persistence linked to generalization.
Key contributions
- Applies persistent homology to analyze the grokking phenomenon in neural networks.
- Discovers a topological signature of grokking: a sharp increase in H1 persistence.
- Reveals the emergence of long-lived topological features reflecting cyclic task structure.
- Demonstrates these topological transitions are linked to generalization, not memorization.
Why it matters
This paper introduces persistent homology as a novel, interpretable framework to analyze how neural networks internalize latent structure. It offers a unified geometric and topological characterization of representation learning, capturing both local and global multi-scale structure. This advances our understanding of generalization.
Original Abstract
We study the grokking phenomenon through the lens of topology. Using persistent homology on point clouds derived from the embedding matrices of a range of models trained on modular arithmetic with varying primes, we identify a clear and consistent topological signature of grokking: a sharp increase in both the maximum and total persistence of first homology ($H_1$). Persistence diagrams reveal the emergence of a dominant long-lived topological feature together with increasingly structured secondary features, reflecting the underlying cyclic structure of the task. Compared to existing spectral and geometric diagnostics -- specifically, Fourier analysis and local intrinsic dimension -- persistent homology provides a unified geometric and topological characterization of representation learning, capturing both local and global multi-scale structure. Ablations across data regimes and control settings show that these topological transitions are tied to generalization rather than memorization. Our results suggest that persistent homology offers a principled and interpretable framework for analyzing how neural networks internalize latent structure during training.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.