ArXiv TLDR

Learning reveals invisible structure in low-rank RNNs

🐦 Tweet
2605.04115

Yoav Ger, Omri Barak

cs.LGcs.AIq-bio.NC

TLDR

A new theory for low-rank RNNs reveals 'loss-invisible' overlaps that govern learning dynamics and store training history, offering testable predictions.

Key contributions

  • Derives gradient-descent dynamics for low-rank RNNs in a reduced overlap space.
  • Formulates a closed-form, low-dimensional ODE system governing learning dynamics.
  • Identifies "loss-invisible" overlaps that store training history and expose network differences.

Why it matters

This work provides a crucial theoretical understanding of learning in low-rank RNNs, a widely used model. By revealing hidden structures, it offers new insights into how neural networks learn and remember. The derived predictions can guide future biological experiments.

Original Abstract

Learning in neural systems arises from synaptic changes that reshape the representations underlying behavior. While low-rank recurrent neural networks (RNNs) have emerged as a powerful framework for linking connectivity to function, a theoretical understanding of their learning process remains elusive. Here, we extend the low-rank framework from activity to learning by deriving gradient-descent dynamics directly in a reduced overlap space. We formulate a closed-form, low-dimensional system of ODEs that governs learning in this space, exact for linear RNNs and asymptotically exact for nonlinear RNNs in the large-$N$ Gaussian limit. Central to our analysis is a distinction between two classes of overlaps: loss-visible overlaps, which fully determine network activity, output, and loss, and loss-invisible overlaps, which do not affect function but are required to describe learning. We illustrate the consequences of this decomposition through two phenomena. First, we show that learning can serve as a perturbation that exposes differences in connectivity between functionally equivalent networks. Second, we show that loss-invisible overlaps can act as memory variables that encode training history, and characterize the conditions under which this occurs. Finally, we present several testable predictions for biological learning experiments derived from our theory.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.