ArXiv TLDR

Learning Hippo: Multi-attractor Dynamics and Stability Effects in a Biologically Detailed CA3 Extension of Hopfield Networks

🐦 Tweet
2604.20679

Daniele Corradetti, Renato Corradetti

cs.NE

TLDR

This paper introduces a biologically detailed CA3 extension of Hopfield networks, demonstrating improved multi-attractor dynamics, associative recall, and stability.

Key contributions

  • Developed a biologically detailed CA3 model with 10 populations, 47 compartments, and multi-rule plasticity.
  • Demonstrated multi-attractor dynamics with realistic inhibition, a feature absent in minimal Hopfield networks.
  • Achieved target-selective associative recall, retrieving B from partial A cues in paired memory tasks.
  • Showed reduced cross-seed variance, indicating enhanced stability compared to baseline models.

Why it matters

This paper significantly advances our understanding of hippocampal memory by demonstrating how biological complexity enhances Hopfield-like network capabilities. It suggests that detailed neural architectures are crucial for achieving robust and flexible memory functions, offering new insights into brain-inspired AI.

Original Abstract

We present a biologically detailed extension of the classical Hopfield/Marr auto-associative memory model for CA3, implementing ten populations (two asymmetric pyramidal subtypes, eight GABAergic interneuron classes), forty-seven compartments, multi-rule plasticity (recurrent Hebb, BCM anti-saturation, mossy-fiber short-term, endocannabinoid iLTD, burst-gated Hebb), and a bimodal cholinergic encoding/consolidation cycle. Evaluated on pattern completion across auto-associative, associative, and temporal regimes, and on a controlled inhibitory-proportion manipulation at $N{=}256$, the full architecture exhibits \emph{three qualitative signatures absent from a minimal Hopfield baseline}: (i)~multi-attractor cross-seed behaviour at $K{=}5$ with biologically realistic inhibitory proportions, where two of five seeds converge to positive attractors with margin ${+}0.10{-}0.22$ (Cohen's $d{=}0.71$, one-sided $p{=}0.08$); (ii)~target-selective associative recall in paired $(A, B)$ memory at $K{\geq}5$, where the full model retrieves $B$ from a partial cue of $A$ while the minimal model echoes $A$ (Pearson margin $Δ{=}{+}0.163$ at $K{=}5$); (iii)~reduced cross-seed variance of the full model below the minimal baseline under clean upstream, with ratios $1.0{-}3.0$. These three signatures are architecture-specific: they appear consistently across independent regimes and are absent from the minimal control.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.