ArXiv TLDR

OmniMouse: Scaling properties of multi-modal, multi-task Brain Models on 150B Neural Tokens

🐦 Tweet
2604.18827

Konstantin F. Willeke, Polina Turishcheva, Alex Gilbert, Goirik Chakrabarty, Hasan A. Bedel + 16 more

q-bio.NCcs.AI

TLDR

OmniMouse models brain activity, showing performance scales with data, not model size, unlike standard AI.

Key contributions

  • Developed OmniMouse, a multi-modal, multi-task model for brain activity prediction and decoding.
  • Leveraged a massive dataset of 150 billion neural tokens from 73 mice visual cortices.
  • Achieved state-of-the-art performance in neural prediction, behavioral decoding, and forecasting.
  • Revealed brain models are data-limited, scaling with data volume but not model size, unlike standard AI.

Why it matters

This paper challenges AI's standard scaling laws, showing brain models are data-limited, not model-size limited. It suggests that future neuroscience AI breakthroughs may depend on even larger, richer datasets, potentially unlocking emergent properties akin to large language models. This shifts focus for brain modeling research.

Original Abstract

Scaling data and artificial neural networks has transformed AI, driving breakthroughs in language and vision. Whether similar principles apply to modeling brain activity remains unclear. Here we leveraged a dataset of 3.1 million neurons from the visual cortex of 73 mice across 323 sessions, totaling more than 150 billion neural tokens recorded during natural movies, images and parametric stimuli, and behavior. We train multi-modal, multi-task models that support three regimes flexibly at test time: neural prediction, behavioral decoding, neural forecasting, or any combination of the three. OmniMouse achieves state-of-the-art performance, outperforming specialized baselines across nearly all evaluation regimes. We find that performance scales reliably with more data, but gains from increasing model size saturate. This inverts the standard AI scaling story: in language and computer vision, massive datasets make parameter scaling the primary driver of progress, whereas in brain modeling -- even in the mouse visual cortex, a relatively simple system -- models remain data-limited despite vast recordings. The observation of systematic scaling raises the possibility of phase transitions in neural modeling, where larger and richer datasets might unlock qualitatively new capabilities, paralleling the emergent properties seen in large language models. Code available at https://github.com/enigma-brain/omnimouse.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.