NeuralBench: A Unifying Framework to Benchmark NeuroAI Models
Hubert Banville, Stéphane d'Ascoli, Simon Dahan, Jérémy Rapin, Marlène Careil + 10 more
TLDR
NeuralBench is a unified open-source framework for systematically benchmarking AI models of brain activity, including a large EEG benchmark.
Key contributions
- Introduces NeuralBench, a unified framework for benchmarking AI models of brain activity.
- Presents NeuralBench-EEG v1.0, a large benchmark with 36 tasks, 14 architectures, and 94 datasets.
- Shows current foundation models only marginally outperform task-specific models on EEG data.
- Reveals many cognitive decoding and clinical prediction tasks remain highly challenging.
Why it matters
This paper addresses the critical need for systematic evaluation of AI models processing brain recordings, which currently suffer from varied pipelines and limited tasks. NeuralBench provides a standardized, extensible framework to unify benchmarking, guiding future NeuroAI research.
Original Abstract
Deep learning and large public datasets have recently catalyzed the proliferation of AI models for processing brain recordings. However, systematically evaluating these models remains a challenge: not only do the preprocessing pipelines, training and finetuning approaches largely vary across studies, but their downstream evaluation is often limited to small sets of tasks and/or datasets. Here, we present NeuralBench: a unified framework for benchmarking AI models of brain activity. We accompany this framework with NeuralBench-EEG v1.0 -- a large EEG benchmark that includes 36 electroencephalography (EEG) tasks and 14 deep learning architectures, and is evaluated on 94 datasets accessed through a standardized interface. This first EEG-focused release already highlights two main findings. First, current foundation models only marginally outperform task-specific models. Second, a large set of tasks (e.g. cognitive decoding, clinical predictions) remain highly challenging, even for the best models. Critically, NeuralBench is designed for the integration of new tasks, datasets, models, and neuroimaging modalities, as illustrated by preliminary extensions to MEG and fMRI datasets and models. Through this white paper, we invite the community to expand this open-source framework and work together toward a unified benchmarking standard for neuroimaging models.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.