ArXiv TLDR

Meta-learning In-Context Enables Training-Free Cross Subject Brain Decoding

🐦 Tweet
2604.08537

Mu Nan, Muquan Yu, Weijian Mai, Jacob S. Prince, Hossein Adeli + 10 more

cs.LGq-bio.NC

TLDR

A meta-learning method allows training-free, cross-subject fMRI brain decoding by inferring individual neural patterns in-context, eliminating fine-tuning.

Key contributions

  • Introduces a meta-optimized approach for training-free, cross-subject fMRI visual decoding.
  • Enables rapid inference of individual neural encoding patterns using a small set of in-context examples.
  • Achieves strong generalization across subjects and scanners without fine-tuning or anatomical alignment.

Why it matters

This paper tackles the challenge of individual neural variability in brain decoding, a major obstacle to generalizable models. It's a critical step towards foundation models for non-invasive brain decoding, potentially accelerating neuroscience research and applications.

Original Abstract

Visual decoding from brain signals is a key challenge at the intersection of computer vision and neuroscience, requiring methods that bridge neural representations and computational models of vision. A field-wide goal is to achieve generalizable, cross-subject models. A major obstacle towards this goal is the substantial variability in neural representations across individuals, which has so far required training bespoke models or fine-tuning separately for each subject. To address this challenge, we introduce a meta-optimized approach for semantic visual decoding from fMRI that generalizes to novel subjects without any fine-tuning. By simply conditioning on a small set of image-brain activation examples from the new individual, our model rapidly infers their unique neural encoding patterns to facilitate robust and efficient visual decoding. Our approach is explicitly optimized for in-context learning of the new subject's encoding model and performs decoding by hierarchical inference, inverting the encoder. First, for multiple brain regions, we estimate the per-voxel visual response encoder parameters by constructing a context over multiple stimuli and responses. Second, we construct a context consisting of encoder parameters and response values over multiple voxels to perform aggregated functional inversion. We demonstrate strong cross-subject and cross-scanner generalization across diverse visual backbones without retraining or fine-tuning. Moreover, our approach requires neither anatomical alignment nor stimulus overlap. This work is a critical step towards a generalizable foundation model for non-invasive brain decoding.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.