Independent-Component-Based Encoding Models of Brain Activity During Story Comprehension
Kamya Hari, Taha Binhuraib, Jin Li, Cory Shain, Anna A. Ivanova
TLDR
This paper introduces an independent component (IC)-based encoding framework to model brain activity during story comprehension, improving on traditional voxelwise methods.
Key contributions
- Introduces an IC-based encoding framework to model fMRI brain activity during story comprehension.
- Predicts IC time series from linguistic input using large language model representations.
- Identified highly predictable ICs, consistent across subjects, reflecting auditory and language networks.
- Offers interpretable, network-level analysis, accommodating individual brain variability.
Why it matters
This paper addresses key limitations of traditional fMRI encoding models by introducing an IC-based approach. It offers a more robust and interpretable way to link continuous stimulus features to neural activity. This method allows for analysis at the functional network level, better accounting for individual differences and providing clearer insights into brain processes during complex tasks like story comprehension.
Original Abstract
Encoding models provide a powerful framework for linking continuous stimulus features to neural activity; however, traditional voxelwise approaches are limited by measurement noise, inter-subject variability, and redundancy arising from spatially correlated voxels encoding overlapping neural signals. Here, we propose an independent component (IC)-based encoding framework that dissociates stimulus-driven and noise-driven signals in fMRI data. We decompose continuous fMRI data from naturalistic story listening into ICs using one subset of the data, and train encoding models on independent data to predict IC time series from large language model representations of linguistic input. Across subjects, a subset of ICs exhibited consistently high predictivity. These ICs were spatially and temporally consistent across subjects and included cognitive networks known to respond during story listening (auditory and language). Auditory component time series were strongly correlated with acoustic stimulus features, highlighting the interpretability of identified component time series. Components identified as noise or motion-related artifacts by ICA-AROMA showed uniformly poor predictive performance, confirming that highly predicted components reflect genuine stimulus-related neural signals rather than confounds. Overall, IC-based encoding models enable analyses at the level of functional networks, accommodating the variability in network locations across individuals and providing interpretable results that are easy to compare across subjects.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.