Are Natural-Domain Foundation Models Effective for Accelerated Cardiac MRI Reconstruction?
Anam Hashmi, Mayug Maniparambil, Julia Dietlmeier, Kathleen M. Curran, Noel E. O'Connor
TLDR
This paper explores if natural-domain foundation models can effectively serve as image priors for accelerated cardiac MRI reconstruction, finding improved robustness in cross-domain scenarios.
Key contributions
- Investigates natural-domain foundation models (CLIP, DINOv2) as priors for accelerated cardiac MRI reconstruction.
- Proposes an unrolled reconstruction framework integrating frozen visual encoders like CLIP and DINOv2.
- Demonstrates foundation models offer improved robustness in challenging cross-domain MRI reconstruction.
- Finds natural-image-pretrained models learn highly transferable structural representations for MRI.
Why it matters
Foundation models have transformed computer vision, but their use in physics-based inverse problems like MRI reconstruction is new. This work demonstrates their potential for robust generalization, especially in challenging cross-domain scenarios. It suggests a promising path for leveraging large-scale priors in medical imaging.
Original Abstract
The emergence of large-scale pretrained foundation models has transformed computer vision, enabling strong performance across diverse downstream tasks. However, their potential for physics-based inverse problems, such as accelerated cardiac MRI reconstruction, remains largely underexplored. In this work, we investigate whether natural-domain foundation models can serve as effective image priors for accelerated cardiac MRI reconstruction, and compare the performance obtained against domain-specific counterparts such as BiomedCLIP. We propose an unrolled reconstruction framework that incorporates pretrained, frozen visual encoders, such as CLIP, DINOv2, and BiomedCLIP, within each cascade to guide the reconstruction process. Through extensive experiments, we show that while task-specific state-of-the-art reconstruction models such as E2E-VarNet achieve superior performance in standard in-distribution settings, foundation-model-based approaches remain competitive. More importantly, in challenging cross-domain scenarios, where models are trained on cardiac MRI and evaluated on anatomically distinct knee and brain datasets--foundation models exhibit improved robustness, particularly under high acceleration factors and limited low-frequency sampling. We further observe that natural-image-pretrained models, such as CLIP, learn highly transferable structural representations, while domain-specific pretraining (BiomedCLIP) provides modest additional gains in more ill-posed regimes. Overall, our results suggest that pretrained foundation models offer a promising source of transferable priors, enabling improved robustness and generalization in accelerated MRI reconstruction.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.