ArXiv TLDR

Virtual Scanning for NSCLC Histology: Investigating the Discriminatory Power of Synthetic PET

🐦 Tweet
2605.02746

Fatih Aksu, Laura Ciuffetti, Francesco Di Feola, Filippo Ruffini, Giulia Romoli + 4 more

cs.CVcs.AI

TLDR

This paper uses a 3D Pix2Pix GAN to generate synthetic PET scans from CT, significantly improving NSCLC subtype classification.

Key contributions

  • Proposed a 3D Pix2Pix GAN to synthesize pseudo-PET volumes from anatomical CT scans.
  • Integrated synthetic PET with CT data using a multi-stage intermediate fusion (MINT) framework.
  • Demonstrated significant improvement in NSCLC subtype classification (ADC/SCC) over CT-only baseline.
  • Achieved AUC increase from 0.489 to 0.591 and GMean from 0.305 to 0.524 on a 714-subject dataset.

Why it matters

This paper offers a novel solution to a critical problem in lung cancer diagnosis: differentiating subtypes without costly and radiation-heavy PET scans. By generating synthetic PET, it provides a valuable alternative, making advanced diagnostic features more accessible. This could lead to more personalized and effective treatments.

Original Abstract

Accurate histological differentiation between adenocarcinoma (ADC) and squamous cell carcinoma (SCC) is critical for personalized treatment in non-small cell lung cancer (NSCLC). While [$^{18}$F]FDG PET/CT is a standard tool for the clinical evaluation of lung cancer, its utility is often limited by high costs and radiation exposure. In this paper, we investigate the feasibility of "virtual scanning" as a feature-enhancement strategy by evaluating whether synthetic PET data can provide complementary feature representations to supplement anatomical CT scans in histological subtype classification. We propose a framework that leverages a 3D Pix2Pix Generative Adversarial Network (GAN), pretrained on the FDG-PET/CT Lesions dataset, to synthesize pseudo-PET volumes from anatomical CT scans. These synthetic volumes are integrated with structural CT data within the MINT framework, a multi-stage intermediate fusion architecture. Our experiments, conducted on a multi-center dataset of 714 subjects, demonstrate that the inclusion of synthetic metabolic features significantly improves classification performance over a CT-only baseline. The multimodal approach achieved a statistically significant increase in the Area Under the Curve (AUC) from 0.489 to 0.591 and improved the Geometric Mean (GMean) from 0.305 to 0.524. These results suggest that synthetic PET scans provide discriminatory metabolic cues that enable deep learning models to exploit complementary cross-modal information, offering a potential feature-enhancement strategy for clinical scenarios where physical PET scans are unavailable.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.