ArXiv TLDR

SAGE: Signal-Amplified Guided Embeddings for LLM-based Vulnerability Detection

🐦 Tweet
2604.19031

Zhengyang Shan, Xu Qian, Jiayun Xin, Minghui Xu, Yue Zhang + 3 more

cs.CR

TLDR

SAGE introduces Signal-Amplified Guided Embeddings to overcome "Signal Submersion" in LLM-based vulnerability detection, achieving SOTA performance.

Key contributions

  • SAGE framework mitigates "Signal Submersion" in LLMs for vulnerability detection.
  • Employs task-conditional Sparse Autoencoders to isolate and amplify faint vulnerability signals.
  • Achieves SOTA performance, boosting a 7B model's MCC by over 300% on diverse datasets.
  • Increases internal Signal-to-Noise Ratio by 12.7x, outperforming 34B models across 13 languages.

Why it matters

LLMs are powerful for vulnerability detection but struggle with "Signal Submersion." SAGE provides a novel mechanistic solution, amplifying critical signals. This enables smaller models to achieve significant performance gains, offering a more efficient and scalable path to software security.

Original Abstract

Software vulnerabilities are a primary threat to modern infrastructure. While static analysis and Graph Neural Networks have long served as the foundation for vulnerability detection, the emergence of Large Language Models (LLMs) has introduced a transformative paradigm driven by superior semantic reasoning and cross-environment generalization. However, in the context of LLM-based vulnerability detection, we identify a fundamental bottleneck in these models termed \textbf{Signal Submersion}: a state where features related to vulnerability are activated internally but numerically overwhelmed by dominant functional semantics. To address this, we propose \textbf{SAGE} (\textbf{S}ignal-\textbf{A}mplified \textbf{G}uided \textbf{E}mbeddings), a framework that shifts from passive signal submersion to active signal recovery. SAGE integrates task-conditional Sparse Autoencoders (SAEs) to isolate and amplify these faint vulnerability signals. Extensive evaluations on BigVul, PrimeVul, and PreciseBugs demonstrate that SAGE achieves state-of-the-art performance. Notably, SAGE mitigates Signal Submersion by increasing the internal Signal-to-Noise Ratio (SNR) by 12.7$\times$ via sparse manifold projection. This mechanistic intervention enables a 7B model to achieve up to 318\% Matthews Correlation Coefficient (MCC) gains on unseen distributions and a 319\% gain on classic datasets. By maintaining robust performance across 13 programming languages and outperforming 34B baselines, SAGE establishes a more efficient and scalable path to software security than simple parameter scaling.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.