SAGE: Signal-Amplified Guided Embeddings for LLM-based Vulnerability Detection
Zhengyang Shan, Xu Qian, Jiayun Xin, Minghui Xu, Yue Zhang + 3 more
TLDR
SAGE introduces Signal-Amplified Guided Embeddings to overcome "Signal Submersion" in LLM-based vulnerability detection, achieving SOTA performance.
Key contributions
- SAGE framework mitigates "Signal Submersion" in LLMs for vulnerability detection.
- Employs task-conditional Sparse Autoencoders to isolate and amplify faint vulnerability signals.
- Achieves SOTA performance, boosting a 7B model's MCC by over 300% on diverse datasets.
- Increases internal Signal-to-Noise Ratio by 12.7x, outperforming 34B models across 13 languages.
Why it matters
LLMs are powerful for vulnerability detection but struggle with "Signal Submersion." SAGE provides a novel mechanistic solution, amplifying critical signals. This enables smaller models to achieve significant performance gains, offering a more efficient and scalable path to software security.
Original Abstract
Software vulnerabilities are a primary threat to modern infrastructure. While static analysis and Graph Neural Networks have long served as the foundation for vulnerability detection, the emergence of Large Language Models (LLMs) has introduced a transformative paradigm driven by superior semantic reasoning and cross-environment generalization. However, in the context of LLM-based vulnerability detection, we identify a fundamental bottleneck in these models termed \textbf{Signal Submersion}: a state where features related to vulnerability are activated internally but numerically overwhelmed by dominant functional semantics. To address this, we propose \textbf{SAGE} (\textbf{S}ignal-\textbf{A}mplified \textbf{G}uided \textbf{E}mbeddings), a framework that shifts from passive signal submersion to active signal recovery. SAGE integrates task-conditional Sparse Autoencoders (SAEs) to isolate and amplify these faint vulnerability signals. Extensive evaluations on BigVul, PrimeVul, and PreciseBugs demonstrate that SAGE achieves state-of-the-art performance. Notably, SAGE mitigates Signal Submersion by increasing the internal Signal-to-Noise Ratio (SNR) by 12.7$\times$ via sparse manifold projection. This mechanistic intervention enables a 7B model to achieve up to 318\% Matthews Correlation Coefficient (MCC) gains on unseen distributions and a 319\% gain on classic datasets. By maintaining robust performance across 13 programming languages and outperforming 34B baselines, SAGE establishes a more efficient and scalable path to software security than simple parameter scaling.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.