Layerwise Convergence Fingerprints for Runtime Misbehavior Detection in Large Language Models
Nay Myat Min, Long H. Pham, Jun Sun
TLDR
LCF is a tuning-free runtime monitor that detects LLM misbehavior like backdoors, jailbreaks, and prompt injections by analyzing hidden-state trajectories.
Key contributions
- Introduces Layerwise Convergence Fingerprinting (LCF) for general LLM runtime misbehavior detection.
- Monitors inter-layer hidden-state trajectories using Mahalanobis distance and Ledoit-Wolf shrinkage.
- Detects backdoors, jailbreaks, and prompt injections across various LLMs with high accuracy.
- Operates without reference models, trigger knowledge, or retraining, with minimal inference overhead.
Why it matters
Current LLM runtime defenses are often threat-specific or require assumptions like trigger knowledge. LCF offers a general, tuning-free solution to protect LLMs from diverse runtime threats. This enhances the safety and reliability of deployed LLMs, making them more robust against unforeseen attacks.
Original Abstract
Large language models deployed at runtime can misbehave in ways that clean-data validation cannot anticipate: training-time backdoors lie dormant until triggered, jailbreaks subvert safety alignment, and prompt injections override the deployer's instructions. Existing runtime defenses address these threats one at a time and often assume a clean reference model, trigger knowledge, or editable weights, assumptions that rarely hold for opaque third-party artifacts. We introduce Layerwise Convergence Fingerprinting (LCF), a tuning-free runtime monitor that treats the inter-layer hidden-state trajectory as a health signal: LCF computes a diagonal Mahalanobis distance on every inter-layer difference, aggregates via Ledoit-Wolf shrinkage, and thresholds via leave-one-out calibration on 200 clean examples, with no reference model, trigger knowledge, or retraining. Evaluated on four architectures (Llama-3-8B, Qwen2.5-7B, Gemma-2-9B, Qwen2.5-14B) across backdoors, jailbreaks, and prompt injection (56 backdoor combinations, 3 jailbreak techniques, and BIPIA email + code-QA), LCF reduces mean backdoor attack success rate (ASR) below 1% on Qwen2.5-7B and Gemma-2 and to 1.3% on Qwen2.5-14B, detects 92-100% of DAN jailbreaks (62-100% for GCG and softer role-play), and flags 100% of text-payload injections across all eight (model, domain) cells, at 12-16% backdoor FPR and <0.1% inference overhead. A single aggregation score covers all three threat families without threat-specific tuning, positioning LCF as a general-purpose runtime safety layer for cloud-served and on-device LLMs.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.