ArXiv TLDR
← All categories

Cryptography & Security

Research on AI security, adversarial attacks, privacy, and cryptographic methods.

cs.CR · 500 papers

VectorSmuggle: Steganographic Exfiltration in Embedding Stores and a Cryptographic Provenance Defense

VectorSmuggle reveals steganographic data exfiltration in RAG embedding stores and proposes VectorPin, a cryptographic defense for embedding integrity.

2605.13764May 13, 2026Jascha Wanger

DisAgg: Distributed Aggregators for Efficient Secure Aggregation in Federated Learning

DisAgg uses distributed client aggregators to securely and efficiently aggregate updates in federated learning, achieving a 4.6x speedup over OPA.

2605.13708May 13, 2026Haaris Mehmood, Giorgos Tatsis, Dimitrios Alexopoulos +4

Identifying AI Web Scrapers Using Canary Tokens

This paper introduces a novel method using canary tokens to reliably identify which web scrapers are feeding data to specific large language models.

2605.13706May 13, 2026Steven Seiden, Triss Ren, Caroline Zhang +3

Limits of Personalizing Differential Privacy Budgets

This paper reveals that personalized differential privacy budgets have significant limitations, showing a simple thresholding method is often superior.

2605.13503May 13, 2026Edwige Cyffers, Juba Ziani

Phantom Force: Injecting Adversarial Tactile Perceptions into Embodied Intelligence via EMI

This paper reveals how electromagnetic interference can inject "phantom forces" into robot tactile sensors, severely compromising embodied intelligence.

2605.13492May 13, 2026Zirui Kong, Youqian Zhang, Sze Yiu Chau

Sleeper Channels and Provenance Gates: Persistent Prompt Injection in Always-on Autonomous AI Agents

Persistent prompt injection in always-on AI agents via 'sleeper channels' is identified, and a tiered defense with provenance gates is proposed.

2605.13471May 13, 2026Narek Maloyan, Dmitry Namiot

Model-Agnostic Lifelong LLM Safety via Externalized Attack-Defense Co-Evolution

EvoSafety introduces a novel framework for lifelong, model-agnostic LLM safety via externalized attack-defense co-evolution to counter adversarial prompts.

2605.13411May 13, 2026Xiaozhe Zhang, Chaozhuo Li, Hui Liu +4

Inducing Overthink: Hierarchical Genetic Algorithm-based DoS Attack on Black-Box Large Language Reasoning Models

A hierarchical genetic algorithm can induce "overthink" in black-box LLMs, creating DoS attacks by significantly increasing response length and resource consumption.

2605.13338May 13, 2026Shuqiang Wang, Wei Cao, Jiaqi Weng +4

Context-Aware Web Attack Detection in Open-Source SIEM Systems via MITRE ATT&CK-Enriched Behavioral Profiling

Smart-SIEM enhances open-source SIEMs with an AI module for context-aware web attack detection using behavioral profiling and MITRE ATT&CK.

2605.13337May 13, 2026Badr Alboushy, Assef Jafar, Mohamad Aljnidi +2

Automatic Detection of Reference Counting Bugs in Linux Kernel Drivers

DrvHorn automatically detects reference counting bugs in Linux kernel drivers, finding 545 bugs (424 new) with a low false positive rate.

2605.13246May 13, 2026Joe Hattori, Naoki Kobayashi, Ken Sakayori

Backdoor Channels Hidden in Latent Space: Cryptographic Undetectability in Modern Neural Networks

This paper shows how to create cryptographically undetectable backdoors in modern neural networks by exploiting latent space geometry, resisting current defenses.

2605.13214May 13, 2026Marte Eggen, Eirik Reiestad, Kristian Gjøsteen +1

PoisonCap: Efficient Hierarchical Temporal Safety for CHERI

PoisonCap enhances CHERI systems with strict use-after-free and initialization safety using a novel 'poison' capability format, without performance overhead.

2605.13210May 13, 2026Yuecheng Wang, Jonathan Woodruff, Alfredo Mazzinghi +5

LoREnc: Low-Rank Encryption for Securing Foundation Models and LoRA Adapters

LoREnc is a training-free framework that secures foundation models and LoRA adapters against IP leakage and model recovery attacks with minimal overhead.

2605.13163May 13, 2026Beomjin Ahn, Jungmin Kwon, Chanyong Jung +1

Code-Centric Detection of Vulnerability-Fixing Commits: A Unified Benchmark and Empirical Study

This study finds code language models struggle to detect vulnerability-fixing commits without commit messages, lacking transferable security understanding from code changes alone.

2605.13138May 13, 2026Nils Loose, Joseph Bienhüls, Kristoffer Hempel +2

Extending Blockchain Untraceability with Plausible Deniability

This paper introduces Deniable Covert Asset Transfer (DCAT) to make blockchain transactions untraceable by blending them into common DeFi MEV activities.

2605.13132May 13, 2026Eunchan Park, Kyonghwa Song, Won Hoi Kim +2

Security Incentivization: An Empirical Study of how Micropayments Impact Code Security

This study shows that team-level incentives tied to automated security metrics significantly improve code security in development teams.

2605.13100May 13, 2026Stefan Rass, Martin Pinzger, Rainer W. Alexandrowicz +5

TextSeal: A Localized LLM Watermark for Provenance & Distillation Protection

TextSeal is a new LLM watermark using dual-key generation and multi-region localization for robust, distortion-free detection and distillation protection.

2605.12456May 12, 2026Tom Sander, Hongyan Chang, Tomáš Souček +10

Attacks and Mitigations for Distributed Governance of Agentic AI under Byzantine Adversaries

This paper analyzes attacks on agentic AI governance from compromised centralized providers and proposes Byzantine-resilient, monitoring, and auditing solutions.

2605.12364May 12, 2026Matthew D. Laws, Alina Oprea, Cristina Nita-Rotaru

Reconstruction of Personally Identifiable Information from Supervised Finetuned Models

This paper reveals that PII can be reconstructed from supervised finetuned LLMs, proposing COVA to enhance reconstruction under prefix attacks.

2605.12264May 12, 2026Sae Furukawa, Alina Oprea

No More, No Less: Task Alignment in Terminal Agents

A new benchmark, TAB, reveals terminal agents struggle with selectively following relevant instructions while ignoring distractors, highlighting a gap in task alignment.

2605.12233May 12, 2026Sina Mavali, David Pape, Jonathan Evertz +5
Page 1 of 25Next

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.