Machine Learning
Papers on learning algorithms, neural networks, deep learning, and optimization.
cs.LG · 1353 papersScaling Laws and Tradeoffs in Recurrent Networks of Expressive Neurons
ELM Networks demonstrate optimal resource allocation in recurrent networks, favoring more complex neurons as scale increases, challenging simple-unit defaults.
Approximation Theory of Laplacian-Based Neural Operators for Reaction-Diffusion System
This paper shows Laplacian-based neural operators efficiently approximate reaction-diffusion systems with polynomial complexity.
SkillSafetyBench: Evaluating Agent Safety under Skill-Facing Attack Surfaces
SkillSafetyBench evaluates how reusable skills in LLM agents create new attack surfaces, revealing vulnerabilities beyond model-level alignment.
Random-Set Graph Neural Networks
This paper introduces Random-Set Graph Neural Networks (RS-GNNs) to model node-level epistemic uncertainty using belief functions for improved predictions.
QDSB: Quantized Diffusion Schrödinger Bridges
QDSB introduces quantized diffusion Schrödinger bridges to efficiently learn generative models from unpaired data, significantly reducing training time.
LOFT: Low-Rank Orthogonal Fine-Tuning via Task-Aware Support Selection
LOFT is a low-rank orthogonal fine-tuning framework that separates adaptation subspace and transformation, improving PEFT efficiency via task-aware support selection.
Multi-Timescale Conductance Spiking Networks: A Sparse, Gradient-Trainable Framework with Rich Firing Dynamics for Enhanced Temporal Processing
Multi-timescale conductance SNNs offer rich dynamics, sparse activity, and direct gradient training, outperforming SOTA in temporal processing.
One-Step Generative Modeling via Wasserstein Gradient Flows
W-Flow introduces a novel one-step generative model using Wasserstein gradient flows, achieving state-of-the-art image generation 100x faster than diffusion models.
Persona-Conditioned Adversarial Prompting: Multi-Identity Red-Teaming for Adversarial Discovery and Mitigation
PCAP uses diverse personas for red-teaming LLMs, significantly boosting attack success and generating robust defense data for improved safety.
Learning U-Statistics with Active Inference
An active inference framework for U-statistics improves estimation efficiency by selectively querying informative labels under budget constraints.
Exact Stiefel Optimization for Probabilistic PLS: Closed-Form Updates, Error Bounds, and Calibrated Uncertainty
Introduces an end-to-end framework for Probabilistic PLS using exact Stiefel optimization, offering calibrated uncertainty and improved accuracy.
EpiCastBench: Datasets and Benchmarks for Multivariate Epidemic Forecasting
EpiCastBench introduces 40 diverse multivariate epidemic datasets and a standardized benchmark for evaluating forecasting models.
A Composite Activation Function for Learning Stable Binary Representations
Introduces HTAF, a smooth composite activation function enabling stable gradient-based training of neural networks with binary representations.
The Evaluation Differential: When Frontier AI Models Recognise They Are Being Tested
This paper introduces the Evaluation Differential, showing AI models behave differently when tested, challenging safety claims from current evaluations.
LPDP: Inference-Time Reward Control for Variable-Length DNA Generation with Edit Flows
LPDP enables training-free, inference-time reward control for variable-length DNA generation using biologically plausible edit flows.
Beyond Manual Curation: Augmenting Targeted Protein Degradation Databases via Agentic Literature Extraction Workflows
A new expert-in-the-loop LLM workflow automates targeted protein degradation data extraction, significantly expanding databases with high accuracy.
Decomposing Evolutionary Mixture-of-LoRA Architectures: The Routing Lever, the Lifecycle Penalty, and a Substrate-Conditional Boundary
This paper decomposes an evolutionary Mixture-of-LoRA system, finding that router improvements, not the evolutionary lifecycle, drive performance gains.
ELF: Embedded Language Flows
ELF proposes a continuous diffusion model for language, leveraging flow matching in embedding space to achieve superior generation quality with fewer steps.
Variational Inference for Lévy Process-Driven SDEs via Neural Tilting
This paper introduces a neural exponential tilting framework for variational inference in Lévy-driven SDEs, addressing challenges in modeling extreme events.
DECO: Sparse Mixture-of-Experts with Dense-Comparable Performance on End-Side Devices
DECO is a sparse MoE model matching dense performance on end-side devices, offering 3x speedup and reduced storage overhead.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.