ArXiv TLDR

Adaptive Norm-Based Regularization for Neural Networks

🐦 Tweet
2605.00171

Muhammad Qasim, Farrukh Javed

stat.MLcs.LGstat.AP

TLDR

This paper introduces adaptive norm-based regularization methods for neural networks that account for input feature covariance and improve predictive performance.

Key contributions

  • Modifies weight decay with a ridge-type L2 penalty incorporating input feature covariance.
  • Combines L1 sparsity with covariance-aware L2 regularization for sparse, structurally informed weights.
  • Improves predictive performance and complexity control, especially with correlated or high-dimensional features.

Why it matters

Standard norm-based regularizers often struggle with correlated or high-dimensional features. This paper introduces adaptive strategies that significantly enhance neural network performance and complexity control by considering input feature dependencies, leading to more robust and accurate models.

Original Abstract

In this paper, we study norm-based regularization methods for neural networks. We compare existing penalization approaches and introduce two regularization strategies that extend classical ridge- and lasso-type penalties to neural network models. The first strategy modifies weight decay by incorporating the covariance structure of the input features into a ridge-type $\ell_2$ penalty, allowing regularization to account for feature dependence. The second combines an $\ell_1$ sparsity penalty with covariance-aware $\ell_2$ regularization, producing neural network weights that are both sparse and structurally informed. Monte Carlo simulations are used to evaluate these methods under different data-generating settings, followed by two real-data applications on building cooling-load prediction and leukemia cell-type classification from high-dimensional gene expression data. Across simulated and real-data examples, the proposed regularizers improve predictive performance on unseen data and provide more effective complexity control than standard norm-based penalties, particularly when features are correlated or high-dimensional.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.