ArXiv TLDR

Fine-tuning Factor Augmented Neural Lasso for Heterogeneous Environments

🐦 Tweet
2604.12288

Jinhang Chai, Jianqing Fan, Cheng Gao, Qishuo Yin

stat.MLcs.LGstat.ME

TLDR

Introduces fine-tuning FAN-Lasso, a transfer learning framework for high-dimensional nonparametric regression with variable selection in heterogeneous environments.

Key contributions

  • Introduces fine-tuning FAN-Lasso for high-dimensional nonparametric regression with variable selection.
  • Manages covariate and posterior shifts using a low-rank factor structure and residual fine-tuning.
  • Derives minimax-optimal excess risk bounds, showing when fine-tuning accelerates learning.
  • Empirically validates performance, outperforming baselines even with limited target data.

Why it matters

This paper addresses a gap in fine-tuning theory for high-dimensional nonparametric settings with variable selection. It introduces a robust framework, FAN-Lasso, that performs well in diverse heterogeneous environments and offers strong theoretical guarantees, advancing transfer learning applications.

Original Abstract

Fine-tuning is a widely used strategy for adapting pre-trained models to new tasks, yet its methodology and theoretical properties in high-dimensional nonparametric settings with variable selection have not yet been developed. This paper introduces the fine-tuning factor augmented neural Lasso (FAN-Lasso), a transfer learning framework for high-dimensional nonparametric regression with variable selection that simultaneously handles covariate and posterior shifts. We use a low-rank factor structure to manage high-dimensional dependent covariates and propose a novel residual fine-tuning decomposition in which the target function is expressed as a transformation of a frozen source function and other variables to achieve transfer learning and nonparametric variable selection. This augmented feature from the source predictor allows for the transfer of knowledge to the target domain and reduces model complexity there. We derive minimax-optimal excess risk bounds for the fine-tuning FAN-Lasso, characterizing the precise conditions, in terms of relative sample sizes and function complexities, under which fine-tuning yields statistical acceleration over single-task learning. The proposed framework also provides a theoretical perspective on parameter-efficient fine-tuning methods. Extensive numerical experiments across diverse covariate- and posterior-shift scenarios demonstrate that the fine-tuning FAN-Lasso consistently outperforms standard baselines and achieves near-oracle performance even under severe target sample size constraints, empirically validating the derived rates.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.