Fine-tuning Factor Augmented Neural Lasso for Heterogeneous Environments
Jinhang Chai, Jianqing Fan, Cheng Gao, Qishuo Yin
TLDR
Introduces fine-tuning FAN-Lasso, a transfer learning framework for high-dimensional nonparametric regression with variable selection in heterogeneous environments.
Key contributions
- Introduces fine-tuning FAN-Lasso for high-dimensional nonparametric regression with variable selection.
- Manages covariate and posterior shifts using a low-rank factor structure and residual fine-tuning.
- Derives minimax-optimal excess risk bounds, showing when fine-tuning accelerates learning.
- Empirically validates performance, outperforming baselines even with limited target data.
Why it matters
This paper addresses a gap in fine-tuning theory for high-dimensional nonparametric settings with variable selection. It introduces a robust framework, FAN-Lasso, that performs well in diverse heterogeneous environments and offers strong theoretical guarantees, advancing transfer learning applications.
Original Abstract
Fine-tuning is a widely used strategy for adapting pre-trained models to new tasks, yet its methodology and theoretical properties in high-dimensional nonparametric settings with variable selection have not yet been developed. This paper introduces the fine-tuning factor augmented neural Lasso (FAN-Lasso), a transfer learning framework for high-dimensional nonparametric regression with variable selection that simultaneously handles covariate and posterior shifts. We use a low-rank factor structure to manage high-dimensional dependent covariates and propose a novel residual fine-tuning decomposition in which the target function is expressed as a transformation of a frozen source function and other variables to achieve transfer learning and nonparametric variable selection. This augmented feature from the source predictor allows for the transfer of knowledge to the target domain and reduces model complexity there. We derive minimax-optimal excess risk bounds for the fine-tuning FAN-Lasso, characterizing the precise conditions, in terms of relative sample sizes and function complexities, under which fine-tuning yields statistical acceleration over single-task learning. The proposed framework also provides a theoretical perspective on parameter-efficient fine-tuning methods. Extensive numerical experiments across diverse covariate- and posterior-shift scenarios demonstrate that the fine-tuning FAN-Lasso consistently outperforms standard baselines and achieves near-oracle performance even under severe target sample size constraints, empirically validating the derived rates.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.