ArXiv TLDR

Spend Less, Fit Better: Budget-Efficient Scaling Law Fitting via Active Experiment Selection

🐦 Tweet
2604.22753

Sijie Li, Shanda Li, Haowei Lin, Weiwei Sun, Ameet Talwalkar + 1 more

cs.LG

TLDR

This paper introduces a budget-efficient method for fitting scaling laws by actively selecting experiments, significantly reducing costs.

Key contributions

  • Formulates scaling-law fitting as budget-aware sequential experimental design.
  • Proposes an uncertainty-aware method for active experiment selection.
  • Outperforms classical baselines on diverse scaling-law tasks.
  • Achieves near full-set performance with only 10% of the budget.

Why it matters

Fitting scaling laws for large models is extremely costly. This paper offers a method to drastically reduce these costs, making multi-million-dollar training runs more budget-efficient. It enables more accessible and efficient planning for large-scale AI development.

Original Abstract

Scaling laws are used to plan multi-million-dollar training runs, but fitting those laws can itself cost millions. In modern large-scale workflows, assembling a sufficiently informative set of pilot experiments is already a major budget-allocation problem rather than a routine preprocessing step. We formulate scaling-law fitting as budget-aware sequential experimental design: given a finite pool of runnable experiments with heterogeneous costs, choose which runs to execute so as to maximize extrapolation accuracy in a high-cost target region. We then propose an uncertainty-aware method for sequentially allocating experimental budget toward the runs most useful for target-region extrapolation. Across a diverse benchmark of scaling-law tasks, our method consistently outperforms classical design-based baselines, and often approaches the performance of fitting on the full experimental set while using only about 10% of the total training budget. Our code is available at https://github.com/PlanarG/active-sl.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.