Why Architecture Choice Matters in Symbolic Regression
TLDR
Architecture choice, not just expressiveness, critically determines the success of gradient-based symbolic regression in recovering target formulas.
Key contributions
- Tree architecture choice critically impacts symbolic regression recovery, with success rates varying from 0% to 100%.
- More expressive tree structures can surprisingly fail on targets that less expressive ones reliably solve.
- Operator choice and its gradient profile significantly influence which target formulas are recovered.
- Balanced (non-chain) tree shapes are consistently not recovered by gradient-based methods.
Why it matters
This paper reveals that the optimization landscape, shaped by architecture choices, is more crucial than expressiveness for gradient-based symbolic regression. It challenges conventional wisdom and provides critical insights for designing more effective and reliable symbolic regression algorithms.
Original Abstract
Symbolic regression discovers mathematical formulas from data. Some methods fix a tree of operators, assign learnable weights, and train by gradient descent. The tree's structure, which determines what operators and variables appear at each position, is chosen once and applied to every target. This paper tests whether that choice affects which targets are actually recovered. Three structures are compared, all sharing the same operator and target language but differing in how variables enter the tree; one is strictly more expressive. Across over 12,700 training runs, one structure recovers a target at 100% while another scores 0%, and the ranking reverses on a different target. Expressiveness guarantees that a solution exists in the search space, but not that gradient descent finds it: the most expressive structure fails on targets that a restricted alternative solves reliably. Switching the operator changes which targets succeed; reversing its gradient profile collapses recovery entirely. Balanced (non-chain) tree shapes are never recovered. These findings show that the optimization landscape, not expressiveness alone, determines what gradient-based symbolic regression recovers.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.