Direct From Darwin: Deriving Advanced Optimizers From Evolutionary First Principles
TLDR
This paper unifies Fisher's and Wright's evolutionary theories to derive advanced gradient optimizers, showing many existing algorithms are evolutionarily compliant.
Key contributions
- Derives advanced gradient-based optimizers directly from evolutionary first principles.
- Introduces Darwinian Lineage Simulations (DLS) to formally unify Fisher's and Wright's evolutionary theories.
- Proves many optimizers (SGD, NGD, Damped Newton) are evolutionarily compatible with DLS noise.
- Shows how to adapt the state-of-the-art Adam optimizer for evolutionary compliance.
Why it matters
This work provides a rigorous evolutionary foundation for advanced gradient-based optimizers, unifying historical theories. It allows for scientifically valid in silico simulations of Darwinian evolution and brings state-of-the-art algorithms into compliance.
Original Abstract
Evolutionary computation has long promised to deliver both high-performance optimization tools as well as rigorous scientific simulations of Darwinian evolution. However, modern algorithms frequently abandon evolutionary fidelity for physics-inspired heuristics or superficial biological metaphors. This paper derives a suite of advanced gradient-based optimization algorithms directly from evolutionary first principles. We introduce Darwinian Lineage Simulations (DLS) to prove that, in an asexual context, Fisher's and Wright's historically opposed views of evolution are actually formally equivalent. This unification requires carefully partitioning Fisher's deterministically-evolving total population into Wright's randomly-drifting sub-populations. We prove that proper bookkeeping requires introducing a specific kind of structured noise (the DLS noise relation). Crucially, however, any bookkeeping choices which satisfy this relation will result in a faithful simulation of evolution. Using this vast representational freedom, we prove that a broad family of battle-tested optimization algorithms are already perfectly compatible with evolutionary dynamics. These include: Stochastic Gradient Descent, Natural Gradient Descent, and the Damped Newton's method among many others. By simply adding DLS noise (i.e., evolutionarily faithful genetic drift), these algorithms become scientifically valid in silico simulations of Darwinian evolution. Finally, we demonstrate that even the state-of-the-art Adam optimizer can be brought into evolutionary compliance through a minor mathematical surgery.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.