Harnessing Agentic Evolution
Jiayi Zhang, Yongfeng Gu, Jianhao Ruan, Maojia Song, Yiran Peng + 8 more
TLDR
AEvo introduces a meta-editing framework that steers agentic evolution by dynamically revising the evolution process, outperforming existing methods.
Key contributions
- Formulates agentic evolution as an interactive environment with process-level state.
- Introduces AEvo, a meta-agent that edits the evolution procedure/agent context, not candidates.
- Provides a unified interface to steer both procedure-based and agent-based evolution.
- Achieves state-of-the-art performance on agentic, reasoning, and open-ended optimization tasks.
Why it matters
Agentic evolution struggles with rigidity or drift, limiting long-term effectiveness. AEvo addresses this by enabling dynamic self-correction of the evolution process itself. This allows for more stable and efficient long-horizon search, significantly improving performance across various complex tasks.
Original Abstract
Agentic evolution has emerged as a powerful paradigm for improving programs, workflows, and scientific solutions by iteratively generating candidates, evaluating them, and using feedback to guide future search. However, existing methods are typically instantiated either as fixed hand-designed procedures that are modular but rigid, or as general-purpose agents that flexibly integrate feedback but can drift in long-horizon evolution. Both forms accumulate rich evidence over time, including candidates, feedback, traces, and failures, yet lack a stable interface for organizing this evidence and revising the mechanism that drives future evolution. We address this limitation by formulating agentic evolution as an interactive environment, where the accumulated evolution context serves as a process-level state. We introduce AEvo, a harnessed meta-editing framework in which a meta-agent observes this state and acts not by directly proposing the next candidate, but by editing the procedure or agent context that controls future evolution. This unified interface enables AEvo to steer both procedure-based and agent-based evolution, making accumulated evidence actionable for long-horizon search. Empirical evaluations on agentic and reasoning benchmarks show that AEvo outperforms five evolution baselines, achieving a 26 relative improvement over the strongest baseline. Across three open-ended optimization tasks, AEvo further outperforms four evolution baselines and achieves state-of-the-art performance under the same iteration budget.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.