ArXiv TLDR

Cut Your Losses! Learning to Prune Paths Early for Efficient Parallel Reasoning

🐦 Tweet
2604.16029

Jiaxi Bi, Tongxu Luo, Wenyu Du, Zhengyang Tang, Benyou Wang

cs.CLcs.LG

TLDR

STOP introduces a learnable internal path pruning method for Large Reasoning Models, significantly improving efficiency and accuracy in parallel reasoning.

Key contributions

  • Proposes the first systematic taxonomy for path pruning in parallel reasoning.
  • Introduces STOP, a novel learnable internal method for early path pruning in LRMs.
  • STOP outperforms baselines in effectiveness and efficiency across LRMs (1.5B-20B).
  • Boosts GPT-OSS-20B accuracy on AIME25 from 84% to 90% under fixed compute budgets.

Why it matters

Parallel reasoning in LRMs is costly due to early errors. This paper introduces STOP, a novel pruning method that significantly improves efficiency and accuracy. It offers a systematic framework and practical guidelines for more effective LRM deployment.

Original Abstract

Parallel reasoning enhances Large Reasoning Models (LRMs) but incurs prohibitive costs due to futile paths caused by early errors. To mitigate this, path pruning at the prefix level is essential, yet existing research remains fragmented without a standardized framework. In this work, we propose the first systematic taxonomy of path pruning, categorizing methods by their signal source (internal vs. external) and learnability (learnable vs. non-learnable). This classification reveals the unexplored potential of learnable internal methods, motivating our proposal of STOP (Super TOken for Pruning). Extensive evaluations across LRMs ranging from 1.5B to 20B parameters demonstrate that STOP achieves superior effectiveness and efficiency compared to existing baselines. Furthermore, we rigorously validate the scalability of STOP under varying compute budgets - for instance, boosting GPT-OSS-20B accuracy on AIME25 from 84% to nearly 90% under fixed compute budgets. Finally, we distill our findings into formalized empirical guidelines to facilitate optimal real-world deployment. Code, data and models are available at https://bijiaxihh.github.io/STOP

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.