ArXiv TLDR

Learning to Reason with Insight for Informal Theorem Proving

🐦 Tweet
2604.16278

Yunhe Li, Hao Shi, Bowen Deng, Wei Wang, Mengzhe Ruan + 6 more

cs.AIcs.CLcs.LG

TLDR

A new framework and dataset enable LLMs to perform insightful informal theorem proving by learning to recognize core techniques.

Key contributions

  • Identifies lack of insight as a key bottleneck in LLM-based informal theorem proving.
  • Proposes a novel framework to cultivate insightful reasoning in large language models.
  • Introduces DeepInsightTheorem, a hierarchical dataset structuring proofs with core techniques.
  • Develops a Progressive Multi-Stage SFT strategy mimicking human learning for insightful thinking.

Why it matters

This paper addresses a critical limitation of LLMs in mathematical reasoning by focusing on "insight." By teaching models to identify and apply core techniques, it significantly improves their ability to solve complex informal proofs. This approach could lead to more robust and human-like AI theorem provers.

Original Abstract

Although most of the automated theorem-proving approaches depend on formal proof systems, informal theorem proving can align better with large language models' (LLMs) strength in natural language processing. In this work, we identify a primary bottleneck in informal theorem proving as a lack of insight, namely the difficulty of recognizing the core techniques required to solve complex problems. To address this, we propose a novel framework designed to cultivate this essential reasoning skill and enable LLMs to perform insightful reasoning. We propose $\mathtt{DeepInsightTheorem}$, a hierarchical dataset that structures informal proofs by explicitly extracting core techniques and proof sketches alongside the final proof. To fully exploit this dataset, we design a Progressive Multi-Stage SFT strategy that mimics the human learning process, guiding the model from basic proof writing to insightful thinking. Our experiments on challenging mathematical benchmarks demonstrate that this insight-aware generation strategy significantly outperforms baselines. These results demonstrate that teaching models to identify and apply core techniques can substantially improve their mathematical reasoning.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.