Improving LLM Code Generation via Requirement-Aware Curriculum Reinforcement Learning
Shouyu Yin, Zhao Tian, Junjie Chen, Shikai Guo
TLDR
RECRL improves LLM code generation by using a requirement-aware curriculum reinforcement learning framework for better training efficiency.
Key contributions
- Automatically perceives model-specific requirement difficulty.
- Optimizes challenging requirements to enhance training data utilization.
- Employs an adaptive curriculum sampling strategy for smooth difficulty batches.
- Achieves 1.23%-5.62% Pass@1 improvement on LLM code generation benchmarks.
Why it matters
LLMs struggle with complex code generation. This paper introduces RECRL, a novel framework that addresses key limitations of existing curriculum reinforcement learning methods. By optimizing requirement difficulty and sampling, RECRL significantly boosts LLM code generation performance. This advancement can lead to more efficient software development.
Original Abstract
Code generation, which aims to automatically generate source code from given programming requirements, has the potential to substantially improve software development efficiency. With the rapid advancement of large language models (LLMs), LLM-based code generation has attracted widespread attention from both academia and industry. However, as programming requirements become increasingly complex, existing LLMs still exhibit notable performance limitations. To address this challenge, recent studies have proposed training-based curriculum reinforcement learning (CRL) strategies to improve LLM code generation performance. Despite their effectiveness, existing CRL approaches suffer from several limitations, including misaligned requirement difficulty perception, the absence of requirement difficulty optimization, and suboptimal curriculum sampling strategies. In CRL-based code generation, programming requirements serve as the sole input to the model, making their quality and difficulty critical to training effectiveness. Motivated by insights from software requirements engineering, we propose RECRL, a novel requirement-aware curriculum reinforcement learning framework for enhancing LLM-based code generation. RECRL automatically perceives model-specific requirement difficulty, optimizes challenging requirements to improve training data utilization, and employs an adaptive curriculum sampling strategy to construct training batches with smoothly varying difficulty. Extensive experiments on five state-of-the-art LLMs across five widely-used code generation benchmarks by comparing with five state-of-the-art baselines, demonstrate the significant effectiveness of RECRL. For example, RECRL achieves an average Pass@1 improvement of 1.23%-5.62% over all state-of-the-art baselines.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.