LeGo-Code: Can Modular Curriculum Learning Advance Complex Code Generation? Insights from Text-to-SQL
Salmane Chafik, Saad Ezzini, Ismail Berrada
TLDR
LeGo-Code uses a Modular Adapter Composition strategy to improve LLM performance on complex Text-to-SQL tasks by sequentially training complexity-specific adapters.
Key contributions
- Identifies that naive curriculum learning fails for complex code generation due to catastrophic forgetting.
- Proposes Modular Adapter Composition (MAC) strategy, training tier-specific adapters incrementally.
- Achieves measurable performance gains on complex Text-to-SQL benchmarks like Spider and BIRD.
- Introduces a flexible, 'Lego-like' architecture for deploying models based on specific schema difficulty.
Why it matters
This paper introduces a novel modular curriculum learning approach that overcomes catastrophic forgetting in LLMs for complex code generation. It offers a flexible architecture, making LLMs more robust and adaptable to real-world, noisy database schemas. This significantly advances Text-to-SQL capabilities.
Original Abstract
Recently, code-oriented large language models (LLMs) have demonstrated strong capabilities in translating natural language into executable code. Text-to-SQL is a significant application of this ability, enabling non-technical users to interact with relational databases using natural language. However, state-of-the-art models continue to struggle with highly complex logic, particularly deeply nested statements involving multiple joins and conditions, as well as with real-world database schemas that are noisy or poorly structured. In this paper, we investigate whether curriculum learning can improve the performance of code-based LLMs on Text-to-SQL tasks. Employing benchmarks including Spider and BIRD, we fine-tune models under different curriculum strategies. Our experiments show that naive curriculum, simply ordering training samples by complexity in a single epoch, fails to surpass standard fine-tuning due to catastrophic forgetting. To overcome this, we propose a Modular Adapter Composition (MAC) strategy. By sequentially training tier-specific adapters on incremental complexity levels (Easy to Extra-Hard), we create a scaffolded learning environment that improves performance on complex queries. Our approach not only produces measurable performance gains on the Spider and BIRD benchmarks but also provides a flexible, "Lego-like" architecture, allowing models to be composed and deployed based on specific schema difficulty requirements. These findings demonstrate that structured, modular learning is a superior alternative to monolithic fine-tuning for mastering the syntax and logic of complex code generation.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.