Memory Transfer Learning: How Memories are Transferred Across Domains in Coding Agents
Kangsan Kim, Minki Kang, Taeil Kim, Yanlai Yang, Mengye Ren + 1 more
TLDR
Memory Transfer Learning (MTL) allows coding agents to use cross-domain memories, boosting performance by transferring meta-knowledge like validation routines.
Key contributions
- Introduces Memory Transfer Learning (MTL) for coding agents using a unified memory pool from heterogeneous domains.
- Cross-domain memory improves average performance by 3.7%, mainly by transferring meta-knowledge (e.g., validation routines).
- Abstraction level dictates transferability: high-level insights generalize well, low-level traces cause negative transfer.
- Transfer effectiveness scales with memory pool size and works across different models.
Why it matters
This paper addresses a key limitation in coding agents by enabling memory utilization across diverse task domains. It provides empirical design principles for effectively transferring knowledge, moving beyond single-domain memory silos. This work is crucial for developing more adaptable and efficient AI coding assistants.
Original Abstract
Memory-based self-evolution has emerged as a promising paradigm for coding agents. However, existing approaches typically restrict memory utilization to homogeneous task domains, failing to leverage the shared infrastructural foundations, such as runtime environments and programming languages, that exist across diverse real-world coding problems. To address this limitation, we investigate \textbf{Memory Transfer Learning} (MTL) by harnessing a unified memory pool from heterogeneous domains. We evaluate performance across 6 coding benchmarks using four memory representations, ranging from concrete traces to abstract insights. Our experiments demonstrate that cross-domain memory improves average performance by 3.7\%, primarily by transferring meta-knowledge, such as validation routines, rather than task-specific code. Importantly, we find that abstraction dictates transferability; high-level insights generalize well, whereas low-level traces often induce negative transfer due to excessive specificity. Furthermore, we show that transfer effectiveness scales with the size of the memory pool, and memory can be transferred even between different models. Our work establishes empirical design principles for expanding memory utilization beyond single-domain silos. Project page: https://memorytransfer.github.io/
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.