ArXiv TLDR

A Survey on Split Learning for LLM Fine-Tuning: Models, Systems, and Privacy Optimizations

🐦 Tweet
2604.24468

Zihan Liu, Yizhen Wang, Rui Wang, Xiu Tang, Sai Wu

cs.CRcs.CLcs.DCcs.LG

TLDR

This survey reviews split learning for LLM fine-tuning, detailing model, system, and privacy optimizations for secure, collaborative adaptation.

Key contributions

  • Presents the first extensive survey on split learning for LLM fine-tuning.
  • Classifies and critiques diverse approaches across model-level optimization.
  • Reviews system-level efficiency techniques for split LLM fine-tuning.
  • Examines privacy preservation methods in collaborative LLM adaptation.

Why it matters

Fine-tuning LLMs is costly and has privacy risks. Split learning offers a solution for resource-constrained organizations to adapt LLMs securely. This survey provides a crucial, structured overview, guiding future research in scalable and secure collaborative LLM adaptation.

Original Abstract

Fine-tuning unlocks large language models (LLMs) for specialized applications, but its high computational cost often puts it out of reach for resource-constrained organizations. While cloud platforms could provide the needed resources, data privacy concerns make sharing sensitive information with third parties risky. A promising solution is split learning for LLM fine-tuning, which divides the model between clients and a server, allowing collaborative and secure training through exchanged intermediate data, thus enabling resource-constrained participants to adapt LLMs safely. % In light of this, a growing body of literature has emerged to advance this paradigm, introducing varied model methods, system optimizations, and privacy defense-attack techniques for split learning. To bring clarity and direction to the field, a comprehensive survey is needed to classify, compare, and critique these diverse approaches. This paper fills the gap by presenting the first extensive survey dedicated to split learning for LLM fine-tuning. We propose a unified, fine-grained training pipeline to pinpoint key operational components and conduct a systematic review of state-of-the-art work across three core dimensions: model-level optimization, system-level efficiency, and privacy preservation. Through this structured taxonomy, we establish a foundation for advancing scalable, robust, and secure collaborative LLM adaptation.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.