ArXiv TLDR

Towards Privacy-Preserving Large Language Model: Text-free Inference Through Alignment and Adaptation

🐦 Tweet
2604.06831

Jeongho Yoon, Chanhee Park, Yongchan Chun, Hyeonseok Moon, Heuiseok Lim

cs.CRcs.AI

TLDR

PPFT enables privacy-preserving LLM inference by transmitting k-pooled prompt embeddings instead of raw text, balancing privacy and utility.

Key contributions

  • Introduces Privacy-Preserving Fine-Tuning (PPFT) for text-free LLM inference.
  • Client transmits k-pooled prompt embeddings, not raw text, to the server.
  • Two-stage training: encoder/projection alignment, then fine-tuning with noise-injected embeddings.
  • Maintains competitive performance with minimal degradation compared to baselines.

Why it matters

This paper addresses the critical privacy risks of sending raw text to LLMs, which often contain sensitive information. It overcomes the privacy-efficiency trade-off of prior methods. PPFT enables secure and efficient LLM services by allowing text-free inference and adaptation.

Original Abstract

Current LLM-based services typically require users to submit raw text regardless of its sensitivity. While intuitive, such practice introduces substantial privacy risks, as unauthorized access may expose personal, medical, or legal information. Although prior defenses strived to mitigate these risks, they often incur substantial computational overhead and degrade model performance. To overcome this privacy-efficiency trade-off, we introduce Privacy-Preserving Fine-Tuning (PPFT), a novel training pipeline that eliminates the need for transmitting raw prompt text while maintaining a favorable balance between privacy preservation and model utility for both clients and service providers. Our approach operates in two stages: first, we train a client-side encoder together with a server-side projection module and LLM, enabling the server to condition on k-pooled prompt embeddings instead of raw text; second, we fine-tune the projection module and LLM on private, domain-specific data using noise-injected embeddings, allowing effective adaptation without exposing plain text prompts and requiring access to the decoder's internal parameters. Extensive experiments on domain-specific and general benchmarks demonstrate that PPFT achieves a striking balance between privacy and utility, maintaining competitive performance with minimal degradation compared to noise-free upper bounds.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.