ArXiv TLDR

DP-FlogTinyLLM: Differentially private federated log anomaly detection using Tiny LLMs

🐦 Tweet
2604.19118

Isaiah Thompson, Tanmay Sen, Ritwik Bhattacharya

cs.CRcs.AI

TLDR

DP-FLogTinyLLM offers a differentially private federated framework for log anomaly detection using Tiny LLMs, enabling collaborative learning without sharing raw data.

Key contributions

  • Proposes DP-FLogTinyLLM, a federated framework for private log anomaly detection.
  • Integrates federated optimization with differential privacy to protect raw log data.
  • Employs LoRA for efficient fine-tuning of Tiny LLMs on resource-constrained clients.
  • Achieves high precision and F1-score, outperforming existing federated baselines.

Why it matters

Centralized log anomaly detection is impractical due to privacy concerns. This paper introduces a privacy-preserving federated framework that allows collaborative learning without sharing sensitive log data. It enables effective anomaly detection in distributed, resource-constrained environments.

Original Abstract

Modern distributed systems generate massive volumes of log data that are critical for detecting anomalies and cyber threats. However, in real world settings, these logs are often distributed across multiple organizations and cannot be centralized due to privacy and security constraints. Existing log anomaly detection methods, including recent large language model (LLM) based approaches, largely rely on centralized training and are not suitable for such environments. In this paper, we propose DP-FLogTinyLLM, a privacy preserving federated framework for log anomaly detection using parameter efficient LLMs. Our approach enables collaborative learning without sharing raw log data by integrating federated optimization with differential privacy. To ensure scalability in resource constrained environments, we employ low rank adaptation (LoRA) for efficient fine tuning of Tiny LLMs at each client. Empirical results on the Thunderbird and BGL datasets show that the proposed framework matches the performance of centralized LLM based methods, while incurring additional computational overhead due to privacy mechanisms. Compared to existing federated baselines, DP-FLogTinyLLM consistently achieves higher precision and F1-score, with particularly strong gains on the Thunderbird dataset, highlighting its effectiveness in detecting anomalies while minimizing false positives.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.