ArXiv TLDR

Distributed Online Convex Optimization with Compressed Communication: Optimal Regret and Applications

🐦 Tweet
2604.09276

Sifan Yang, Dan-Yue Li, Lijun Zhang

cs.LG

TLDR

This paper introduces optimal algorithms and lower bounds for distributed online convex optimization with compressed communication, reducing communication costs.

Key contributions

  • Established optimal lower bounds for D-OCO with compressed communication: Ω(δ⁻¹/²√T) for convex and Ω(δ⁻¹logT) for strongly convex.
  • Proposed an optimal algorithm achieving matching regret bounds for both convex and strongly convex loss functions.
  • Integrated error feedback into FTRL and an online compression strategy to handle various compression errors.
  • Provided first convergence guarantees for distributed non-smooth optimization with compressed communication and domain constraints.

Why it matters

Distributed online convex optimization faces significant communication bottlenecks in large-scale applications. This paper provides a foundational framework for addressing this by introducing compressed communication. It offers optimal algorithms and theoretical guarantees, paving the way for more efficient and scalable distributed learning systems.

Original Abstract

Distributed online convex optimization (D-OCO) is a powerful paradigm for modeling distributed scenarios with streaming data. However, the communication cost between local learners and the central server is substantial in large-scale applications. To alleviate this bottleneck, we initiate the study of D-OCO with compressed communication. Firstly, to quantify the compression impact, we establish the $Ω(δ^{-1/2}\sqrt{T})$ and $Ω(δ^{-1}\log{T})$ lower bounds for convex and strongly convex loss functions, respectively, where $δ\in (0,1]$ is the compression ratio. Secondly, we propose an optimal algorithm, which enjoys regret bounds of $O(δ^{-1/2}\sqrt{T})$ and $O(δ^{-1} \log T)$ for convex and strongly convex loss functions, respectively. Our method incorporates the error feedback mechanism into the Follow-the-Regularized-Leader framework to address the coupling between the compression error and the projection error. Furthermore, we employ the online compression strategy to mitigate the accumulated error arising from the bidirectional compression. Our online method has great generality, and can be extended to the offline stochastic setting via online-to-batch conversion. We establish convergence rates of $O(δ^{-1/2}T^{-1/2})$ and $O(δ^{-1} T^{-1})$ for convex and strongly convex loss functions, respectively, providing the first guarantees for distributed non-smooth optimization with compressed communication and domain constraints.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.