Order-Optimal Sequential 1-Bit Mean Estimation in General Tail Regimes
TLDR
This paper introduces an order-optimal adaptive 1-bit mean estimator for general tail regimes, achieving near-minimax sample complexity.
Key contributions
- Proposes a novel adaptive 1-bit mean estimator based solely on randomized threshold queries.
- Achieves order-optimal sample complexity for general tail regimes (bounded k-th central moments).
- Establishes new information-theoretic lower bounds, demonstrating fundamental limits and a significant adaptivity gap.
Why it matters
This work addresses the critical problem of mean estimation under strict 1-bit communication, vital for resource-constrained settings. By achieving order-optimal sample complexity and establishing fundamental limits, it significantly advances our understanding of 1-bit quantization. It also highlights the necessity of adaptive approaches for efficiency.
Original Abstract
In this paper, we study the problem of mean estimation under strict 1-bit communication constraints. We propose a novel adaptive mean estimator based solely on randomized threshold queries, where each 1-bit outcome indicates whether a given sample exceeds a sequentially chosen threshold. Our estimator is $(ε, δ)$-PAC for any distribution with a bounded mean $μ\in [-λ, λ]$ and a bounded $k$-th central moment $\mathbb{E}[|X-μ|^k] \le σ^k$ for any fixed $k > 1$. Crucially, our sample complexity is order-optimal in all such tail regimes, i.e., for every such $k$ value. For $k \neq 2$, our estimator's sample complexity matches the unquantized minimax lower bounds plus an unavoidable $O(\log(λ/σ))$ localization cost. For the finite-variance case ($k=2$), our estimator's sample complexity has an extra multiplicative $O(\log(σ/ε))$ penalty, and we establish a novel information-theoretic lower bound showing that this penalty is a fundamental limit of 1-bit quantization. We also establish a significant adaptivity gap: for both threshold queries and more general interval queries, the sample complexity of any non-adaptive estimator must scale linearly with the search space parameter $λ/σ$, rendering it vastly less sample efficient than our adaptive approach. Finally, we present algorithmic variants that (i) handle an unknown sampling budget, (ii) adapt to an unknown scale parameter~$σ$ given (possibly loose) bounds, and (iii) require only two stages of adaptivity at the expense of more complicated general 1-bit queries.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.