ArXiv TLDR

When Are Trade-Off Functions Testable from Finite Samples?

🐦 Tweet
2605.10774

Kaining Shi, Qiaosen Wang, Cong Ma

math.STstat.ML

TLDR

This paper identifies conditions under which trade-off functions for binary testing are testable from finite samples, crucial for statistical inference.

Key contributions

  • Identifies a sharp condition for finite-sample testability of trade-off functions.
  • Shows finite VC dimension of the rejection region class is necessary and sufficient.
  • Develops a test with nonasymptotic error guarantees and simultaneous confidence bands.
  • Extends the framework to approximate attainability, covering log-concave distributions.

Why it matters

This paper solves a fundamental problem in statistical inference by identifying conditions for testing trade-off functions from finite samples. It provides nonasymptotic error guarantees and confidence bands, enabling robust analysis in settings where previous methods were impossible. This advances statistical testing for binary classification.

Original Abstract

We study finite-sample inference for the trade-off function of two unknown probability distributions, the function that traces the optimal type I/type II error frontier in binary testing. Given samples from distributions $P$ and $Q$, we consider the problem of testing whether their trade-off function lies above a benchmark curve $f_0$ or falls below a weaker benchmark $f_1$. Without structural restrictions, this problem is impossible uniformly over nonparametric classes. We identify a sharp condition under which it becomes possible. The key structural assumption is that the Neyman--Pearson rejection regions for $(P,Q)$ are attainable, up to null sets, by a prescribed class $S$ of measurable sets. Within this exact attainability framework, finite Vapnik--Chervonenkis dimension of $S$ is both sufficient and necessary for nontrivial finite-sample testing. We construct a test with nonasymptotic error guarantees: type I error control is valid without assuming attainability, while power holds uniformly over attainable alternatives satisfying an explicit separation condition. By inverting the test, we also obtain simultaneous confidence bands for the whole trade-off curve. Finally, we study the sharpness and robustness of the procedure. In the monotone likelihood-ratio model, we derive local separation rates and prove matching lower bounds up to logarithmic factors. We also allow approximate, rather than exact, attainability; this extension yields finite-sample guarantees for univariate log-concave distributions by approximating their rejection regions with unions of intervals.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.