Prediction-powered Inference by Mixture of Experts
Yanwu Gu, Linglong Kong, Dong Xia
TLDR
Introduces a Mixture of Experts (MOE) framework for semi-supervised inference, enhancing Prediction-Powered Inference (PPI) with improved variance reduction.
Key contributions
- Proposes a Mixture of Experts (MOE) framework for semi-supervised inference, building on Prediction-Powered Inference (PPI).
- Achieves variance reduction by optimally combining diverse AI predictors, adapting to their unknown individual performance.
- Offers a 'best-expert guarantee' and benefits from the collective predictive power of multiple models.
- Applicable to mean estimation, linear regression, quantile estimation, and general M-estimation with theoretical guarantees.
Why it matters
This paper introduces a robust semi-supervised inference method that effectively combines multiple AI predictors. It significantly improves upon existing PPI by adapting to predictor performance and offering strong theoretical guarantees. This is crucial for scenarios with limited labeled data, enabling more accurate and reliable predictions.
Original Abstract
The rapidly expanding artificial intelligence (AI) industry has produced diverse yet powerful prediction tools, each with its own network architecture, training strategy, data-processing pipeline, and domain-specific strengths. These tools create new opportunities for semi-supervised inference, in which labeled data are limited and expensive to obtain, whereas unlabeled data are abundant and widely available. Given a collection of predictors, we treat them as a mixture of experts (MOE) and introduce an MOE-powered semi-supervised inference framework built upon prediction-powered inference (PPI). Motivated by the variance reduction principle underlying PPI, the proposed framework seeks the mixture of experts that achieves the smallest possible variance. Compared with standard PPI, the MOE-powered inference framework adapts to the unknown performance of individual predictors, benefits from their collective predictive power, and enjoys a best-expert guarantee. The framework is flexible and applies to mean estimation, linear regression, quantile estimation, and general M-estimation. We develop non-asymptotic theory for the MOE-powered inference framework and establish upper bounds on the coverage error of the resulting confidence intervals. Numerical experiments demonstrate the practical effectiveness of MOE-powered inference and corroborate our theoretical findings.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.