Generative Augmented Inference
Cheng Lu, Mengxin Wang, Dennis J. Zhang, Heng Zhang
TLDR
GAI is a new framework that efficiently integrates AI-generated data as features to improve estimation and reduce human labeling costs in data-driven operations.
Key contributions
- Introduces Generative Augmented Inference (GAI) to integrate AI outputs as features for human-labeled outcomes.
- Uses orthogonal moment construction for consistent estimation and valid inference with flexible AI-human relationships.
- Achieves "safe default" property, improving efficiency over human-only data when AI signals are predictive.
- Empirically reduces estimation error by 50% and human labeling by over 75% across diverse applications.
Why it matters
This paper offers a principled solution to leverage inexpensive AI-generated data effectively, addressing the challenge of its complex relationship with human labels. GAI significantly reduces the need for costly human annotations while improving model accuracy and inference validity. This makes data-driven operations more scalable and accessible.
Original Abstract
Data-driven operations management often relies on parameters estimated from costly human-generated labels. Recent advances in large language models (LLMs) and other AI systems offer inexpensive auxiliary data, but introduce a new challenge: AI outputs are not direct observations of the target outcomes, but could involve high-dimensional representations with complex and unknown relationships to human labels. Conventional methods leverage AI predictions as direct proxies for true labels, which can be inefficient or unreliable when this relationship is weak or misspecified. We propose Generative Augmented Inference (GAI), a general framework that incorporates AI-generated outputs as informative features for estimating models of human-labeled outcomes. GAI uses an orthogonal moment construction that enables consistent estimation and valid inference with flexible, nonparametric relationship between LLM-generated outputs and human labels. We establish asymptotic normality and show a "safe default" property: relative to human-data-only estimators, GAI weakly improves estimation efficiency under arbitrary auxiliary signals and yields strict gains whenever the auxiliary information is predictive. Empirically, GAI outperforms benchmarks across diverse settings. In conjoint analysis with weak auxiliary signals, GAI reduces estimation error by about 50% and lowers human labeling requirements by over 75%. In retail pricing, where all methods access the same auxiliary inputs, GAI consistently outperforms alternative estimators, highlighting the value of its construction rather than differences in information. In health insurance choice, it cuts labeling requirements by over 90% while maintaining decision accuracy. Across applications, GAI improves confidence interval coverage without inflating width. Overall, GAI provides a principled and scalable approach to integrating AI-generated information.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.