On the Learning Curves of Revenue Maximization
Steve Hanneke, Alkis Kalavasis, Shay Moran, Grigoris Velegkas
TLDR
This paper characterizes learning curves for revenue maximization in auctions, revealing varied convergence rates based on valuation distributions.
Key contributions
- Introduces learning curves for revenue maximization beyond PAC worst-case bounds.
- Proves existence of Bayes-consistent algorithms with universal but possibly slow convergence.
- Shows optimal decay rate ~1/√n when optimal revenue is from a finite price.
- Demonstrates near-exponential learning curve decay for discrete valuation distributions.
Why it matters
Understanding learning curves in revenue maximization reveals how quickly algorithms improve with data, guiding better auction design and pricing strategies. This work bridges theory and practice by detailing convergence rates beyond worst-case scenarios.
Original Abstract
Learning curves are a fundamental primitive in supervised learning, describing how an algorithm's performance improves with more data and providing a quantitative measure of its generalization ability. Formally, a learning curve plots the decay of an algorithm's error for a fixed underlying distribution as a function of the number of training samples. Prior work on revenue-maximizing learning algorithms, starting with the seminal work of Cole and Roughgarden [STOC, 2014], adopts a distribution-free perspective, which parallels the PAC learning framework in learning theory. This approach evaluates performance against the hardest possible sequence of valuation distributions, one for each sample size, effectively defining the upper envelope of learning curves over all possible distributions, thus leading to error bounds that do not capture the shape of the learning curves. In this work we initiate the study of learning curves for revenue maximization and provide a near-complete characterization of their rate of decay in the basic setting of a single item and a single buyer. In the absence of any restriction on the valuation distribution, we show that there exists a Bayes-consistent algorithm, meaning that its learning curve converges to zero for any arbitrary valuation distribution as the number of samples $n \to \infty$. However, this convergence must be arbitrarily slow, even if the optimal revenue is finite. In contrast, if the optimal revenue is achieved by a finite price, then the optimal rate of decay is roughly $1/\sqrt{n}$. Finally, for distributions supported on discrete sets of values, we show that learning curves decay almost exponentially fast, a rate unattainable under the PAC framework.
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.