ArXiv TLDR

What is Learnable in Valiant's Theory of the Learnable?

🐦 Tweet
2605.13840

Steve Hanneke, Anay Mehrotra, Grigoris Velegkas, Manolis Zampetakis

stat.MLcs.DScs.LGmath.STstat.CO

TLDR

This paper characterizes learnability in Valiant's original model, showing membership queries expand learnable classes and providing a new algorithm for halfspaces.

Key contributions

  • Introduces a new "adaptive query-compression scheme" to characterize learnability in Valiant's model for finite domains.
  • Shows learnability in Valiant's model is strictly between PAC and Valiant's model without membership queries.
  • Demonstrates membership queries fundamentally change the set of learnable classes, a rare occurrence in learning theory.
  • Presents the first algorithm for learning d-dimensional halfspaces in Valiant's model, which are not learnable without queries.

Why it matters

This paper uncovers a surprisingly rich theory behind Valiant's original learnability model. It shows that membership queries can fundamentally expand the set of learnable classes, a rare and significant finding. The new characterization and algorithm for halfspaces advance our understanding of learning theory.

Original Abstract

Valiant's 1984 paper is widely credited with introducing the PAC learning model, but it, in fact, introduced a different model: unlike PAC learning, the learner receives only positives, may issue membership queries, and must output a hypothesis with no false positives. Prior work characterized variants, including the case without queries. We revisit Valiant's original model and ask: *Which classes are learnable in it?* For every finite domain, including Valiant's Boolean-hypercube setting, we show that a class is learnable if and only if every realizable positive sample can be certified by a poly-size adaptive query-compression scheme. This is a new variant of sample compression where the learner certifies samples via a short interaction with the membership oracle. Our characterization shows that learnability in Valiant's model is strictly sandwiched between learnability in the PAC model and the variant of Valiant's model without membership queries. This is one of the rare cases where introducing membership queries changes the set of learnable classes, and not just the sample or computational complexity. Next, we study the natural extension of the model to arbitrary domains. While we do not obtain an exact characterization, our techniques readily generalize and show that the same strict sandwiching persists. Finally, we show that $d$-dimensional halfspaces, which are not learnable without queries, are learnable with queries: we give a $\mathrm{poly}(d) \tilde{O}(1/ε)$ sample and $\mathrm{poly}(d) \mathrm{polylog}(1/ε)$ query algorithm, and prove that at least $Ω(d)$ samples or queries are necessary. To our knowledge, this is the first algorithm for halfspaces in Valiant's model. Together, these results uncover a surprisingly rich theory behind Valiant's original notion of learnability and introduce ideas that may be of independent interest in learning theory.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.