When Hardness of Approximation Meets Hardness of Learning

Eran Malach, Shai Shalev-Shwartz

Research output: Contribution to journalArticlepeer-review

Abstract

A supervised learning algorithm has access to a distribution of labeled examples, and needs to return a function (hypothesis) that correctly labels the examples. The hypothesis of the learner is taken from some fixed class of functions (e.g., linear classifiers, neural networks etc.). A failure of the learning algorithm can occur due to two possible reasons: wrong choice of hypothesis class (hardness of approximation), or failure to find the best function within the hypothesis class (hardness of learning). Although both approximation and learnability are important for the success of the algorithm, they are typically studied separately. In this work, we show a single hardness property that implies both hardness of approximation using linear classes and shallow networks, and hardness of learning using correlation queries and gradient-descent. This allows us to obtain new results on hardness of approximation and learnability of parity functions, DNF formulas and AC0 circuits.

Original languageAmerican English
Article number91
Number of pages24
JournalJournal of Machine Learning Research
Volume23
StatePublished - 2022

Keywords

  • Hardness of learning
  • approximation
  • gradient-descent
  • neural networks
  • statistical-queries

All Science Journal Classification (ASJC) codes

  • Software
  • Artificial Intelligence
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'When Hardness of Approximation Meets Hardness of Learning'. Together they form a unique fingerprint.

Cite this