Super-Teaching in Machine Learning

Dina Barak-Pelleg, Daniel Berend, Aryeh Kontorovich

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In machine learning, efficiently leveraging large datasets is essential, particularly when computational resources are limited. Sample compression techniques, which involve using carefully chosen subsets of data to train models, provide significant reductions in runtime and memory requirements while preserving generalization performance. This paper investigates a concept known as “super-teaching,” where a knowledgeable teacher selectively provides an optimal subset of training data to a learner, thereby enhancing learning outcomes. Building on the work of Ma et al. (2018), who demonstrated that a teacher with complete knowledge of the data distribution could significantly improve a learner’s performance, our study extends this idea beyond the scenario in their paper. We present a more robust and generalized approach, offering insights into how super-teaching can be effectively applied to a broader range of machine learning problems, potentially leading to better training efficiency and improved generalization.

Original languageAmerican English
Title of host publicationCyber Security, Cryptology, and Machine Learning - 8th International Symposium, CSCML 2024, Proceedings
EditorsShlomi Dolev, Michael Elhadad, Mirosław Kutyłowski, Giuseppe Persiano
PublisherSpringer Science and Business Media Deutschland GmbH
Pages335-342
Number of pages8
ISBN (Print)9783031769337
DOIs
StatePublished - 1 Jan 2025
Event8th International Symposium on Cyber Security, Cryptology, and Machine Learning, CSCML 2024 - Be'er Sheva, Israel
Duration: 19 Dec 202420 Dec 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume15349 LNCS

Conference

Conference8th International Symposium on Cyber Security, Cryptology, and Machine Learning, CSCML 2024
Country/TerritoryIsrael
CityBe'er Sheva
Period19/12/2420/12/24

Keywords

  • concentration
  • Machine learning
  • super-teacher
  • teacher
  • unimodal distribution

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • General Computer Science

Cite this