Boosting for Straggling and Flipping Classifiers

Yuval Cassuto, Yogjune Kim

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Boosting is a well-known method in machine learning for combining multiple weak classifiers into one strong classifier. When used in distributed setting, accuracy is hurt by classifiers that flip or straggle due to communication and/or computation unreliability. While unreliability in the form of noisy data is well-treated by the boosting literature, the unreliability of the classifier outputs has not been explicitly addressed. Protecting the classifier outputs with an error/erasure-correcting code requires reliable encoding of multiple classifier outputs, which is not feasible in common distributed settings. In this paper we address the problem of training boosted classifiers subject to straggling or flips at classification time. We propose two approaches: one based on minimizing the usual exponential loss but in expectation over the classifier errors, and one by defining and minimizing a new worst-case loss for a specified bound on the number of unreliable classifiers.

Original languageEnglish
Title of host publication2021 IEEE International Symposium on Information Theory, ISIT 2021 - Proceedings
Pages2441-2446
Number of pages6
ISBN (Electronic)9781538682098
DOIs
StatePublished - 12 Jul 2021
Event2021 IEEE International Symposium on Information Theory, ISIT 2021 - Virtual, Melbourne, Australia
Duration: 12 Jul 202120 Jul 2021

Publication series

NameIEEE International Symposium on Information Theory - Proceedings
Volume2021-July

Conference

Conference2021 IEEE International Symposium on Information Theory, ISIT 2021
Country/TerritoryAustralia
CityVirtual, Melbourne
Period12/07/2120/07/21

All Science Journal Classification (ASJC) codes

  • Theoretical Computer Science
  • Information Systems
  • Modelling and Simulation
  • Applied Mathematics

Fingerprint

Dive into the research topics of 'Boosting for Straggling and Flipping Classifiers'. Together they form a unique fingerprint.

Cite this