Fiduciary bandits

Gal Bahar, Omer Ben Porat, Kevin Leyton Brown, Moshe Tennenholtz

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Recommendation systems often face explorationexploitation tradeoffs: The system can only learn about the desirability of new options by recommending them to some user. Such systems can thus be modeled as multi-Armed bandit settings; however, users are self-interested and cannot be made to follow recommendations. We ask whether exploration can nevertheless be performed in a way that scrupulously respects agents interests-i.e., by a system that acts as a fiduciary. More formally, we introduce a model in which a recommendation system faces an explorationexploitation tradeoff under the constraint that it can never recommend any action that it knows yields lower reward in expectation than an agent would achieve if it acted alone. Our main contribution is a positive result: An asymptotically optimal, incentive compatible, and ex-Ante individually rational recommendation algorithm.

Original languageEnglish
Title of host publication37th International Conference on Machine Learning, ICML 2020
EditorsHal Daume, Aarti Singh
Pages495-504
Number of pages10
ISBN (Electronic)9781713821120
StatePublished - 2020
Event37th International Conference on Machine Learning, ICML 2020 - Virtual, Online
Duration: 13 Jul 202018 Jul 2020

Publication series

Name37th International Conference on Machine Learning, ICML 2020
VolumePartF168147-1

Conference

Conference37th International Conference on Machine Learning, ICML 2020
CityVirtual, Online
Period13/07/2018/07/20

All Science Journal Classification (ASJC) codes

  • Computational Theory and Mathematics
  • Human-Computer Interaction
  • Software

Cite this