Tight Lower Bounds for Combinatorial Multi-Armed Bandits

Nadav Merlis, Shie Mannor

Research output: Contribution to journalConference articlepeer-review

Abstract

The Combinatorial Multi-Armed Bandit problem is a sequential decision-making problem in which an agent selects a set of arms at each round, observes feedback for each of these arms and aims to maximize a known reward function of the arms it chose. While previous work proved regret upper bounds in this setting for general reward functions, only a few works provided matching lower bounds, all for specific reward functions. In this work, we prove regret lower bounds for combinatorial bandits that hold under mild assumptions for all smooth reward functions. We derive both problem-dependent and problem-independent bounds and show that the recently proposed Gini-weighted smoothness parameter (Merlis and Mannor, 2019) also determines the lower bounds for monotone reward functions. Notably, this implies that our lower bounds are tight up to log-factors.

Original languageEnglish
Pages (from-to)2830-2857
Number of pages28
JournalProceedings of Machine Learning Research
Volume125
StatePublished - 2020
Event33rd Conference on Learning Theory, COLT 2020 - Virtual, Online, Austria
Duration: 9 Jul 202012 Jul 2020

Keywords

  • Combinatorial Multi-Armed Bandits
  • Gini-Weighted Smoothness
  • Lower Bounds

All Science Journal Classification (ASJC) codes

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Tight Lower Bounds for Combinatorial Multi-Armed Bandits'. Together they form a unique fingerprint.

Cite this