Abstract
We consider the use of no-regret algorithms to compute equilibria for particular classes of convex-concave games. While standard regret bounds would lead to convergence rates on the order of O(T−1/2), recent work (Rakhlin and Sridharan, 2013b; Syrgkanis et al., 2015) has established O(1/T) rates by taking advantage of a particular class of optimistic prediction algorithms. In this work we go further, showing that for a particular class of games one achieves a O(1/T2) rate, and we show how this applies to the Frank-Wolfe method and recovers a similar bound (Garber and Hazan, 2015). We also show that such no-regret techniques can even achieve a linear rate, O(exp(−T)), for equilibrium computation under additional curvature assumptions.
| Original language | English |
|---|---|
| Pages (from-to) | 1595-1625 |
| Number of pages | 31 |
| Journal | Proceedings of Machine Learning Research |
| Volume | 75 |
| State | Published - 2018 |
| Externally published | Yes |
| Event | 31st Annual Conference on Learning Theory, COLT 2018 - Stockholm, Sweden Duration: 6 Jul 2018 → 9 Jul 2018 |
Keywords
- Frank-Wolfe
- Online learning
- fast rates
- zero-sum games
All Science Journal Classification (ASJC) codes
- Software
- Artificial Intelligence
- Control and Systems Engineering
- Statistics and Probability