Table of Contents
Multi-Armed Bandits
Surveys
Theory
Related Pages
Multi-Armed Bandits
See
Wikipedia - Multi-armed Bandit
.
Surveys
Bubeck & Cesa-Bianchi 2012 - Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems
Very good survey
Theory
Mei et al 2023 - Stochastic Gradient Succeeds for Bandits
Related Pages
Reinforcement Learning
Online Learning
Regret Bounds