Adversarial Bandits: Theory and Algorithms

Published: 01 January 1970
on channel: Simons Institute
3,018
44

Haipeng Luo (USC)
https://simons.berkeley.edu/talks/adv...
Data-Driven Decision Processes Boot Camp

The adversarial (a.k.a. non-stochastic) multi-armed bandit problem is an influential marriage between the online learning literature that concerns sequential decision making without distributional assumptions and the bandit literature that concerns learning with partial information feedback. This tutorial will give an overview of the theory and algorithms on this topic, starting from classical algorithms and their analysis and then moving on to advances in recent years on data-dependent regret guarantees, structural bandits, bandit with switching costs, combining bandit algorithms, and others. Special focus will be given to highlighting the similarities and differences between online learning with full-information feedback and that with bandit feedback.


Watch video Adversarial Bandits: Theory and Algorithms online without registration, duration hours minute second in high quality. This video was added by user Simons Institute 01 January 1970, don't forget to share it with your friends and acquaintances, it has been viewed on our site 3,018 once and liked it 44 people.