On Explore-Then-Commit Strategies

Abstract : We study the problem of minimising regret in two-armed bandit problems with Gaussian rewards. Our objective is to use this simple setting to illustrate that strategies based on an exploration phase (up to a stopping time) followed by exploitation are necessarily suboptimal. The results hold regardless of whether or not the difference in means between the two arms is known. Besides the main message, we also refine existing deviation inequalities, which allow us to design fully sequential strategies with finite-time regret guarantees that are (a) asymptotically optimal as the horizon grows and (b) order-optimal in the minimax sense. Furthermore we provide empirical evidence that the theory also holds in practice and discuss extensions to non-gaussian and multiple-armed case.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-01322906
Contributor : Emilie Kaufmann <>
Submitted on : Monday, November 14, 2016 - 1:19:04 PM
Last modification on : Monday, April 29, 2019 - 4:47:06 PM
Long-term archiving on : Tuesday, March 21, 2017 - 12:39:18 AM

Files

nips_final.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-01322906, version 2
  • ARXIV : 1605.08988

Citation

Aurélien Garivier, Emilie Kaufmann, Tor Lattimore. On Explore-Then-Commit Strategies. NIPS, Dec 2016, Barcelona, Spain. ⟨hal-01322906v2⟩

Share

Metrics

Record views

381

Files downloads

175