Pure Exploration in Multi-armed Bandits Problems

@inproceedings{Bubeck2009PureEI,
  title={Pure Exploration in Multi-armed Bandits Problems},
  author={S{\'e}bastien Bubeck and R{\'e}mi Munos and Gilles Stoltz},
  booktitle={ALT},
  year={2009}
}
We consider the framework of stochastic multi-armed bandit problems and study the possibilities and limitations of strategies that explore sequentially the arms. The strategies are assessed in terms of their simple regrets, a regret notion that captures the fact that exploration is only constrained by the number of available rounds (not necessarily known in advance), in contrast to the case when the cumulative regret is considered and when exploitation needs to be performed at the same time. We… CONTINUE READING
Highly Influential
This paper has highly influenced 25 other papers. REVIEW HIGHLY INFLUENTIAL CITATIONS
Highly Cited
This paper has 218 citations. REVIEW CITATIONS

Citations

Publications citing this paper.
Showing 1-10 of 156 extracted citations

Principles of Metalevel Control

View 8 Excerpts
Highly Influenced

Strategic Choices in Optimization

J. Inf. Sci. Eng. • 2014
View 4 Excerpts
Highly Influenced

219 Citations

02040'11'13'15'17'19
Citations per Year
Semantic Scholar estimates that this publication has 219 citations based on the available data.

See our FAQ for additional information.

Similar Papers

Loading similar papers…