Skip to search formSkip to main contentSkip to account menu

Multi-armed bandit

Known as: N-armed bandit, Two-armed bandit, K-armed bandit 
In probability theory, the multi-armed bandit problem (sometimes called the K- or N-armed bandit problem) is a problem in which a gambler at a row of… 
Wikipedia (opens in a new tab)

Papers overview

Semantic Scholar uses AI to extract papers important to this topic.
Highly Cited
2014
Highly Cited
2014
The stochastic multi-armed bandit model is a simple abstraction that has proven useful in many different contexts in statistics… 
Highly Cited
2010
Highly Cited
2010
We consider the problem of finding the best arm in a stochastic multi-armed bandit game. The regret of a forecaster is here… 
Highly Cited
2010
Highly Cited
2010
A multi-armed bandit is an experiment with the goal of accumulating rewards from a payoff distribution with unknown parameters… 
Highly Cited
2009
Highly Cited
2009
We formulate and study a decentralized multi-armed bandit (MAB) problem. There are M distributed players competing for N… 
Highly Cited
2008
Highly Cited
2008
Algorithms for learning to rank Web documents usually assume a document's relevance is independent of other documents. This leads… 
Highly Cited
2006
Highly Cited
2006
We incorporate statistical confidence intervals in both the multi-armed bandit and the reinforcement learning problems. In the… 
Highly Cited
2005
Highly Cited
2005
The multi-armed bandit problem for a gambler is to decide which arm of a K-slot machine to pull to maximize his total reward in a… 
Highly Cited
2002
Highly Cited
2002
Reinforcement learning policies face the exploration versus exploitation dilemma, i.e. the search for a balance between exploring… 
Highly Cited
1998
Highly Cited
1998
We consider the situation in which individuals in a finite population must repeatedly choose an action yielding an uncertain… 
Highly Cited
1995
Highly Cited
1995
In the multi-armed bandit problem, a gambler must decide which arm of K non-identical slot machines to play in a sequence of…