Multi-armed bandit

Known as: N-armed bandit, Two-armed bandit, K-armed bandit 
In probability theory, the multi-armed bandit problem (sometimes called the K- or N-armed bandit problem) is a problem in which a gambler at a row of… (More)
Wikipedia

Topic mentions per year

Topic mentions per year

1968-2017
010020030019682017

Papers overview

Semantic Scholar uses AI to extract papers important to this topic.
2015
2015
This paper is about the study of Multi–Armed Bandit (MAB) approaches for pricing applications, where a seller needs to identify… (More)
  • figure 1
  • figure 2
  • figure 3
  • figure 5
Is this relevant?
Highly Cited
2012
Highly Cited
2012
Multi-armed bandit problems are the most basic examples of sequential decision problems with an exploration–exploitation trade… (More)
  • figure 3.1
  • figure 4.1
Is this relevant?
Highly Cited
2012
Highly Cited
2012
The multi-armed bandit problem is a popular model for studying exploration/exploitation trade-off in sequential decision problems… (More)
  • figure 1
Is this relevant?
Highly Cited
2010
Highly Cited
2010
We formulate and study a decentralized multi-armed bandit (MAB) problem. There are M distributed players competing for N… (More)
  • figure 1
  • figure 2
  • figure 3
  • figure 4
  • figure 6
Is this relevant?
2008
2008
This paper considers the multi-armed bandit problem with multiple simultaneous arm pulls. We develop a new ‘irrevocable… (More)
  • table 1
Is this relevant?
Highly Cited
2007
Highly Cited
2007
Multi-armed bandit (MAB) problems are a class of sequential resource allocation problems concerned with allocating one or more… (More)
Is this relevant?
Highly Cited
2007
Highly Cited
2007
We provide a framework to exploit dependencies among arms in multi-armed bandit problems, when the dependencies are in the form… (More)
  • figure 1
  • figure 3
  • figure 4
  • figure 5
Is this relevant?
Highly Cited
2005
Highly Cited
2005
The multi-armed bandit problem for a gambler is to decide which arm of a K-slot machine to pull to maximize his total reward in a… (More)
  • table 1
Is this relevant?
Highly Cited
2002
Highly Cited
2002
The bandit problem is revisited and considered under the PAC model. Our main contribution in this part is to show that given n… (More)
Is this relevant?
Highly Cited
1987
Highly Cited
1987
The multi-armed bandit problem arises in sequentially allocating effcot to one of N prefects and sequentially asngning patients… (More)
Is this relevant?