Skip to search formSkip to main contentSkip to account menu

Multi-armed bandit

Known as: N-armed bandit, Two-armed bandit, K-armed bandit 
In probability theory, the multi-armed bandit problem (sometimes called the K- or N-armed bandit problem) is a problem in which a gambler at a row of… 
Wikipedia (opens in a new tab)

Papers overview

Semantic Scholar uses AI to extract papers important to this topic.
2013
2013
In the classic Multi-Armed Bandit (MAB) problem, there is a given set of arms with unknown reward distributions. At each time, a… 
2012
2012
We study recommendations for persistent groups that repeatedly engage in a joint activity. We approach this as a multi-arm bandit… 
2012
2012
To investigate the influence of information about fellow group members in a constrained decision-making context, we develop four… 
2010
2010
We propose a model - the "tug-of-war (TOW) model" - to conduct unique parallel searches using many nonlocally correlated search… 
2010
2010
We consider decentralized multi-armed bandit problems with multiple distributed players. At each time, each player chooses one of… 
2006
2006
How should a decision-maker perform repeated choices so as to optimize the average cost or benefit of those choices in the long… 
2005
2005
Purpose – This paper seeks to apply results from the study of bandit processes to cases of information technology (IT) project… 
1998
1998
Functional effects of human a5 nicotinic ACh receptor (AChR) subunits coassembled with a3 and b2 or with a3 and b4 subunits, were… 
Highly Cited
1992
Highly Cited
1992
We have studied the cytoskeletal association of intercellular adhesion molecule-1 (ICAM-1, CD54), an integral membrane protein… 
1983
1983
There are N independent machines. Machine i is described by a sequence {Xi(s), Fi(s)} where xi(s) is the immediate reward and Fi…