Skip to search form
Skip to main content
Skip to account menu
Semantic Scholar
Semantic Scholar's Logo
Search 218,179,913 papers from all fields of science
Search
Sign In
Create Free Account
Andrew Barto
Known as:
Andrew G. Barto
, Barto
, Barto, Andrew G.
Andrew Barto is a professor of computer science at University of Massachusetts Amherst, and chair of the department since January 2007. His main…
Expand
Wikipedia
(opens in a new tab)
Create Alert
Alert
Related topics
Related topics
4 relations
Computer science
Constructing skill trees
Reinforcement learning
Skill chaining
Papers overview
Semantic Scholar uses AI to extract papers important to this topic.
2015
2015
Portable Option Discovery for Automated Learning Transfer in Object-Oriented Markov Decision Processes
Nicholay Topin
,
Nicholas Haltmeyer
,
S. Squire
,
J. Winder
,
Marie desJardins
,
J. MacGlashan
International Joint Conference on Artificial…
2015
Corpus ID: 17921651
We introduce a novel framework for option discovery and learning transfer in complex domains that are represented as object…
Expand
Highly Cited
2012
Highly Cited
2012
Linear Off-Policy Actor-Critic
T. Degris
,
Martha White
,
R. Sutton
International Conference on Machine Learning
2012
Corpus ID: 10513082
This paper presents the first actor-critic algorithm for off-policy reinforcement learning. Our algorithm is online and…
Expand
Highly Cited
2005
Highly Cited
2005
Lazy Approximation for Solving Continuous Finite-Horizon MDPs
Lihong Li
,
M. Littman
AAAI Conference on Artificial Intelligence
2005
Corpus ID: 14092393
Solving Markov decision processes (MDPs) with continuous state spaces is a challenge due to, among other problems. the well-known…
Expand
Highly Cited
2000
Highly Cited
2000
Acquisition of stand-up behavior by a real robot using hierarchical reinforcement learning
J. Morimoto
,
K. Doya
Robotics Auton. Syst.
2000
Corpus ID: 5937494
Highly Cited
1998
Highly Cited
1998
Solving Large POMDPs using Real Time Dynamic Programming
Hector Geffner
,
Blai Bonet
1998
Corpus ID: 14063723
Partially Observable Markov Decision Processes (pomdps) are general models of sequential decision problems in which both actions…
Expand
Highly Cited
1998
Highly Cited
1998
An Analysis of Actor/Critic Algorithms Using Eligibility Traces: Reinforcement Learning with Imperfect Value Function
H. Kimura
,
S. Kobayashi
International Conference on Machine Learning
1998
Corpus ID: 29097000
We present an analysis of actor/critic algorithms, in which the actor updates its policy using eligibility traces of the policy…
Expand
Highly Cited
1995
Highly Cited
1995
Adaptive critic designs: A case study for neurocontrol
D. Prokhorov
,
R. Santiago
,
D. Wunsch
Neural Networks
1995
Corpus ID: 11934898
Highly Cited
1993
Highly Cited
1993
Learning to Solve Markovian Decision Processes
Satinder Singh
1993
Corpus ID: 59715909
This dissertation is about building learning control architectures for agents embedded in finite, stationary, and Markovian…
Expand
Highly Cited
1989
Highly Cited
1989
Generalization and Scaling in Reinforcement Learning
D. Ackley
,
M. Littman
Neural Information Processing Systems
1989
Corpus ID: 16277413
In associative reinforcement learning, an environment generates input vectors, a learning system generates possible output…
Expand
Highly Cited
1971
Highly Cited
1971
Aadaptive lactate dehydrogenase variation in the crested blenny, Anoplarchus1
M. Johnson
Heredity
1971
Corpus ID: 33581229
GEL electrophoresis of proteins has demonstrated much heterozygosity in populations (e.g. Harris, 1966; Lewontin and Hubby, 1966…
Expand
By clicking accept or continuing to use the site, you agree to the terms outlined in our
Privacy Policy
(opens in a new tab)
,
Terms of Service
(opens in a new tab)
, and
Dataset License
(opens in a new tab)
ACCEPT & CONTINUE