An adaptive synchronization approach for weights of deep reinforcement learning
@article{Badran2020AnAS, title={An adaptive synchronization approach for weights of deep reinforcement learning}, author={Saeed Badran and Mansoor Rezghi}, journal={ArXiv}, year={2020}, volume={abs/2008.06973} }
Deep Q-Networks (DQN) is one of the most well-known methods of deep reinforcement learning, which uses deep learning to approximate the action-value function. Solving numerous Deep reinforcement learning challenges such as moving targets problem and the correlation between samples are the main advantages of this model. Although there have been various extensions of DQN in recent years, they all use a similar method to DQN to overcome the problem of moving targets. Despite the advantages…
15 References
Deep Reinforcement Learning with Double Q-Learning
- Computer ScienceAAAI
- 2016
This paper proposes a specific adaptation to the DQN algorithm and shows that the resulting algorithm not only reduces the observed overestimations, as hypothesized, but that this also leads to much better performance on several games.
Dueling Network Architectures for Deep Reinforcement Learning
- Computer ScienceICML
- 2016
This paper presents a new neural network architecture for model-free reinforcement learning that leads to better policy evaluation in the presence of many similar-valued actions and enables the RL agent to outperform the state-of-the-art on the Atari 2600 domain.
Human-level control through deep reinforcement learning
- Computer ScienceNature
- 2015
This work bridges the divide between high-dimensional sensory inputs and actions, resulting in the first artificial agent that is capable of learning to excel at a diverse array of challenging tasks.
Self-improving reactive agents based on reinforcement learning, planning and teaching
- Computer ScienceMachine Learning
- 2004
This paper compares eight reinforcement learning frameworks: Adaptive heuristic critic (AHC) learning due to Sutton, Q-learning due to Watkins, and three extensions to both basic methods for speeding up learning and two extensions are experience replay, learning action models for planning, and teaching.
Reinforcement Learning: An Introduction
- Computer ScienceIEEE Transactions on Neural Networks
- 2005
This book provides a clear and simple account of the key ideas and algorithms of reinforcement learning, which ranges from the history of the field's intellectual foundations to the most recent developments and applications.
Dopamine: A Research Framework for Deep Reinforcement Learning
- Computer ScienceArXiv
- 2018
Dopamine is an open-source, TensorFlow-based, and compact and reliable implementations of some state-of-the-art deep RL agents that complement this offering with a taxonomy of the different research objectives in deep RL research.
Double Q-learning
- Computer ScienceNIPS
- 2010
An alternative way to approximate the maximum expected value for any set of random variables is introduced and the obtained double estimator method is shown to sometimes underestimate rather than overestimate themaximum expected value.
A Distributional Perspective on Reinforcement Learning
- Computer ScienceICML
- 2017
This paper argues for the fundamental importance of the value distribution: the distribution of the random return received by a reinforcement learning agent, and designs a new algorithm which applies Bellman's equation to the learning of approximate value distributions.
Rainbow: Combining Improvements in Deep Reinforcement Learning
- Computer ScienceAAAI
- 2018
This paper examines six extensions to the DQN algorithm and empirically studies their combination, showing that the combination provides state-of-the-art performance on the Atari 2600 benchmark, both in terms of data efficiency and final performance.
Prioritized Experience Replay
- Computer ScienceICLR
- 2016
A framework for prioritizing experience, so as to replay important transitions more frequently, and therefore learn more efficiently, in Deep Q-Networks, a reinforcement learning algorithm that achieved human-level performance across many Atari games.