#### Filter Results:

#### Publication Year

2003

2017

#### Publication Type

#### Co-author

#### Publication Venue

#### Data Set Used

#### Key Phrases

Learn More

- Timothy P. Lillicrap, Jonathan J. Hunt, +5 authors Daan Wierstra
- ArXiv
- 2015

We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the de-terministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20… (More)

- Volodymyr Mnih, Nicolas Heess, Alex Graves, Koray Kavukcuoglu
- NIPS
- 2014

Applying convolutional neural networks to large images is computationally expensive because the amount of computation scales linearly with the number of image pixels. We present a novel recurrent neural network model that is capable of extracting information from an image or video by adaptively selecting a sequence of regions or locations and only… (More)

- S. M. Ali Eslami, Nicolas Heess, Christopher K. I. Williams, John M. Winn
- International Journal of Computer Vision
- 2012

A good model of object shape is essential in applications such as segmentation, detection, inpainting and graphics. For example, when performing segmentation, local constraints on the shapes can help where object boundaries are noisy or unclear, and global constraints can resolve ambiguities where background clutter looks similar to parts of the objects. In… (More)

In this paper we consider deterministic policy gradient algorithms for reinforcement learning with continuous actions. The deterministic policy gradient has a particularly appealing form: it is the expected gradient of the action-value function. This simple form means that the deter-ministic policy gradient can be estimated much more efficiently than the… (More)

- Nicolas Le Roux, Nicolas Heess, Jamie Shotton, John M. Winn
- Neural Computation
- 2011

Computer vision has grown tremendously in the past two decades. Despite all efforts, existing attempts at matching parts of the human visual system's extraordinary ability to understand visual scenes lack either scope or power. By combining the advantages of general low-level generative models and powerful layer-based and hierarchical models, this work aims… (More)

We present a unified framework for learning continuous control policies using backpropagation. It supports stochastic control by treating stochasticity in the Bellman equation as a deterministic function of exogenous noise. The product is a spectrum of general policy gradient algorithms that range from model-free methods with value functions to model-based… (More)

- Jyri J. Kivinen, Christopher K. I. Williams, Nicolas Heess
- AISTATS
- 2014

This paper investigates visual boundary detection , i.e. prediction of the presence of a boundary at a given image location. We develop a novel neurally-inspired deep architecture for the task. Notable aspects of our work are (i) the use of " covariance features " [Ranzato and Hinton, 2010] which depend on the squared response of a filter to the input… (More)

Partially observed control problems are a challenging aspect of reinforcement learning. We extend two related, model-free algorithms for continuous control – deterministic policy gradient and stochastic value gradient – to solve partially observed domains using recurrent neural networks trained with backpropagation through time. We demonstrate that this… (More)

- Bogdan Alexe, Nicolas Heess, Yee Whye Teh, Vittorio Ferrari
- NIPS
- 2012

The dominant visual search paradigm for object class detection is sliding windows. Although simple and effective, it is also wasteful, unnatural and rigidly hardwired. We propose strategies to search for objects which intelligently explore the space of windows by making sequential observations at locations decided based on previous observations. Our… (More)

- John Schulman, Nicolas Heess, Theophane Weber, Pieter Abbeel
- NIPS
- 2015

In a variety of problems originating in supervised, unsupervised, and reinforcement learning, the loss function is defined by an expectation over a collection of random variables, which might be part of a probabilistic model or the external world. Estimating the gradient of this loss function, using samples, lies at the core of gradient-based learning… (More)