#### Filter Results:

- Full text PDF available (42)

#### Publication Year

2003

2017

- This year (13)
- Last 5 years (35)
- Last 10 years (42)

#### Publication Type

#### Co-author

#### Journals and Conferences

#### Data Set Used

#### Key Phrases

#### Method

#### Organism

Learn More

- Timothy P. Lillicrap, Jonathan J. Hunt, +5 authors Daan Wierstra
- ArXiv
- 2015

We adapt the ideas underlying the success of Deep Q-Learning to the continuous action domain. We present an actor-critic, model-free algorithm based on the deterministic policy gradient that can operate over continuous action spaces. Using the same learning algorithm, network architecture and hyper-parameters, our algorithm robustly solves more than 20… (More)

- Volodymyr Mnih, Nicolas Heess, Alex Graves, Koray Kavukcuoglu
- NIPS
- 2014

Applying convolutional neural networks to large images is computationally expensive because the amount of computation scales linearly with the number of image pixels. We present a novel recurrent neural network model that is capable of extracting information from an image or video by adaptively selecting a sequence of regions or locations and only… (More)

In this paper we consider deterministic policy gradient algorithms for reinforcement learning with continuous actions. The deterministic policy gradient has a particularly appealing form: it is the expected gradient of the action-value function. This simple form means that the deterministic policy gradient can be estimated much more efficiently than the… (More)

- S. M. Ali Eslami, Nicolas Heess, Christopher K. I. Williams, John M. Winn
- International Journal of Computer Vision
- 2012

A good model of object shape is essential in applications such as segmentation, detection, inpainting and graphics. For example, when performing segmentation, local constraints on the shapes can help where object boundaries are noisy or unclear, and global constraints can resolve ambiguities where background clutter looks similar to parts of the objects. In… (More)

- S. M. Ali Eslami, Nicolas Heess, +4 authors Geoffrey E. Hinton
- NIPS
- 2016

We present a framework for efficient inference in structured image models that explicitly reason about objects. We achieve this by performing probabilistic inference using a recurrent neural network that attends to scene elements and processes them one at a time. Crucially, the model itself learns to choose the appropriate number of inference steps. We use… (More)

- Jyri J. Kivinen, Christopher K. I. Williams, Nicolas Heess
- AISTATS
- 2014

This paper investigates visual boundary detection, i.e. prediction of the presence of a boundary at a given image location. We develop a novel neurally-inspired deep architecture for the task. Notable aspects of our work are (i) the use of “covariance features” [Ranzato and Hinton, 2010] which depend on the squared response of a filter to the input image,… (More)

We present a unified framework for learning continuous control policies using backpropagation. It supports stochastic control by treating stochasticity in the Bellman equation as a deterministic function of exogenous noise. The product is a spectrum of general policy gradient algorithms that range from model-free methods with value functions to model-based… (More)

A key goal of computer vision is to recover the underlying 3D structure from 2D observations of the world. In this paper we learn strong deep generative models of 3D structures, and recover these structures from 3D and 2D images via probabilistic inference. We demonstrate high-quality samples and report log-likelihoods on several datasets, including… (More)

- Nicolas Le Roux, Nicolas Heess, Jamie Shotton, John M. Winn
- Neural Computation
- 2011

Computer vision has grown tremendously in the past two decades. Despite all efforts, existing attempts at matching parts of the human visual system's extraordinary ability to understand visual scenes lack either scope or power. By combining the advantages of general low-level generative models and powerful layer-based and hierarchical models, this work aims… (More)

- John Schulman, Nicolas Heess, Theophane Weber, Pieter Abbeel
- NIPS
- 2015

In a variety of problems originating in supervised, unsupervised, and reinforcement learning, the loss function is defined by an expectation over a collection of random variables, which might be part of a probabilistic model or the external world. Estimating the gradient of this loss function, using samples, lies at the core of gradient-based learning… (More)