Stochastic Backpropagation and Approximate Inference in Deep Generative Models
@inproceedings{JimenezRezende2014StochasticBA, title={Stochastic Backpropagation and Approximate Inference in Deep Generative Models}, author={Danilo Jimenez Rezende and Shakir Mohamed and Daan Wierstra}, booktitle={ICML}, year={2014} }
We marry ideas from deep neural networks and approximate Bayesian inference to derive a generalised class of deep, directed generative models, endowed with a new algorithm for scalable inference and learning. [] Key Method We develop stochastic back-propagation -- rules for back-propagation through stochastic variables -- and use this to develop an algorithm that allows for joint optimisation of the parameters of both the generative and recognition model. We demonstrate on several real-world data sets that…
3,814 Citations
Towards Local Learning and MCMC Inference in Biologically Plausible Deep Generative Networks
- Computer Science
- 2018
An ansatz for approximate Markov Chain Monte Carlo inference and learning in multi-layered (deep) generative models that exhibit biologically plausible dynamics is proposed and provides new perspectives on how biologically plausible learning algorithms could be reconciled with statistical learning theory.
Learning Deep Generative Models with Doubly Stochastic MCMC
- Computer ScienceArXiv
- 2015
This work presents doubly stochastic gradient MCMC, a simple and generic method for (approximate) Bayesian inference of deep generative models (DGMs) in a collapsed continuous parameter space that outperforms many state-of-the-art competitors.
Structured Inference Networks for Nonlinear State Space Models
- Computer ScienceAAAI
- 2017
A unified algorithm is introduced to efficiently learn a broad class of linear and non-linear state space models, including variants where the emission and transition distributions are modeled by deep neural networks.
Neural Variational Inference and Learning in Undirected Graphical Models
- Computer ScienceNIPS
- 2017
This work proposes black-box learning and inference algorithms for undirected models that optimize a variational approximation to the log-likelihood of the model via a unified variational inference framework and empirically demonstrates the effectiveness of the method on several popular generative modeling datasets.
Asymptotically exact inference in differentiable generative models
- Computer ScienceAISTATS
- 2017
A method for performing efficient MCMC inference in generative models when conditioning on observations of the model output is presented, using a constrained variant of Hamiltonian Monte Carlo which leverages the smooth geometry of the manifold to coherently move between inputs exactly consistent with observations.
Inference in Deep Networks in High Dimensions
- Computer Science2018 IEEE International Symposium on Information Theory (ISIT)
- 2018
The main contribution shows that the mean-squared error (MSE) of ML-VAMP can be exactly predicted in a certain large system limit and matches the Bayes optimal value recently postulated by Reeves when certain fixed point equations have unique solutions.
Structured VAEs: Composing Probabilistic Graphical Models and Variational Autoencoders
- Computer Science
- 2016
A new framework for unsupervised learning is developed that composes probabilistic graphical models with deep learning methods and combines their respective strengths to learn flexible feature models and bottom-up recognition networks.
Variational Gaussian Process
- Computer ScienceICLR
- 2016
The variational Gaussian process is constructed, a Bayesian nonparametric model which adapts its shape to match complex posterior distributions, and is proved a universal approximation theorem for the VGP, demonstrating its representative power for learning any model.
Composing graphical models with neural networks for structured representations and fast inference
- Computer ScienceNIPS
- 2016
A general modeling and inference framework that composes probabilistic graphical models with deep learning methods and combines their respective strengths is proposed, giving a scalable algorithm that leverages stochastic variational inference, natural gradients, graphical model message passing, and the reparameterization trick.
Probabilistic Models with Deep Neural Networks
- Computer ScienceEntropy
- 2021
An overview of the main concepts, methods, and tools needed to use deep neural networks within a probabilistic modeling framework is provided.
References
SHOWING 1-10 OF 44 REFERENCES
Deep AutoRegressive Networks
- Computer ScienceICML
- 2014
An efficient approximate parameter estimation method based on the minimum description length (MDL) principle is derived, which can be seen as maximising a variational lower bound on the log-likelihood, with a feedforward neural network implementing approximate inference.
Auto-Encoding Variational Bayes
- Computer ScienceICLR
- 2014
A stochastic variational inference and learning algorithm that scales to large datasets and, under some mild differentiability conditions, even works in the intractable case is introduced.
Deep Generative Stochastic Networks Trainable by Backprop
- Computer ScienceICML
- 2014
Theorems that generalize recent work on the probabilistic interpretation of denoising autoencoders are provided and obtain along the way an interesting justification for dependency networks and generalized pseudolikelihood.
Stochastic variational inference
- Computer ScienceJ. Mach. Learn. Res.
- 2013
Stochastic variational inference lets us apply complex Bayesian models to massive data sets, and it is shown that the Bayesian nonparametric topic model outperforms its parametric counterpart.
Bayesian Learning via Stochastic Gradient Langevin Dynamics
- Computer ScienceICML
- 2011
In this paper we propose a new framework for learning from large scale datasets based on iterative learning from small mini-batches. By adding the right amount of noise to a standard stochastic…
Deep Gaussian Processes
- Computer ScienceAISTATS
- 2013
Deep Gaussian process (GP) models are introduced and model selection by the variational bound shows that a five layer hierarchy is justified even when modelling a digit data set containing only 150 examples.
A Deep and Tractable Density Estimator
- Computer ScienceICML
- 2014
This work introduces an efficient procedure to simultaneously train a NADE model for each possible ordering of the variables, by sharing parameters across all these models.
Practical Variational Inference for Neural Networks
- Computer ScienceNIPS
- 2011
This paper introduces an easy-to-implement stochastic variational method (or equivalently, minimum description length loss function) that can be applied to most neural networks and revisits several common regularisers from a variational perspective.
Variational Bayesian Inference with Stochastic Search
- Computer ScienceICML
- 2012
This work presents an alternative algorithm based on stochastic optimization that allows for direct optimization of the variational lower bound and demonstrates the approach on two non-conjugate models: logistic regression and an approximation to the HDP.
Unsupervised Variational Bayesian Learning of Nonlinear Models
- Computer ScienceNIPS
- 2004
This paper presents a framework for using multi-layer perceptron (MLP) networks in nonlinear generative models trained by variational Bayesian learning using a Gauss-Hermite quadrature at the hidden neurons for accurate approximation for cases of large posterior variance.