• Corpus ID: 195767445

# The Thermodynamic Variational Objective

@inproceedings{Masrani2019TheTV,
title={The Thermodynamic Variational Objective},
author={Vaden Masrani and Tuan Anh Le and Frank D. Wood},
booktitle={Neural Information Processing Systems},
year={2019}
}
• Published in
Neural Information Processing…
28 June 2019
• Computer Science
We introduce the thermodynamic variational objective (TVO) for learning in both continuous and discrete deep generative models. The TVO arises from a key connection between variational inference and thermodynamic integration that results in a tighter lower bound to the log marginal likelihood than the standard variational variational evidence lower bound (ELBO) while remaining as broadly applicable. We provide a computationally efficient gradient estimator for the TVO that applies to continuous…

## Figures and Tables from this paper

• Computer Science
ICML
• 2020
An exponential family interpretation of the geometric mixture curve underlying the TVO and various path sampling methods is proposed, which allows the gap in TVO likelihood bounds as a sum of KL divergences and derives a doubly reparameterized gradient estimator which improves model learning and allows the TVo to benefit from more refined bounds.
• Computer Science
NeurIPS
• 2021
NVI is developed, a family of methods that learn proposals for nested importance samplers by minimizing an forward or reverse KL divergence at each level of nesting, and it is observed that optimizing nested objectives leads to improved sample quality in terms of log average weight and effective sample size.
• Computer Science
NeurIPS
• 2020
This paper introduces a bespoke Gaussian process bandit optimization method that automates their one-time selection, but also dynamically adapts their positions over the course of optimization, leading to improved model learning and inference.
• Computer Science
ICML
• 2020
A novel vari­ ational inference algorithm for sequential data is presented, which performs well even when the density from the model is not differentiable, for instance, due to the use of discrete random variables.
• Computer Science
ArXiv
• 2022
This paper builds bridges between two families of probabilistic algorithms: (hi-erarchical) variational inference (VI), which is typically used to model distributions over continuous spaces, and
• Computer Science
NeurIPS
• 2020
NVAE is the first successful VAE applied to natural images as large as 256$\times$256 pixels and achieves state-of-the-art results among non-autoregressive likelihood-based models on the MNIST, CIFAR-10, CelebA 64, and CelebA HQ datasets and it provides a strong baseline on FFHQ.
• Computer Science
ICML
• 2022
This work argues theoretically that the resulting algorithm allows an intuitive trade-off between inference and computational cost, and shows that it performs well in practice and is well-suited for black-box inference in probabilistic programming frameworks.
• Computer Science
ArXiv
• 2020
Using importance weighting and an analysis of the objective of semi-supervised VAEs, it is shown that they use the posterior of the learned generative model to guide the inference model in learning the partially observed latent variable.
• Computer Science
CIKM
• 2021
This paper proposes a novel Semi-deterministic and Contrastive Variational Graph autoencoder (SCVG) for item recommendation, and empirically shows that the contrastive regularization makes learned user/item latent representation more personalized and helps to smooth the training process.
• Computer Science
ArXiv
• 2020
It is demonstrated that EnKOs outperform the SMC based methods in terms of predictive ability for three benchmark nonlinear dynamics systems tasks and can identify the latent dynamics given fewer particles because of its rich particle diversity.

## References

SHOWING 1-10 OF 47 REFERENCES

• Computer Science
AISTATS
• 2019
UIVI considers an implicit variational distribution obtained in a hierarchical manner using a simple reparameterizable distribution whose variational parameters are defined by arbitrarily flexible deep neural networks and directly optimizes the evidence lower bound (ELBO).
• Computer Science
AISTATS
• 2018
The VSMC family is a variational family that can approximate the posterior arbitrarily well, while still allowing for efficient optimization of its parameters, and is demonstrated its utility on state space models, stochastic volatility models for financial data, and deep Markov models of brain neural circuits.
• Computer Science
ICML
• 2016
The first unbiased gradient estimator designed for importance-sampled objectives is developed, which is both simpler and more effective than the NVIL estimator proposed for the single-sample variational objective, and is competitive with the currently used biases.
• Computer Science
ICML
• 2014
This work proposes a fast non-iterative approximate inference method that uses a feedforward network to implement efficient exact sampling from the variational posterior and shows that it outperforms the wake-sleep algorithm on MNIST and achieves state-of-the-art results on the Reuters RCV1 document dataset.
• Computer Science
ICML
• 2018
We provide theoretical and empirical evidence that using tighter evidence lower bounds (ELBOs) can be detrimental to the process of learning an inference network by reducing the signal-to-noise ratio
• Computer Science
NIPS
• 2017
A family of lower bounds defined by a particle filter's estimator of the marginal likelihood, the filtering variational objectives (FIVOs), are considered, which take the same arguments as the ELBO, but can exploit a model's sequential structure to form tighter bounds.
• Computer Science
ICLR
• 2019
A computationally efficient, unbiased drop-in gradient estimator that reduces the variance of the IWAE gradient, the reweighted wake-sleep update (RWS), and the jackknife variational inference (JVI) gradient (Nowozin, 2018).
• Computer Science
ICML
• 2016
A new model is proposed which guarantees that the top-down and bottom-up distributions can efficiently invert each other, and which results in state of the art generative models which prefer significantly deeper architectures while it allows for orders of magnitude more efficient approximate inference.
• Computer Science, Mathematics
AISTATS
• 2017
This work proposes a new method that lets us leverage reparameterization gradients even when variables are outputs of a acceptance-rejection sampling algorithm, and shows that the variance of the estimator of the gradient is significantly lower than other state-of-the-art methods, leading to faster convergence of stochastic gradient variational inference.
• Computer Science
ICML '08
• 2008
It is shown that Annealed Importance Sampling (AIS) can be used to efficiently estimate the partition function of an RBM, and a novel AIS scheme for comparing RBM's with different architectures is presented.