Improved Training of Wasserstein GANs
- Ishaan Gulrajani, Faruk Ahmed, Martín Arjovsky, Vincent Dumoulin, Aaron C. Courville
- Computer ScienceNIPS
- 31 March 2017
This work proposes an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input, which performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning.
Invariant Risk Minimization
- Martín Arjovsky, L. Bottou, Ishaan Gulrajani, David Lopez-Paz
- Computer ScienceArXiv
- 5 July 2019
This work introduces Invariant Risk Minimization, a learning paradigm to estimate invariant correlations across multiple training distributions and shows how the invariances learned by IRM relate to the causal structures governing the data and enable out-of-distribution generalization.
In Search of Lost Domain Generalization
- Ishaan Gulrajani, David Lopez-Paz
- Computer ScienceInternational Conference on Learning…
- 2 July 2020
This paper implements DomainBed, a testbed for domain generalization including seven multi-domain datasets, nine baseline algorithms, and three model selection criteria, and finds that, when carefully implemented, empirical risk minimization shows state-of-the-art performance across all datasets.
Ask Me Anything: Dynamic Memory Networks for Natural Language Processing
- A. Kumar, Ozan Irsoy, R. Socher
- Computer ScienceInternational Conference on Machine Learning
- 24 June 2015
The dynamic memory network (DMN), a neural network architecture which processes input sequences and questions, forms episodic memories, and generates relevant answers, is introduced.
SampleRNN: An Unconditional End-to-End Neural Audio Generation Model
- Soroush Mehri, Kundan Kumar, Yoshua Bengio
- Computer ScienceInternational Conference on Learning…
- 4 November 2016
It is shown that the model, which profits from combining memory-less modules, namely autoregressive multilayer perceptrons, and stateful recurrent neural networks in a hierarchical structure is able to capture underlying sources of variations in the temporal sequences over very long time spans, on three datasets of different nature.
PixelVAE: A Latent Variable Model for Natural Images
- Ishaan Gulrajani, Kundan Kumar, Aaron C. Courville
- Computer ScienceInternational Conference on Learning…
- 4 November 2016
Natural image modeling is a landmark challenge of unsupervised learning. Variational Autoencoders (VAEs) learn a useful latent representation and model global structure well but have difficulty…
GANSynth: Adversarial Neural Audio Synthesis
- Jesse Engel, Kumar Krishna Agrawal, Shuo Chen, Ishaan Gulrajani, Chris Donahue, Adam Roberts
- Computer ScienceInternational Conference on Learning…
- 1 February 2019
Through extensive empirical investigations on the NSynth dataset, it is demonstrated that GANs are able to outperform strong WaveNet baselines on automated and human evaluation metrics, and efficiently generate audio several orders of magnitude faster than their autoregressive counterparts.
Diffusion-LM Improves Controllable Text Generation
- Xiang Lisa Li, John Thickstun, Ishaan Gulrajani, Percy Liang, Tatsunori Hashimoto
- Computer ScienceArXiv
- 27 May 2022
A new non-autoregressive language model based on continuous diffusions that iteratively denoises a sequence of Gaussian vectors into word vectors, yielding a sequences of intermediate latent variables that enables a simple gradient-based algorithm to perform complex, controllable generation tasks.
Towards GAN Benchmarks Which Require Generalization
- Ishaan Gulrajani, Colin Raffel, Luke Metz
- Computer ScienceInternational Conference on Learning…
- 10 January 2020
A necessary condition for an evaluation metric not to behave this way is clarified: estimating the function must require a large sample from the model, so the resulting benchmarks cannot be "won" by training set memorization, while still being perceptually correlated and computable only from samples.
ARGE S AMPLES
- Ishaan Gulrajani, Colin Raffel, Luke Metz
- Computer Science
- 2018
The problem of evaluating a generative model using only a finite sample from the model is studied, and the tradeoff between an evaluation function’s ability to permit meaningful generalization and its ability to be estimated from a finite samples is clarified.
...
...