A Style-Based Generator Architecture for Generative Adversarial Networks

@article{Karras2018ASG,
  title={A Style-Based Generator Architecture for Generative Adversarial Networks},
  author={Tero Karras and Samuli Laine and Timo Aila},
  journal={2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2018},
  pages={4396-4405}
}
  • Tero KarrasS. LaineTimo Aila
  • Published 12 December 2018
  • Computer Science
  • 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
We propose an alternative generator architecture for generative adversarial networks, borrowing from style transfer literature. [] Key Method To quantify interpolation quality and disentanglement, we propose two new, automated methods that are applicable to any generator architecture. Finally, we introduce a new, highly varied and high-quality dataset of human faces.

Figures and Tables from this paper

A StyleGAN-2 inspired Generative Adversarial Network for the PCA-controllable generation of drums samples for content-based retrieval

This project validated a StyleGAN-2 inspired architecture for the unlimited generation of high-quality magnitude spectrogram images, for the purpose of content-based retrieval, and demonstrates that when trained on drum loops, some of these controllable latent dimensions directly relate to highly semantic factors.

Unsupervised K-modal styled content generation

The emergence of deep generative models has recently enabled the automatic generation of massive amounts of graphical content, both in 2D and in 3D. Generative Adversarial Networks (GANs) and style

An Image-based Generator Architecture for Synthetic Image Refinement

These are alternative generator architectures for Boundary Equilibrium Generative Adversarial Networks, motivated by Learning from Simulated and Unsupervised Images through adversarial Training, that disentangles the need for a noise-based latent space and attempts to resolve the latent space's poorly understood properties.

Unsupervised multi-modal Styled Content Generation

This paper introduces UMMGAN, a novel architecture designed to better model multi-modal distributions, in an unsupervised fashion, and demonstrates that this approach is capable of effectively approximating a complex distribution as a superposition of multiple simple ones.

Cluster-guided Image Synthesis with Unconditional Models

This work focuses on controllable image generation by leveraging GANs that are well-trained in an unsupervised fashion and discovers that the representation space of intermediate layers of the generator forms a number of clusters that separate the data according to semantically meaningful attributes.

Mask-Guided Discovery of Semantic Manifolds in Generative Models

This work presents a method to explore the manifolds of changes of spatially localized regions of the face and discovers smoothly varying sequences of latent vectors along these manifolds suitable for creating animations.

Unsupervised Controllable Generation with Self-Training

This work proposes an unsupervised framework to learn a distribution of latent codes that control the generator through self-training, and exhibits better disentanglement compared to other variants such as the variational autoencoder, and is able to discover semantically meaningful latent codes without any supervision.

Self-Attention Generative Adversarial Networks

The proposed SAGAN achieves the state-of-the-art results, boosting the best published Inception score from 36.8 to 52.52 and reducing Frechet Inception distance from 27.62 to 18.65 on the challenging ImageNet dataset.

Autoencoding Generative Adversarial Networks

The Autoencoding Generative Adversarial Network (AEGAN), a four-network model which learns a bijective mapping between a specified latent space and a given sample space by applying an adversarial loss and a reconstruction loss to both the generated images and the generated latent vectors is proposed.

OpenGAN: Open Set Generative Adversarial Networks

This work proposes an open set GAN architecture that is conditioned per-input sample with a feature embedding drawn from a metric space, and shows that classifier performance can be significantly improved by augmenting the training data with OpenGAN samples on classes that are outside of the GAN training distribution.
...

References

SHOWING 1-10 OF 65 REFERENCES

Self-Attention Generative Adversarial Networks

The proposed SAGAN achieves the state-of-the-art results, boosting the best published Inception score from 36.8 to 52.52 and reducing Frechet Inception distance from 27.62 to 18.65 on the challenging ImageNet dataset.

Generative Multi-Adversarial Networks

Generative adversarial networks (GANs) are a framework for producing a generative model by way of a two-player minimax game. In this paper, we propose the \emph{Generative Multi-Adversarial Network}

Gaussian Mixture Generative Adversarial Networks for Diverse Datasets, and the Unsupervised Clustering of Images

Gaussian Mixture GAN (GM-GAN) is proposed, a variant of the basic GAN model, where the probability distribution over the latent space is a mixture of Gaussians, and a feature is demonstrated which further sets this model apart from other GAN models: the option to control the quality-diversity trade-off by altering, post-training, the likelihood distribution of the latentspace.

Generative adversarial interpolative autoencoding: adversarial training on latent space interpolations encourage convex latent distributions

A neural network architecture based upon the Autoencoder and Generative Adversarial Network that promotes a convex latent distribution by training adversarially on latent space interpolations to preserve realistic resemblances to the network inputs is presented.

On Self Modulation for Generative Adversarial Networks

This work proposes and study an architectural modification, self-modulation, which improves GAN performance across different data sets, architectures, losses, regularizers, and hyperparameter settings.

Online Adaptative Curriculum Learning for GANs

Experimental results show that the proposed novel framework for training the generator against an ensemble of discriminator networks improves samples quality and diversity over existing baselines by effectively learning a curriculum and supports the claim that weaker discriminators have higher entropy improving modes coverage.

The GAN Landscape: Losses, Architectures, Regularization, and Normalization

This work reproduces the current state of the art of GANs from a practical perspective, discusses common pitfalls and reproducibility issues, and goes beyond fairly exploring the GAN landscape.

Progressive Growing of GANs for Improved Quality, Stability, and Variation

A new training methodology for generative adversarial networks is described, starting from a low resolution, and adding new layers that model increasingly fine details as training progresses, allowing for images of unprecedented quality.

Improved Training of Wasserstein GANs

This work proposes an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input, which performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning.

Adversarial Feature Learning

Bidirectional Generative Adversarial Networks are proposed as a means of learning the inverse mapping of GANs, and it is demonstrated that the resulting learned feature representation is useful for auxiliary supervised discrimination tasks, competitive with contemporary approaches to unsupervised and self-supervised feature learning.
...