• Corpus ID: 231879656

Disentangled Representations from Non-Disentangled Models

@article{Khrulkov2021DisentangledRF,
  title={Disentangled Representations from Non-Disentangled Models},
  author={Valentin Khrulkov and Leyla Mirvakhabova and I. Oseledets and Artem Babenko},
  journal={ArXiv},
  year={2021},
  volume={abs/2102.06204}
}
Constructing disentangled representations is known to be a difficult task, especially in the unsupervised scenario. The dominating paradigm of unsupervised disentanglement is currently to train a generative model that separates different factors of variation in its latent space. This separation is typically enforced by training with specific regularization terms in the model’s objective function. These terms, however, introduce additional hyperparameters responsible for the trade-off between… 
Learning Disentangled Representation by Exploiting Pretrained Generative Models: A Contrastive Learning View
TLDR
Disentaglement via Contrast (DisCo) achieves the state-of-the-art disentangled representation learning and distinct direction discovering, given pretrained nondisentangled generative models including GAN, VAE, and Flow.
Generative Modeling Helps Weak Supervision (and Vice Versa)
TLDR
This work proposes a model fusing programmatic weak supervision and generative adversarial networks and provides theoretical justification motivating this fusion, and is the first approach to enable data augmentation through weakly supervised synthetic images and pseudolabels.
Self-supervised Enhancement of Latent Discovery in GANs
TLDR
Scale Ranking Estimator (SRE) is proposed, which is trained using self-supervision and enhances the disentanglement in directions obtained by existing unsupervised disentangled techniques.
Visual Concepts Tokenization
TLDR
An unsupervised transformer-based Visual Concepts Tokenization framework, dubbed VCT, to perceive an image into a set of disentangled visual concept tokens, with each concept token responding to one type of independent visual concept.

References

SHOWING 1-10 OF 54 REFERENCES
Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations
TLDR
This paper theoretically shows that the unsupervised learning of disentangled representations is fundamentally impossible without inductive biases on both the models and the data, and trains more than 12000 models covering most prominent methods and evaluation metrics on seven different data sets.
Learning Deep Disentangled Embeddings with the F-Statistic Loss
TLDR
A new paradigm for discovering disentangled representations of class structure is proposed and a novel loss function based on the $F$ statistic is proposed, which describes the separation of two or more distributions.
InfoGAN-CR and ModelCentrality: Self-supervised Model Training and Selection for Disentangling GANs
TLDR
This work designs a novel approach for training disentangled GANs with self-supervision and proposes contrastive regularizer, which is inspired by a natural notion of disentanglement: latent traversal, and proposes an unsupervised model selection scheme called ModelCentrality, which uses generated synthetic samples to compute the medoid of a collection of models.
Semi-Supervised StyleGAN for Disentanglement Learning
TLDR
The impact of limited supervision is investigated, new metrics to quantify generator controllability are proposed, and there may exist a crucial trade-off between disentangled representation learning and controllable generation.
Variational Inference of Disentangled Latent Concepts from Unlabeled Observations
TLDR
This work considers the problem of unsupervised learning of disentangled representations from large pool of unlabeled observations, and proposes a variational inference based approach to inferdisentangled latent factors.
Disentangling Disentanglement in Variational Autoencoders
We develop a generalisation of disentanglement in VAEs---decomposition of the latent representation---characterising it as the fulfilment of two factors: a) the latent encodings of the data having an
A Sober Look at the Unsupervised Learning of Disentangled Representations and their Evaluation
TLDR
Theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without inductive biases on both the models and the data, and investigate concrete benefits of enforcing disentanglement of the learned representations and consider a reproducible experimental setup covering several data sets.
Structured Disentangled Representations
TLDR
Experiments on a variety of datasets demonstrate that the proposed two-level hierarchical objective can not only disentangle discrete variables, but that doing so also improves disentanglement of other variables and, importantly, generalization even to unseen combinations of factors.
Isolating Sources of Disentanglement in Variational Autoencoders
We decompose the evidence lower bound to show the existence of a term measuring the total correlation between latent variables. We use this to motivate our $\beta$-TCVAE (Total Correlation
The Hessian Penalty: A Weak Prior for Unsupervised Disentanglement
TLDR
A model-agnostic, unbiased stochastic approximation of this term based on Hutchinson's estimator to compute it efficiently during training and provides empirical evidence that the Hessian Penalty encourages substantial shrinkage when applied to over-parameterized latent spaces.
...
...