Corpus ID: 3783306

Unsupervised Image-to-Image Translation Networks

@inproceedings{Liu2017UnsupervisedIT,
  title={Unsupervised Image-to-Image Translation Networks},
  author={Ming-Yu Liu and Thomas M. Breuel and Jan Kautz},
  booktitle={NIPS},
  year={2017}
}
Unsupervised image-to-image translation aims at learning a joint distribution of images in different domains by using images from the marginal distributions in individual domains. [...] Key Method We compare the proposed framework with competing approaches and present high quality image translation results on various challenging unsupervised image translation tasks, including street scene image translation, animal image translation, and face image translation. We also apply the proposed framework to domain…Expand
Multimodal Image-to-Image Translation via Mutual Information Estimation and Maximization
TLDR
A novel framework that can achieve multimodal image-to-image translation by simply encouraging the statistical dependence between the latent code and the output image in conditional generative adversarial networks by incorporating a U-net generator is presented. Expand
In2I: Unsupervised Multi-Image-to-Image Translation Using Generative Adversarial Networks
TLDR
This paper introduces a Generative Adversarial Network (GAN) based framework along with a multi-modal generator structure and a new loss term, latent consistency loss, and shows that leveraging multiple inputs generally improves the visual quality of the translated images. Expand
Unsupervised Image-to-Image Translation Using Domain-Specific Variational Information Bound
TLDR
This work proposes an unsupervised image-to-image translation framework which maximizes a domain-specific variational information bound and learns the target domain-invariant representation of the two domain. Expand
Unsupervised Image-to-Image Translation with Self-Attention Networks
  • Taewon Kang, K. H. Lee
  • Computer Science
  • 2020 IEEE International Conference on Big Data and Smart Computing (BigComp)
  • 2020
TLDR
This paper proposes an unsupervised image-to-image translation with self-attention networks, in which long range dependency helps to not only capture strong geometric change but also generate details using cues from all feature locations, and qualitatively and quantitatively shows superiority of the proposed method compared to existing state-of-the-art unsuper supervised image- to- image translation task. Expand
Exemplar Guided Unsupervised Image-to-Image Translation
TLDR
The Exemplar Guided UNsupervised Image-to-image Translation (EGUNIT) network is proposed which conditions the image translation process on an image in the target domain and transfers style information specific to one domain through exemplar-based adaptive instance normalization. Expand
Rethinking the Truly Unsupervised Image-to-Image Translation
TLDR
This paper proposes the truly unsupervised image-to-image translation method (TUNIT) that simultaneously learns to separate image domains via an information-theoretic approach and generate corresponding images using the estimated domain labels. Expand
Multimodal Unsupervised Image-to-Image Translation
TLDR
A Multimodal Unsupervised Image-to-image Translation (MUNIT) framework that assumes that the image representation can be decomposed into a content code that is domain-invariant, and a style code that captures domain-specific properties. Expand
Unaligned Image-to-Image Translation by Learning to Reweight
  • Shaoan Xie, Mingming Gong, Yanwu Xu, Kun Zhang
  • Computer Science
  • ArXiv
  • 2021
TLDR
This paper proposes to select images based on importance reweighting and develop a method to learn the weights and perform translation simultaneously and automatically and compares the proposed method with state-of-the-art image translation approaches and presents qualitative and quantitative results on different tasks with unaligned domains. Expand
Unsupervised Multi-Domain Multimodal Image-to-Image Translation with Explicit Domain-Constrained Disentanglement
TLDR
This paper proposes a unified framework for learning to generate diverse outputs using unpaired training data and allow for simultaneous multi-domain translation via a single model and investigates how to extract domain-level signal as explicit supervision so as to achieve better image-to-image translation. Expand
Exemplar Guided Unsupervised Image-to-Image Translation with Semantic Consistency
TLDR
The Exemplar Guided & Semantically Consistent Image-to-image Translation (EGSC-IT) network is proposed which conditions the translation process on an exemplar image in the target domain and introduces the concept of feature masks that provide coarse semantic guidance without requiring the use of any semantic labels. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 42 REFERENCES
Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks
TLDR
This work presents an approach for learning to translate an image from a source domain X to a target domain Y in the absence of paired examples, and introduces a cycle consistency loss to push F(G(X)) ≈ X (and vice versa). Expand
Image-to-Image Translation with Conditional Adversarial Networks
TLDR
Conditional adversarial networks are investigated as a general-purpose solution to image-to-image translation problems and it is demonstrated that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks. Expand
Unsupervised Cross-Domain Image Generation
TLDR
The Domain Transfer Network (DTN) is presented, which employs a compound loss function that includes a multiclass GAN loss, an f-constancy component, and a regularizing component that encourages G to map samples from T to themselves. Expand
Conditional Image Generation with PixelCNN Decoders
TLDR
The gated convolutional layers in the proposed model improve the log-likelihood of PixelCNN to match the state-of-the-art performance of PixelRNN on ImageNet, with greatly reduced computational cost. Expand
Coupled Generative Adversarial Networks
TLDR
This work proposes coupled generative adversarial network (CoGAN), which can learn a joint distribution without any tuple of corresponding images, and applies it to several joint distribution learning tasks, and demonstrates its applications to domain adaptation and image transformation. Expand
Perceptual Losses for Real-Time Style Transfer and Super-Resolution
TLDR
This work considers image transformation problems, and proposes the use of perceptual loss functions for training feed-forward networks for image transformation tasks, and shows results on image style transfer, where aFeed-forward network is trained to solve the optimization problem proposed by Gatys et al. in real-time. Expand
Image Style Transfer Using Convolutional Neural Networks
TLDR
A Neural Algorithm of Artistic Style is introduced that can separate and recombine the image content and style of natural images and provide new insights into the deep image representations learned by Convolutional Neural Networks and demonstrate their potential for high level image synthesis and manipulation. Expand
Attribute2Image: Conditional Image Generation from Visual Attributes
TLDR
A layered generative model with disentangled latent variables that can be learned end-to-end using a variational auto-encoder is developed and shows excellent quantitative and visual results in the tasks of attribute-conditioned image reconstruction and completion. Expand
Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space
TLDR
This paper introduces an additional prior on the latent code, improving both sample quality and sample diversity, leading to a state-of-the-art generative model that produces high quality images at higher resolutions than previous generative models, and does so for all 1000 ImageNet categories. Expand
Learning from Simulated and Unsupervised Images through Adversarial Training
TLDR
This work develops a method for S+U learning that uses an adversarial network similar to Generative Adversarial Networks (GANs), but with synthetic images as inputs instead of random vectors, and makes several key modifications to the standard GAN algorithm to preserve annotations, avoid artifacts, and stabilize training. Expand
...
1
2
3
4
5
...