• Corpus ID: 182953096

What and Where to Translate: Local Mask-based Image-to-Image Translation

  title={What and Where to Translate: Local Mask-based Image-to-Image Translation},
  author={Wonwoong Cho and Seunghwan Choi and Junwoo Park and David Keetae Park and Tao Qin and Jaegul Choo},
Recently, image-to-image translation has obtained significant attention. Among many, those approaches based on an exemplar image that contains the target style information has been actively studied, due to its capability to handle multimodality as well as its applicability in practical use. However, two intrinsic problems exist in the existing methods: what and where to transfer. First, those methods extract style from an entire exemplar which includes noisy information, which impedes a… 


Multimodal Unsupervised Image-to-Image Translation
A Multimodal Unsupervised Image-to-image Translation (MUNIT) framework that assumes that the image representation can be decomposed into a content code that is domain-invariant, and a style code that captures domain-specific properties.
Show, Attend, and Translate: Unsupervised Image Translation With Self-Regularization and Attention
This work constrain the problem with the assumption that the translated image needs to be perceptually similar to the original image and also appears to be drawn from the new domain, and proposes a simple yet effective image translation model consisting of a single generator trained with a self-regularization term and an adversarial term.
Image-To-Image Translation via Group-Wise Deep Whitening-And-Coloring Transformation
An end-to-end approach tailored for image translation that efficiently approximates this transformation with the novel regularization methods is proposed and is fast, both in training and inference, and highly effective in reflecting the style of an exemplar.
Toward Multimodal Image-to-Image Translation
This work aims to model a distribution of possible outputs in a conditional generative modeling setting that helps prevent a many-to-one mapping from the latent code to the output during training, also known as the problem of mode collapse.
Conditional Image-to-Image Translation
This paper twists two conditional translation models together for inputs combination and reconstruction while preserving domain independent features and carries out experiments on men's faces from-to women's faces translation and edges to shoes&bags translations.
Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks
This work presents an approach for learning to translate an image from a source domain X to a target domain Y in the absence of paired examples, and introduces a cycle consistency loss to push F(G(X)) ≈ X (and vice versa).
Unsupervised Attention-guided Image to Image Translation
This work empirically demonstrates that their approach is able to attend to relevant regions in the image without requiring any additional supervision, and that by doing so it achieves more realistic mappings compared to recent approaches.
Unsupervised Image-to-Image Translation Networks
This work makes a shared-latent space assumption and proposes an unsupervised image-to-image translation framework based on Coupled GANs that achieves state-of-the-art performance on benchmark datasets.
Arbitrary Style Transfer in Real-Time with Adaptive Instance Normalization
This paper presents a simple yet effective approach that for the first time enables arbitrary style transfer in real-time, comparable to the fastest existing approach, without the restriction to a pre-defined set of styles.
Image-to-Image Translation with Conditional Adversarial Networks
Conditional adversarial networks are investigated as a general-purpose solution to image-to-image translation problems and it is demonstrated that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks.