Corpus ID: 207869889

Cali-Sketch: Stroke Calibration and Completion for High-Quality Face Image Generation from Poorly-Drawn Sketches

@article{Xia2019CaliSketchSC,
  title={Cali-Sketch: Stroke Calibration and Completion for High-Quality Face Image Generation from Poorly-Drawn Sketches},
  author={Weihao Xia and Yujiu Yang and Jing-Hao Xue},
  journal={ArXiv},
  year={2019},
  volume={abs/1911.00426}
}
Image generation task has received increasing attention because of its wide application in security and entertainment. Sketch-based face generation brings more fun and better quality of image generation due to supervised interaction. However, When a sketch poorly aligned with the true face is given as input, existing supervised image-to-image translation methods often cannot generate acceptable photo-realistic face images. To address this problem, in this paper we propose Cali-Sketch, a poorly… Expand
TediGAN: Text-Guided Diverse Face Image Generation and Manipulation
TLDR
This work proposes TediGAN, a novel framework for multi-modal image generation and manipulation with textual descriptions using a control mechanism based on style-mixing, and proposes the Multi-Modal CelebA-HQ, a large-scale dataset consisting of real face images and corresponding semantic segmentation map, sketch, and textual descriptions. Expand
Towards Open-World Text-Guided Face Image Generation and Manipulation
TLDR
This work proposes a unified framework for both face image generation and manipulation that produces diverse and high-quality images with an unprecedented resolution at 1024 from multimodal inputs and supports open-world scenarios, including both image and text, without any re-training, fine-tuning, or post-processing. Expand
TediGAN: Text-Guided Diverse Image Generation and Manipulation
TLDR
This work proposes TediGAN, a novel framework for multi-modal image generation and manipulation with textual descriptions, and proposes the Multi-Modal CelebA-HQ, a large-scale dataset consisting of real face images and corresponding semantic segmentation map, sketch, and textual descriptions. Expand
Deep Learning for Free-Hand Sketch: A Survey
TLDR
A comprehensive survey of the deep learning techniques oriented at free-hand sketch data, and the applications that they enable. Expand
M6-UFC: Unifying Multi-Modal Controls for Conditional Image Synthesis
TLDR
A new two-stage architecture, M6UFC, is proposed to unify any number of multi-modal controls to enhance the holistic consistency of the synthesized image, to support preserving specified image blocks, and to improve the synthesis speed. Expand
UFC-BERT: Unifying Multi-Modal Controls for Conditional Image Synthesis
TLDR
A new two-stage architecture to unify any number of multi-modal controls, UFC-BERT, which adopts non-autoregressive generation at the second stage to enhance the holistic consistency of the synthesized image, to support preserving specified image blocks, and to improve the synthesis speed. Expand
Deep Image Synthesis from Intuitive User Input: A Review and Perspectives
TLDR
This paper reviews recent works for image synthesis given intuitive user input, covering advances in input versatility, image generation methodology, benchmark datasets, and evaluation metrics. Expand
GAN Inversion: A Survey
TLDR
This paper provides an overview of GAN inversion with a focus on its recent algorithms and applications and further elaborate on some trends and challenges for future directions. Expand

References

SHOWING 1-10 OF 70 REFERENCES
High-Quality Facial Photo-Sketch Synthesis Using Multi-Adversarial Networks
TLDR
A novel synthesis framework called Photo-Sketch Synthesis using Multi-Adversarial Networks, (PS2-MAN) that iteratively generates low resolution to high resolution images in an adversarial way that leverages the pair information using CycleGAN framework. Expand
Image Generation from Sketch Constraint Using Contextual GAN
TLDR
This paper addresses image generation guided by hand sketch using a novel joint image completion approach, where the sketch provides the image context for completing, or generating the output image, using a generated adversarial network. Expand
Sketch-to-Image Generation Using Deep Contextual Completion
TLDR
The proposed sketch-to-image generation, where the output edges do not necessarily follow the input edges, can generate more realistic images than the state-ofthe- arts on challenging inputs and generalize well on common categories. Expand
SketchyGAN: Towards Diverse and Realistic Sketch to Image Synthesis
TLDR
This work proposes a novel Generative Adversarial Network approach that synthesizes plausible images from 50 categories including motorcycles, horses and couches and introduces a new network building block suitable for both the generator and discriminator which improves the information flow by injecting the input image at multiple scales. Expand
SC-FEGAN: Face Editing Generative Adversarial Network With User’s Sketch and Color
TLDR
This work trained the network with an additional style loss, which made it possible to generate realistic results despite large portions of the image being removed, and is well suited for generating high-quality synthetic images using intuitive user inputs. Expand
High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs
TLDR
A new method for synthesizing high-resolution photo-realistic images from semantic label maps using conditional generative adversarial networks (conditional GANs) is presented, which significantly outperforms existing methods, advancing both the quality and the resolution of deep image synthesis and editing. Expand
PoseShop: Human Image Database Construction and Personalized Content Synthesis
TLDR
PoseShop - a pipeline to construct segmented human image database with minimal manual intervention is presented, which contains 400 thousands human figures that are segmented out of their background. Expand
Free-Form Image Inpainting With Gated Convolution
TLDR
The proposed gated convolution solves the issue of vanilla convolution that treats all input pixels as valid ones, generalizes partial convolution by providing a learnable dynamic feature selection mechanism for each channel at each spatial location across all layers. Expand
Quality-Aware Unpaired Image-to-Image Translation
TLDR
A unifiedquality-aware GAN-based framework for unpaired image-to-image translation, where a quality-aware loss is explicitly incorporated by comparing each source image and the reconstructed image at the domain level is proposed. Expand
InstaGAN: Instance-aware Image-to-Image Translation
TLDR
A novel method is proposed, coined instance-aware GAN (InstaGAN), that incorporates the instance information and improves multi-instance transfiguration and introduces a context preserving loss that encourages the network to learn the identity function outside of target instances. Expand
...
1
2
3
4
5
...