3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in sketch-to-image-translation-4
Use these libraries to find sketch-to-image-translation-4 models and implementations
No subtasks available.
S spatially-adaptive normalization is proposed, a simple but effective layer for synthesizing photorealistic images given an input semantic layout that allows users to easily control the style and content of image synthesis results as well as create multi-modal results.
A new method for synthesizing high-resolution photo-realistic images from semantic label maps using conditional generative adversarial networks (conditional GANs) is presented, which significantly outperforms existing methods, advancing both the quality and the resolution of deep image synthesis and editing.
This work introduces the first method for automatic image generation from scene-level freehand sketches that allows for controllable image generation by specifying the synthesis goal via free hand sketches and builds a large-scale composite dataset called SketchyCOCO to support and evaluate the solution.
This paper proposes a simple yet effective open-domain sampling and optimization strategy to "fool" the generator into treating fake sketches as real ones, and shows impressive results in synthesizing realistic color, texture, and maintaining the geometric composition for various categories of open- domain sketches.
Examination of the proposed pretraining-based image-to-image translation (PITI) is shown to be capable of synthesizing images of unprecedented realism and faithfulness, and an adversarial training to enhance the texture synthesis in the diffusion model training is proposed.
The paper shows that intermediate self-attention maps of a masked generative transformer encode important structural information of the input image, such as scene layout and object shape, and proposes a novel sampling method based on this observation to enable structure-guided generation in MaskSketch.
An interactive GAN-based sketch-to-image translation method that helps novice users easily create images of simple objects and introduces a gating-based approach for class conditioning, which allows for distinct classes without feature mixing, from a single generator network.
DeepSIM, a generative model for conditional image manipulation based on a single image, finds that extensive augmentation is key for enabling single image training, and incorporates the use of thin-plate-spline (TPS) as an effective augmentation.
Adding a benchmark result helps the community track progress.