3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in image-generation
Use these libraries to find image-generation models and implementations
No subtasks available.
This work offers a new generation paradigm that allows flexible control of the output image, such as spatial layout and palette, while maintaining the synthesis quality and model creativity, and serves as a general framework and facilitates a wide range of classical generative tasks without retraining.
The novel Pose Guided Person Generation Network (PG$^2$) that allows to synthesize person images in arbitrary poses, based on an image of that person and a novel pose, is proposed.
This work combines 3D animation rendering and a pose transfer model to synthesize conducting video from a single given user's image and proposes Audio Motion Correspondence Network and adversarial-perceptual learning to learn the cross-modal relationship and generate diverse, plausible, music-synchronized motion.
A novel, two-stage reconstruction pipeline is proposed that learns a disentangled representation of the aforementioned image factors and generates novel person images at the same time and can manipulate the foreground, background and pose of the input image, and also sample new embedding features to generate targeted manipulations, that provide more control over the generation process.
This work addresses the problem of guided image-to-image translation where an input image is translated into another while respecting the constraints provided by an external, user-provided guidance image and presents a bi-directional feature transformation (bFT) scheme.
A novel approach for the task of human pose transfer, which aims at synthesizing a new image of a person from an input image of that person and a target pose, is presented, which addresses the issues of limited correspondences identified between keypoints only and invisible pixels due to self-occlusion.
A new generative adversarial network to the problem of pose transfer, i.e., transferring the pose of a given person to a target one, which can generate training images for person re-identification, alleviating data insufficiency.
The Attribute-Decomposed GAN is introduced, a novel generative model for controllable person image synthesis, which can produce realistic person images with desired human attributes provided in various source inputs and its superiority over the state of the art in pose transfer and its effectiveness in the brand-new task of component attribute transfer.
A novel Generative Adversarial Network (XingGAN or CrossingGAN) for person image generation tasks, i.e., translating the pose of a given person to a desired one to mutually improve each other is proposed.
This paper introduces deformable skip connections in the generator of the Generative Adversarial Network and proposes a nearest-neighbour loss instead of the common L1 and L2 losses in order to match the details of the generated image with the target image.
Adding a benchmark result helps the community track progress.