3260 papers • 126 benchmarks • 313 datasets
Image source: Multi-View Motion Synthesis via Applying Rotated Dual-Pixel Blur Kernels
(Image credit: Papersgraph)
These leaderboards are used to track progress in motion-synthesis-7
Use these libraries to find motion-synthesis-7 models and implementations
It is shown that, surprisingly, state of the art performance can be achieved by a simple baseline that does not attempt to model motion at all, and a simple and scalable RNN architecture is proposed that obtains state-of-the-art performance on human motion prediction.
A novel sequence-to-sequence model for probabilistic human motion prediction, trained with a modified version of improved Wasserstein generative adversarial networks (WGAN-GP), in which the model learns a probability density function of future human poses conditioned on previous poses.
This work follows the trend of rendering the NIMAT effect by introducing a modification on the blur synthesis procedure in portrait mode that enables a high-quality synthesis of multi-view bokeh from a single image by applying rotated blurring kernels.
Experimental qualitative and quantitative results demonstrate that the proposed synthesis-by-analysis learning framework can synthesize realistic, diverse, style-consistent, and beat-matching dances from music.
This work designs a Transformer-based architecture, ACTOR, for encoding and decoding a sequence of parametric SMPL human body models estimated from action recognition datasets, and learns an action-aware latent representation for human motions by training a generative variational autoencoder (VAE).
A generic audio-driven facial animation approach that achieves highly realistic motion synthesis results for the entire face by disentangles audio-correlated and audio-uncorrelated information based on a novel cross-modality loss.
The proposed MotionDiffuse is one of the first diffusion model-based text-driven motion generation frameworks, which demonstrates several desired properties over existing methods, and outperforms existing SoTA methods by convincing margins on text-driven motion generation and action-conditioned motion generation.
This paper introduces three forms of composition based on diffusion priors: sequential, parallel, and model composition, and introduces DiffusionBlending, an interpolation mechanism to effectively blend several such models to enable flexible and efficient fine-grained joint and trajectory-level control and editing.
Adding a benchmark result helps the community track progress.