3260 papers • 126 benchmarks • 313 datasets
Yan et al. (2019) CSGN: "When the dancer is stepping, jumping and spinning on the stage, attentions of all audiences are attracted by the streamof the fluent and graceful movements. Building a model that is capable of dancing is as fascinating a task as appreciating the performance itself. In this paper, we aim to generate long-duration human actions represented as skeleton sequences, e.g. those that cover the entirety of a dance, with hundreds of moves and countless possible combinations." ( Image credit: Convolutional Sequence Generation for Skeleton-Based Action Synthesis )
(Image credit: Papersgraph)
These leaderboards are used to track progress in human-action-generation-5
Use these libraries to find human-action-generation-5 models and implementations
The conditional version of generative adversarial nets is introduced, which can be constructed by simply feeding the data, y, to the generator and discriminator, and it is shown that this model can generate MNIST digits conditioned on class labels.
A framework of an autoencoder and a generative adversarial network to produce multiple and consecutive human actions conditioned on the initial state and the given class label is proposed.
This paper focuses on skeleton-based action generation and proposes to model smooth and diverse transitions on a latent space of action sequences with much lower dimensionality, and is learned with a bi-directional generative-adversarial-net framework.
This work proposes a variant of GCNs to leverage the powerful self-attention mechanism to adaptively sparsify a complete action graph in the temporal space, and demonstrates the superiority of this method on two standard human action datasets compared with existing methods.
This paper aims to generate plausible human motion sequences in 3D given a prescribed action type, and proposes a temporal Variational Auto-Encoder (VAE) that encourages a diverse sampling of the motion space.
This work designs a Transformer-based architecture, ACTOR, for encoding and decoding a sequence of parametric SMPL human body models estimated from action recognition datasets, and learns an action-aware latent representation for human motions by training a generative variational autoencoder (VAE).
Kinetic-GAN notably surpasses the state-of-the-art methods in terms of distribution quality metrics while having the ability to synthesise more than one order of magnitude regarding the number of different actions.
This work introduces MUGL, a novel deep neural model for large-scale, diverse generation of single and multi- person pose-based action sequences with locomotion, and incorporates duration-aware feature representations to enable variable-length sequence generation.
Adding a benchmark result helps the community track progress.