3260 papers • 126 benchmarks • 313 datasets
Image: Cudeiro et al
(Image credit: Papersgraph)
These leaderboards are used to track progress in 3d-face-animation-4
Use these libraries to find 3d-face-animation-4 models and implementations
No subtasks available.
Faces Learned with an Articulated Model and Expressions is low-dimensional but more expressive than the FaceWarehouse model and the Basel Face Model and is compared to these models by fitting them to static 3D scans and 4D sequences using the same optimization method.
This work presents the first approach that regresses 3D face shape and animatable details that are specific to an individual but change with expression, and introduces a novel detail-consistency loss that disentangles person-specific details from expression-dependent wrinkles.
A generic audio-driven facial animation approach that achieves highly realistic motion synthesis results for the entire face by disentangles audio-correlated and audio-uncorrelated information based on a novel cross-modality loss.
A novel speech-to-motion generation framework in which the face, body, and hands are modeled separately, and a cross-conditional autoregressive model that generates body poses and hand gestures, leading to coherent and realistic motions is proposed.
An end-to-end neural network to disentangle different emotions in speech so as to generate rich 3D facial expressions and outperforms state-of-the-art methods and exhibits more diverse facial movements.
Experiments demonstrate that EMAGE generates holistic gestures with state-of-the-art performance and is flexible in accepting predefined spatial-temporal gesture inputs, generating complete, audio-synchronized results.
This paper proposes an approach to robustly compute correspondences between a large set of facial motion sequences in a fully automatic way using a multilinear model as statistical prior and synthesizes new motion sequences and performs expression recognition.
A unique 4D face dataset with about 29 minutes of 4D scans captured at 60 fps and synchronized audio from 12 speakers is introduced and VOCA (Voice Operated Character Animation) is learned, the only realistic 3D facial animation model that is readily applicable to unseen subjects without retargeting.
A deep neural network model that takes an audio signal A of a source person and a very short video V of a target person as input, and outputs a synthesized high-quality talking face video with personalized head pose, making use of the visual information in V, expression and lip synchronization.
Experimental results and user studies show the proposed talking face generation method can generate realistic talking face videos with not only synchronized lip motions, but also natural head movements and eye blinks, with better qualities than the results of state-of-the-art methods.
Adding a benchmark result helps the community track progress.