3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in video-emotion-recognition-31
No benchmarks available.
Use these libraries to find video-emotion-recognition-31 models and implementations
No subtasks available.
This paper proposes to combine micro- and macro-motion features to improve video emotion recognition with a two-stream recurrent network, named MIMAMO (Micro-Macro-Motion) Net, achieves state of the art performance on two video emotion datasets, the OMG emotion dataset and the Aff-Wild dataset.
This work attempts to explore different neural networks to improve accuracy of emotion recognition and finds (CNN+RNN) + 3DCNN multi-model architecture which processes audio spectrograms and corresponding video frames giving emotion prediction accuracy of 54.0% among 4 emotions and 71.75% among 3 emotions using IEMOCAP[2] dataset.
This work uses MIMAMO Net \cite{deng2020mimamo} model to achieve information about micro-motion and macro-motion for improving video emotion recognition and achieve Concordance Correlation Coefficient of 0.415 and 0.511 for valence and arousal on the reselected validation set.
A fully multimodal video-to-emotion system for fast yet effective recognition inference, whose benefits are threefold: the adoption of the hierarchical attention method upon the sound spectra breaks through the limited contribution of the acoustic modality, and outperforms the existing models’ performance on both IEMOCAP and CMU-MOSEI datasets.
Adding a benchmark result helps the community track progress.