3260 papers • 126 benchmarks • 313 datasets
Facial Micro-Expression Recognition is a challenging task in identifying suppressed emotion in a high-stake environment, often comes in very brief duration and subtle changes.
(Image credit: Papersgraph)
These leaderboards are used to track progress in micro-expression-recognition-6
Use these libraries to find micro-expression-recognition-6 models and implementations
No subtasks available.
An Enriched Long-term Recurrent Convolutional Network (ELRCN) that first encodes each micro- expression frame into a feature vector through CNN module(s), then predicts the micro-expression by passing the feature vectorthrough a Long Short-term Memory (LSTM) module.
Micro-expressions are spontaneous, brief and subtle facial muscle movements that exposes underlying emotions. Motivated by recent exploits into deep learning for micro-expression analysis, we propose a lightweight dual-stream shallow network in the form of a pair of truncated CNNs with heterogeneous input features. The merging of the convolutional features allows for discriminative learning of micro-expression classes stemming from both streams. Using activation heatmaps, we further demonstrate that salient facial areas are well emphasized, and correspond closely to relevant action units belonging to emotion classes. We empirically validate the proposed network on three benchmark databases, obtaining state-of-the-art performance on the CASME II and SAMM while remaining competitive on the SMIC. Further observations point towards the sufficiency of utilizing shallower deep networks for micro-expression recognition.
The results show that GPT-4V has high accuracy in facial action unit recognition and micro-expression detection while its general facial expression recognition performance is not accurate, and provides valuable insights into the potential applications and challenges of MLLMs in human-centric computing.
This work proposes a novel attention mechanism called micro-attention cooperating with residual network that enables the network to learn to focus on facial areas of interest covering different action units and pushes the boundary of automatic recognition of micro-expression.
A Shallow Triple Stream Three-dimensional CNN (STSTNet) that is computationally light whilst capable of extracting discriminative high level features and details of micro-expressions is designed.
Two 3D-CNN methods are proposed: MicroExpSTCNN and MicroExpFuseNet, for spontaneous facial micro-expression recognition by exploiting the spatiotemporal information in CNN framework, which outperforms the state-of-the-art methods.
A novel Apex-Time Network (ATNet) is proposed to recognize micro-expression based on spatial information from the apex frame as well as on temporalInformation from the respective-adjacent frames so that the model with such temporal information is more robust in cross-dataset validations.
Micro-expressions are reflections of people's true feelings and motives, which attract an increasing number of researchers into the study of automatic facial micro-expression recognition. The short detection window, the subtle facial muscle movements, and the limited training samples make micro-expression recognition challenging. To this end, we propose a novel Identity-aware and Capsule-Enhanced Generative Adversarial Network with graph-based reasoning (ICE-GAN), introducing micro-expression synthesis as an auxiliary task to assist recognition. The generator produces synthetic faces with controllable micro-expressions and identity-aware features, whose long-ranged dependencies are captured through the graph reasoning module (GRM), and the discriminator detects the image authenticity and expression classes. Our ICE-GAN was evaluated on Micro-Expression Grand Challenge 2019 (MEGC2019) with a significant improvement (12.9%) over the winner and surpassed other state-of-the-art methods.
Experimental results show that the proposed TGSR learns the discriminative and explicable regions, and outperforms most state-of-the-art subspace-learning-based domain-adaptive methods for CDMER.
A protocol to automatically synthesize large scale MiE training data that allow us to train improved recognition models for real-world test data and discover three types of Action Units (AUs) that can constitute trainable MiEs.
Adding a benchmark result helps the community track progress.