3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in hand-gesture-recognition-17
Use these libraries to find hand-gesture-recognition-17 models and implementations
No subtasks available.
A hierarchical structure enabling offline-working convolutional neural network (CNN) architectures to operate online efficiently by using sliding window approach is proposed by proposing a lightweight CNN architecture to detect gestures and a classifier which is a deep CNN to classify the detected gestures.
This work proposes a two-stage convolutional neural network architecture for robust recognition of hand gestures, called HGR-Net, where the first stage performs accurate semantic segmentation to determine hand regions, and the second stage identifies the gesture.
This paper proposes a new method of interaction with computing devices having a consumer grade camera that uses two colored markers worn on tips of the fingers to generate desired hand gestures, and for marker detection and tracking the authors used template matching with kalman filter.
Two new deep models termed as F-BLSTM and F-BGRU are proposed, which can effectively classify the gesture based on analyzing the acceleration and angular velocity data of the human gestures.
The 3D ResNets trained on the Kinetics did not suffer from overfitting despite the large number of parameters of the model, and achieved better performance than relatively shallow networks, such as C3D.
This paper proposes a data level fusion strategy, Motion Fused Frames (MFFs), designed to fuse motion information into static images as better representatives of spatio-temporal states of an action.
This work presents an efficient approach for leveraging the knowledge from multiple modalities in training unimodal 3D convolutional neural networks (3D-CNNs) for the task of dynamic hand gesture recognition, and introduces a "spatiotemporal semantic alignment" loss (SSA) to align the content of the features from different networks.
This work combines image entropy and density clustering to exploit the key frames from hand gesture video for further feature extraction, which can improve the efficiency of recognition.
A Dynamic Graph-Based Spatial-Temporal Attention (DG-STA) method for hand gesture recognition to first construct a fully-connected graph from a hand skeleton, where the node features and edges are automatically learned via a self-attention mechanism that performs in both spatial and temporal domains.
Adding a benchmark result helps the community track progress.