3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in one-shot-3d-action-recognition-5
Use these libraries to find one-shot-3d-action-recognition-5 models and implementations
No subtasks available.
This work introduces a large-scale dataset for RGB+D human action recognition, which is collected from 106 distinct subjects and contains more than 114 thousand video samples and 8 million frames, and investigates a novel one-shot 3D activity recognition problem on this dataset.
This work forms the one-shot action recognition problem as a deep metric learning problem and proposes a novel image-based skeleton representation that performs well in a metric learning setting and trains a model that projects the image representations into an embedding space.
This work proposes a metric learning approach to reduce the action recognition problem to a nearest neighbor search in embedding space, which generalizes well in experiments on the UTD-MHAD dataset for inertial, skeleton and fused data and the Simitate dataset for motion capturing data.
A pretraining stage in which a motion encoder is trained to recover the underlying 3D motion from noisy partial 2D observations is proposed, which achieves state-of-the-art performance on all three downstream tasks by simply finetuning the pretrained motion encoding with a simple regression head, which demonstrates the versatility of the learned motion representations.
This work presents a novel approach for one-shot action recognition in the wild that computes motion representations robust to variable kinematic conditions and develops a set of complementary steps that boost the action recognition performance in the most challenging scenarios.
Adding a benchmark result helps the community track progress.