3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in video-alignment-9
Use these libraries to find video-alignment-9 models and implementations
No subtasks available.
A self-supervised approach for learning representations and robotic behaviors entirely from unlabeled videos recorded from multiple viewpoints is proposed, and it is demonstrated that this representation can be used by a robot to directly mimic human poses without an explicit correspondence, and that it can be use as a reward function within a reinforcement learning algorithm.
This work applies the proposed method to the problem of weakly supervised learning of actions and actors from movies together with corresponding movie scripts and proposes an online optimization algorithm based on the Block-Coordinate Frank-Wolfe algorithm.
It is shown that the learned embeddings enable few-shot classification of these action phases, significantly reducing the supervised training requirements; and TCC is complementary to other methods of self-supervised learning in videos, such as Shuffle and Learn and Time-Contrastive Networks.
An approach for learning a compact view-invariant embedding space from 2D joint keypoints alone, without explicitly predicting 3D poses, and uses probabilistic embeddings to model this input uncertainty.
This work proposes an approach to learning a compact view-invariant embedding space from 2D body joint keypoints, without explicitly predicting 3D poses, and investigates different keypoint occlusion augmentation strategies during training.
This paper considers a learnable approach for comparing and aligning videos. Our architecture builds upon and revisits temporal match kernels within neural networks: we propose a new temporal layer that finds temporal alignments by maximizing the scores between two sequences of vectors, according to a time-sensitive similarity metric parametrized in the Fourier domain. We learn this layer with a temporal proposal strategy, in which we minimize a triplet loss that takes into account both the localization accuracy and the recognition rate. We evaluate our approach on video alignment, copy detection and event retrieval. Our approach outperforms the state on the art on temporal video alignment and video copy detection datasets in comparable setups. It also attains the best reported results for particular event search, while precisely aligning videos.
This work presents an audio-to-video method for automating speech to lips alignment, stretching and compressing the audio signal to match the lip movements, based on deep audio-visual features.
This work proposes a novel approach to learn a task-agnostic skill embedding space from unlabeled multi-view videos by using an adversarial loss, and shows that the learned embedding enables training of continuous control policies to solve novel tasks that require the interpolation of previously seen skills.
This paper introduces a novel contrastive action representation learning (CARL) framework to learn frame-wise action representations, especially for long videos, in a self-supervised manner and shows outstanding performance on video alignment and fine-grained frame retrieval tasks.
A 3D Token Representation Layer (3DTRL) is proposed that estimates the 3D positional information of the visual tokens and leverages it for learning viewpoint-agnostic representations and outperform their backbone Transformers in all the tasks with minimal added computation.
Adding a benchmark result helps the community track progress.