3260 papers • 126 benchmarks • 313 datasets
Video Quality Assessment is a computer vision task aiming to mimic video-based human subjective perception. The goal is to produce a mos score, where higher score indicates better perceptual quality. Some well-known benchmarks for this task are KoNViD-1k, LIVE-VQC, YouTube-UGC and LSVQ. SROCC/PLCC/RMSE are usually used to evaluate the performance of different models.
(Image credit: Papersgraph)
These leaderboards are used to track progress in video-quality-assessment-34
Use these libraries to find video-quality-assessment-34 models and implementations
No subtasks available.
This work studies the adversarial robustness of neural networks through the lens of robust optimization, and suggests the notion of security against a first-order adversary as a natural and broad security guarantee.
A new dataset of human perceptual similarity judgments is introduced and it is found that deep features outperform all previous metrics by large margins on this dataset, and suggests that perceptual similarity is an emergent property shared across deep visual representations.
A novel visualization technique is introduced that gives insight into the function of intermediate feature layers and the operation of the classifier in large Convolutional Network models, used in a diagnostic role to find model architectures that outperform Krizhevsky et al on the ImageNet classification benchmark.
The proposed approach relies on the success (and retraining) of proven, state-of-the-art deep object recognition networks and can be used to not only score images reliably and with high correlation to human perception, but also to assist with adaptation and optimization of photo editing/enhancement algorithms in a photographic pipeline.
This paper reports on the 2018 PIRM challenge on perceptual super-resolution (SR), held in conjunction with the Perceptual Image Restoration and Manipulation (PIRM) workshop at ECCV 2018, and concludes with an analysis of the current trends in perceptual SR, as reflected from the leading submissions.
This work conducts a comprehensive evaluation of leading no-reference/blind VQA (BVQA) features and models on a fixed evaluation architecture, yielding new empirical insights on both subjective video quality studies and objective V QA model design.
Grid Mini-patch Sampling (GMS), which allows consideration of local quality by sampling patches at their raw resolution and covers global quality with contextual relations via mini-patches sampled in uniform grids, and enables efficient end-to-end deep VQA and learns effective video-quality-related representations.
This work proposes a unified scheme, spatial-temporal grid mini-cube sampling (St-GMS), and the resultant samples are named fragments, a network architecture tailored specifically for fragments that achieves up to 1612× efficiency than the existing state-of-the-art VQA benchmarks.
This work turns a single unlabeled test sample into a self-supervised learning problem, on which the model parameters are updated before making a prediction, which leads to improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts.
The Disentangled Objective Video Quality Evaluator (DOVER) is proposed, the first approach to provide reliable clear-cut quality evaluations from a single aesthetic or technical perspective, and the first approach to provide reliable clear-cut quality evaluations from a single aesthetic or technical perspective.
Adding a benchmark result helps the community track progress.