3260 papers • 126 benchmarks • 313 datasets
A Benchmark for the: Robustness of Object Detection Models to Image Corruptions and Distortions To allow fair comparison of robustness enhancing methods all models have to use a standard ResNet50 backbone because performance strongly scales with backbone capacity. If requested an unrestricted category can be added later. Benchmark Homepage: https://github.com/bethgelab/robust-detection-benchmark Metrics: mPC [AP]: Mean Performance under Corruption [measured in AP] rPC [%]: Relative Performance under Corruption [measured in %] Test sets: Coco: val 2017; Pascal VOC: test 2007; Cityscapes: val; ( Image credit: Benchmarking Robustness in Object Detection )
(Image credit: Papersgraph)
These leaderboards are used to track progress in robust-object-detection
Use these libraries to find robust-object-detection models and implementations
No subtasks available.
We present some updates to YOLO! We made a bunch of little design changes to make it better. We also trained this new network that's pretty swell. It's a little bigger than last time but more accurate. It's still fast though, don't worry. At 320x320 YOLOv3 runs in 22 ms at 28.2 mAP, as accurate as SSD but three times faster. When we look at the old .5 IOU mAP detection metric YOLOv3 is quite good. It achieves 57.9 mAP@50 in 51 ms on a Titan X, compared to 57.5 mAP@50 in 198 ms by RetinaNet, similar performance but 3.8x faster. As always, all the code is online at this https URL
This work introduces a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals and further merge RPN and Fast R-CNN into a single network by sharing their convolutionAL features.
In training, Random Erasing randomly selects a rectangle region in an image and erases its pixels with random values and yields consistent improvement over strong baselines in image classification, object detection and person re-identification.
AugMix significantly improves robustness and uncertainty measures on challenging image classification benchmarks, closing the gap between previous methods and the best possible performance in some cases by more than half.
This work investigates how learned, specialized data augmentation policies improve generalization performance for detection models, and reveals that a learned augmentation policy is superior to state-of-the-art architecture regularization methods for object detection, even when considering strong baselines.
This work builds on the recent state-of-the-art Faster R-CNN model, and design two domain adaptation components, on image level and instance level, to reduce the domain discrepancy, based on $$-divergence theory.
This work proposes Iterative Normalization (IterNorm), which employs Newton’s iterations for much more efficient whitening, while simultaneously avoiding the eigen-decomposition, and exclusively introduces Stochastic Normalization Disturbance (SND), which measures the inherent stochastic uncertainty of samples when applied to normalization operations.
It is shown that a range of standard object detection models suffer a severe performance loss on corrupted images (down to 30--60\% of the original performance), however, a simple data augmentation trick---stylizing the training images---leads to a substantial increase in robustness across corruption type, severity and dataset.
This work proposes to utilize free meta-data in conjunction with associated UAV images to learn domain-robust features via an adversarial training framework dubbed Nuisance Disentangled Feature Transform (NDFT), for the specific challenging problem of object detection in Uav images, achieving a substantial gain in robustness to those nuisances.
IBN-Net is presented, a novel convolutional architecture, which remarkably enhances a CNN’s modeling ability on one domain as well as its generalization capacity on another domain without finetuning.
Adding a benchmark result helps the community track progress.