1
DaViT: Dual Attention Vision Transformers
3
On Causally Disentangled Representations
4
Swin Transformer V2: Scaling Up Capacity and Resolution
5
Are Transformers More Robust Than CNNs?
6
Adversarial Robustness Comparison of Vision Transformer and MLP-Mixer to CNNs
7
Do Vision Transformers See Like Convolutional Neural Networks?
8
ConvNets vs. Transformers: Whose Visual Representations are More Transferable?
9
BEiT: BERT Pre-Training of Image Transformers
10
Reveal of Vision Transformers Robustness against Adversarial Attacks
11
Intriguing Properties of Vision Transformers
12
Vision Transformers are Robust Learners
13
Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction without Convolutions
14
Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet
15
Training data-efficient image transformers & distillation through attention
16
Learning Contextual Causality from Time-consecutive Images
17
MorphGAN: One-Shot Face Synthesis GAN for Detecting Recognition Bias
18
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
19
ThreeDWorld: A Platform for Interactive Multi-Modal Physical Simulation
20
Gender Slopes: Counterfactual Fairness for Computer Vision Models by Attribute Manipulation
21
Compositional Convolutional Neural Networks: A Deep Architecture With Innate Robustness to Partial Occlusion
22
Natural Adversarial Examples
23
Image Counterfactual Sensitivity Analysis for Detecting Unintended Bias
24
On the Transfer of Inductive Bias from Simulation to the Real World: a New Disentanglement Dataset
25
Combining Compositional Models and Deep Networks For Robust Object Classification under Occlusion
26
EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks
27
Robustness of Object Recognition under Extreme Occlusion in Humans and Computational Models
28
Meta-Sim: Learning to Generate Synthetic Datasets
29
Counterfactual Visual Explanations
30
Synthetic Examples Improve Generalization for Rare Classes
31
Counterfactual Sensitivity and Robustness
32
Benchmarking Neural Network Robustness to Common Corruptions and Perturbations
33
Moment Matching for Multi-Source Domain Adaptation
34
Strike (With) a Pose: Neural Networks Are Easily Fooled by Strange Poses of Familiar Objects
36
ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness
37
Learning dexterous in-hand manipulation
38
VisDA: A Synthetic-to-Real Benchmark for Visual Domain Adaptation
39
Synthesizing Programs for Images using Reinforced Adversarial Learning
40
Learning to Adapt Structured Output Space for Semantic Segmentation
41
Path-Specific Counterfactual Fairness
42
Disentangling by Factorising
43
Empirically Analyzing the Effect of Dataset Biases on Deep Face Recognition Systems
44
CyCADA: Cycle-Consistent Adversarial Domain Adaptation
45
CARLA: An Open Urban Driving Simulator
46
Adversarial Variational Optimization of Non-Differentiable Simulators
47
No More Discrimination: Cross City Adaptation of Road Scene Segmenters
48
Counterfactual Fairness
49
Adversarial Discriminative Domain Adaptation
50
CLEVR: A Diagnostic Dataset for Compositional Language and Elementary Visual Reasoning
51
Densely Connected Convolutional Networks
52
Playing for Data: Ground Truth from Computer Games
53
The SYNTHIA Dataset: A Large Collection of Synthetic Images for Semantic Segmentation of Urban Scenes
54
VirtualWorlds as Proxy for Multi-object Tracking Analysis
55
Deep Residual Learning for Image Recognition
56
A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation
57
Domain-Adversarial Training of Neural Networks
58
Very Deep Convolutional Networks for Large-Scale Image Recognition
59
ImageNet classification with deep convolutional neural networks
60
3D Object Detection and Viewpoint Estimation with a Deformable 3D Cuboid Model
61
ImageNet: A large-scale hierarchical image database
63
Learning methods for generic object recognition with invariance to pose and lighting
65
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
66
ObjectNet: A large-scale bias-controlled dataset for pushing the limits of object recognition models
67
NVLabs. Falor3d - isaac3d
68
dsprites: Disentanglement testing sprites dataset
69
Establishing Good Benchmarks and Baselines for Face Recognition
70
Gradient-based learning applied to document recognition
71
Did you discuss any potential negative societal impacts of your work? [Yes] Extended discussion in supp. material
72
(a) Did you state the full set of assumptions of all theoretical results
73
Do the main claims made in the abstract and introduction accurately reflect the paper's contributions and scope?
74
Did you report error bars (e.g., with respect to the random seed after running experiments multiple times
75
Have you read the ethics review guidelines and ensured that your paper conforms to them
76
Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content?
77
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable?
78
If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots
79
Did you describe the limitations of your work
80
c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation?