1
Masked Autoencoders for Point Cloud Self-supervised Learning
2
CrossPoint: Self-Supervised Cross-Modal Contrastive Learning for 3D Point Cloud Understanding
3
data2vec: A General Framework for Self-supervised Learning in Speech, Vision and Language
4
UniFormer: Unifying Convolution and Self-Attention for Visual Recognition
5
Masked Feature Prediction for Self-Supervised Visual Pre-Training
6
PointCLIP: Point Cloud Understanding by CLIP
7
Point-BERT: Pre-training 3D Point Cloud Transformers with Masked Point Modeling
8
SimMIM: a Simple Framework for Masked Image Modeling
9
iBOT: Image BERT Pre-Training with Online Tokenizer
10
Masked Autoencoders Are Scalable Vision Learners
11
An End-to-End Transformer Model for 3D Object Detection
12
Spatio-temporal Self-Supervised Representation Learning for 3D Point Clouds
13
PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers
14
BEiT: BERT Pre-Training of Image Transformers
15
SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers
16
Dual-stream Network for Visual Recognition
17
Learning Transferable Visual Models From Natural Language Supervision
18
Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision
19
Fast Convergence of DETR with Spatially Modulated Co-Attention
20
Self-Supervised Pretraining of 3D Features on any Point-Cloud
21
Training data-efficient image transformers & distillation through attention
22
PCT: Point cloud transformer
23
Exploring Simple Siamese Representation Learning
24
End-to-End Object Detection with Adaptive Clustering Transformer
26
An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
27
Deformable DETR: Deformable Transformers for End-to-End Object Detection
28
Unsupervised Point Cloud Pre-training via Occlusion Completion
29
Self-Supervised Learning of Point Clouds via Orientation Estimation
30
PointContrast: Unsupervised Pre-training for 3D Point Cloud Understanding
31
Language Models are Few-Shot Learners
32
End-to-End Object Detection with Transformers
33
Self-Supervised Learning for Domain Adaptation on Point Clouds
34
A Simple Framework for Contrastive Learning of Visual Representations
35
Momentum Contrast for Unsupervised Visual Representation Learning
36
DensePoint: Learning Densely Contextual Representation for Efficient Point Cloud Processing
37
Revisiting Point Cloud Classification: A New Benchmark Dataset and Classification Model on Real-World Data
38
Multi-Angle Point Cloud-VAE: Unsupervised Feature Learning for 3D Point Clouds From Multiple Angles by Joint Self-Reconstruction and Half-to-Half Prediction
39
VoteNet: A Deep Learning Label Fusion Method for Multi-Atlas Segmentation
40
Relation-Shape Convolutional Neural Network for Point Cloud Analysis
41
Self-Supervised Deep Learning on Point Clouds by Reconstructing Space
42
View Inter-Prediction GAN: Unsupervised Representation Learning for 3D Shapes by Learning Global Shape Memories to Support Local View Predictions
43
SO-Net: Self-Organizing Network for Point Cloud Analysis
44
Dynamic Graph CNN for Learning on Point Clouds
45
PointCNN: Convolution On X-Transformed Points
46
FoldingNet: Point Cloud Auto-Encoder via Deep Grid Deformation
47
Attention is All you Need
48
PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space
49
ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes
50
A Point Set Generation Network for 3D Object Reconstruction from a Single Image
51
PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation
52
A scalable active framework for region annotation in 3D shape collections
53
Learning a Probabilistic Latent Space of Object Shapes via 3D Generative-Adversarial Modeling
54
Discrete Variational Autoencoders
55
ShapeNet: An Information-Rich 3D Model Repository
56
U-Net: Convolutional Networks for Biomedical Image Segmentation
57
Adam: A Method for Stochastic Optimization
58
3D ShapeNets: A deep representation for volumetric shapes
59
MonoDETR: Depth-aware Transformer for Monocular 3D Object Detection
60
Learning Localized Representations of Point Clouds With Graph-Convolutional Generative Adversarial Networks
61
Swin Transformer: Hierarchical Vision Transformer using Shifted Windows
62
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
63
Language Models are Unsupervised Multitask Learners
64
Improving Language Understanding by Generative Pre-Training
65
Visualizing Data using t-SNE
66
Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)?
67
Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation
68
Did you include the full text of instructions given to participants and screenshots, if applicable?
69
Do the main claims made in the abstract and introduction accurately reflect the paper's contributions and scope?
70
Have you read the ethics review guidelines and ensured that your paper conforms to them
71
Checklist 1. For all authors
72
code, data, models) or curating/releasing new assets... (a) If your work uses existing assets
73
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable?
74
(a) Did you state the full set of assumptions of all theoretical results?
75
c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)?
76
Did you discuss any potential negative societal impacts of your work