1
Planning-oriented Autonomous Driving
2
BEVPoolv2: A Cutting-edge Implementation of BEVDet Toward Deployment
3
PlanT: Explainable Planning Transformers via Object-Level Representations
4
Model-Based Imitation Learning for Urban Driving
5
Motion Transformer with Global Intention Localization and Local Movement Refinement
6
BEVStereo: Enhancing Depth Estimation in Multi-view 3D Object Detection with Dynamic Temporal Stereo
7
Delving Into the Devils of Bird’s-Eye-View Perception: A Review, Evaluation and Recipe
8
Safety-Enhanced Autonomous Driving Using Interpretable Sensor Fusion Transformer
9
ST-P3: End-to-end Vision-based Autonomous Driving via Spatial-Temporal Feature Learning
10
MMFN: Multi-Modal-Fusion-Net for End-to-End Driving
11
IBISCape: A Simulated Benchmark for multi-modal SLAM Systems Evaluation in Large-scale Dynamic Environments
12
BEVDepth: Acquisition of Reliable Depth for Multi-view 3D Object Detection
13
Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline
14
TransFuser: Imitation With Transformer-Based Sensor Fusion for Autonomous Driving
15
BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation
16
Cross-view Transformers for real-time Map-view Semantic Segmentation
17
M2BEV: Multi-Camera Joint 3D Detection and Segmentation with Unified Birds-Eye View Representation
18
BEVFormer: Learning Bird's-Eye-View Representation from Multi-Camera Images via Spatiotemporal Transformers
19
Learning from All Vehicles
20
PersFormer: 3D Lane Detection via Perspective Transformer and the OpenLane Benchmark
21
PETR: Position Embedding Transformation for Multi-View 3D Object Detection
22
DINO: DETR with Improved DeNoising Anchor Boxes for End-to-End Object Detection
23
A ConvNet for the 2020s
24
BEVDet: High-performance Multi-camera 3D Object Detection in Bird-Eye-View
25
AFDetV2: Rethinking the Necessity of the Second Stage for Object Detection from Point Clouds
26
GRI: General Reinforced Imitation and its Application to Vision-Based Autonomous Driving
27
Structured Bird’s-Eye-View Traffic Scene Understanding from Onboard Images
28
NEAT: Neural Attention Fields for End-to-End Autonomous Driving
29
End-to-End Urban Driving by Imitating a Reinforcement Learning Coach
30
RSN: Range Sparse Net for Efficient, Accurate LiDAR 3D Object Detection
31
Learning to drive from a world on rails
32
Multi-Modal Fusion Transformer for End-to-End Autonomous Driving
33
LiDAR R-CNN: An Efficient and Universal 3D Object Detector
34
Categorical Depth Distribution Network for Monocular 3D Object Detection
35
MP3: A Unified Model to Map, Perceive, Predict and Plan
36
LookOut: Diverse Multi-Future Prediction and Planning for Self-Driving
37
Voxel R-CNN: Towards High Performance Voxel-based 3D Object Detection
38
From Goals, Waypoints & Paths To Long Term Human Trajectory Forecasting
39
IDE-Net: Interactive Driving Event and Pattern Extraction From Human Data
40
Fighting Copycat Agents in Behavioral Cloning from Observation Histories
41
Deformable DETR: Deformable Transformers for End-to-End Object Detection
42
Perceive, Predict, and Plan: Safe Motion Planning Through Interpretable Semantic Representations
43
DSDNet: Deep Structured self-Driving Network
44
Lift, Splat, Shoot: Encoding Images From Arbitrary Camera Rigs by Implicitly Unprojecting to 3D
45
Center-based 3D Object Detection and Tracking
46
MANTRA: Memory Augmented Networks for Multiple Trajectory Prediction
47
End-to-End Object Detection with Transformers
48
TITAN: Future Forecast Using Action Priors
49
RAFT: Recurrent All-Pairs Field Transforms for Optical Flow
50
PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection
52
PyTorch: An Imperative Style, High-Performance Deep Learning Library
53
End-to-End Model-Free Reinforcement Learning for Urban Driving Using Implicit Affordances
54
STD: Sparse-to-Dense 3D Object Detector for Point Cloud
55
End-To-End Interpretable Neural Motion Planner
56
BASNet: Boundary-Aware Salient Object Detection
57
Exploring the Limitations of Behavior Cloning for Autonomous Driving
58
PointRCNN: 3D Object Proposal Generation and Detection From Point Cloud
59
ChauffeurNet: Learning to Drive by Imitating the Best and Synthesizing the Worst
60
3D-LaneNet: End-to-End 3D Multiple Lane Detection
61
Orthographic Feature Transform for Monocular 3D Object Detection
62
SECOND: Sparsely Embedded Convolutional Detection
63
R³Net: Recurrent Residual Refinement Network for Saliency Detection
64
Conditional Affordance Learning for Driving in Urban Environments
65
Path Aggregation Network for Instance Segmentation
66
VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection
67
Decoupled Weight Decay Regularization
68
CARLA: An Open Urban Driving Simulator
69
End-to-End Driving Via Conditional Imitation Learning
70
A Stagewise Refinement Model for Detecting Salient Objects in Images
71
PWC-Net: CNNs for Optical Flow Using Pyramid, Warping, and Cost Volume
72
FlowNet 2.0: Evolution of Optical Flow Estimation with Deep Networks
73
DHSNet: Deep Hierarchical Saliency Network for Salient Object Detection
74
End to End Learning for Self-Driving Cars
75
Deep Residual Learning for Image Recognition
76
Scheduled Sampling for Sequence Prediction with Recurrent Neural Networks
77
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
78
U-Net: Convolutional Networks for Biomedical Image Segmentation
80
FlowNet: Learning Optical Flow with Convolutional Networks
81
On the Properties of Neural Machine Translation: Encoder–Decoder Approaches
82
Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation
83
Off-Road Obstacle Avoidance through End-to-End Learning
85
Policy Pre-training for End-to-end Autonomous Driving via Self-supervised Geometric Modeling
86
Towards Capturing the Temporal Dynamics for Trajectory Prediction: a Coarse-to-Fine Approach
87
Multi-Agent Trajectory Prediction by Combining Egocentric and Allocentric Views
88
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
89
MMCV: OpenMMLab computer vision foundation
91
ALVINN, an autonomous land vehicle in a neural network
92
The refined trajectory notices the jay-walker and leads to a emergency stop during the lane-changing process
93
The refined trajectory leaves more advance for the merging, which leads to a safer and smoother driving
94
For the jay-walker with the nearby vehicle's occlusion, the refined trajectory leads to a deacceleration compared to the original one
95
Figure 1. Visualization for the predictions from different layers of decoder. Larger and brighter dots are from deeper layers
96
license agreement with IEEE. Restrictions apply
98
Authorized licensed use limited to the terms of the applicable license agreement with IEEE
99
On the properties of neural machine [27 21991 Authorized licensed use limited to the terms of the applicable