3260 papers • 126 benchmarks • 313 datasets
This task has no description! Would you like to contribute one?
(Image credit: Papersgraph)
These leaderboards are used to track progress in lidar-semantic-segmentation-1
Use these libraries to find lidar-semantic-segmentation-1 models and implementations
No subtasks available.
The approach, named SSD, discretizes the output space of bounding boxes into a set of default boxes over different aspect ratios and scales per feature map location, which makes SSD easy to train and straightforward to integrate into systems that require a detection component.
KPConv is a new design of point convolution, i.e. that operates on point clouds without any intermediate representation, that outperform state-of-the-art classification and segmentation approaches on several datasets.
This paper introduces RandLA-Net, an efficient and lightweight neural architecture to directly infer per-point semantics for large-scale point clouds, and introduces a novel local feature aggregation module to progressively increase the receptive field for each 3D point, thereby effectively preserving geometric details.
This work proposes Sparse Point-Voxel Convolution (SPVConv), a lightweight 3D module that equips the vanilla Sparse Convolution with the high-resolution point-based branch, and presents 3D Neural Architecture Search (3D-NAS) to search the optimal network architecture over this diverse design space efficiently and effectively.
This work proposes a new LiDAR-specific, KNN-free segmentation algorithm - PolarNet, which greatly increases the mIoU in three drastically different real urban LiDar single-scan segmentation datasets while retaining ultra low latency and near real-time throughput.
This paper proposes LENet, a lightweight andcient projection-based LiDAR semantic segmentation network, which has an encoder-decoder architecture, and introduces multiple auxiliary segmentation heads to further improve the network accuracy.
This work develops a 3D cylinder partition and a3D cylinder convolution based framework, termed as Cylinder3D, which exploits the 3D topology relations and structures of driving-scene point clouds and introduces a dimension-decomposition based context modeling module to explore the high-rank context information in point clouds in a progressive manner.
This paper proposes using scribbles to annotate LiDAR point clouds and releases ScribbleKITTI, the first scribble-annotated dataset for LiDar semantic segmentation, and presents a pipeline to reduce the performance gap that arises when using such weak annotations.
CPGNet is proposed, which ensures both effectiveness and efficiency mainly by the following two techniques: the novel Point-Grid (PG) fusion block extracts semantic features mainly on the 2D projected grid for efficiency, while the proposed transformation consistency loss narrows the gap between the single-time model inference and TTA.
This paper presents Point Transformer V3 (PTv3), which prioritizes simplicity and efficiency over the accuracy of certain mechanisms that are minor to the over-all performance after scaling, such as replacing the precise neighbor search by KNN with an efficient serialized neighbor mapping of point clouds organized with specific patterns.
Adding a benchmark result helps the community track progress.