作者
Sourabh Vora, Alex H Lang, Bassam Helou, Oscar Beijbom
发表日期
2020
研讨会论文
Proceedings of the IEEE/CVF conference on computer vision and pattern recognition
页码范围
4604-4612
简介
Camera and lidar are important sensor modalities for robotics in general and self-driving cars in particular. The sensors provide complementary information offering an opportunity for tight sensor-fusion. Surprisingly, lidar-only methods outperform fusion methods on the main benchmark datasets, suggesting a gap in the literature. In this work, we propose PointPainting: a sequential fusion method to fill this gap. PointPainting works by projecting lidar points into the output of an image-only semantic segmentation network and appending the class scores to each point. The appended (painted) point cloud can then be fed to any lidar-only method. Experiments show large improvements on three different state-of-the art methods, Point-RCNN, VoxelNet and PointPillars on the KITTI and nuScenes datasets. The painted version of PointRCNN represents a new state of the art on the KITTI leaderboard for the bird's-eye view detection task. In ablation, we study how the effects of Painting depends on the quality and format of the semantic segmentation output, and demonstrate how latency can be minimized through pipelining.
引用总数
学术搜索中的文章
S Vora, AH Lang, B Helou, O Beijbom - Proceedings of the IEEE/CVF conference on computer …, 2020