Lidarmultinet: Towards a unified multi-task network for lidar perception

D Ye, Z Zhou, W Chen, Y Xie, Y Wang… - Proceedings of the …, 2023 - ojs.aaai.org
LiDAR-based 3D object detection, semantic segmentation, and panoptic segmentation are
usually implemented in specialized networks with distinctive architectures that are difficult to …

A review of 3D object detection for autonomous driving of electric vehicles

D Dai, Z Chen, P Bao, J Wang - World Electric Vehicle Journal, 2021 - mdpi.com
In recent years, electric vehicles have achieved rapid development. Intelligence is one of the
important trends to promote the development of electric vehicles. As a result, autonomous …

Image2point: 3d point-cloud understanding with 2d image pretrained models

C Xu, S Yang, T Galanti, B Wu, X Yue, B Zhai… - … on Computer Vision, 2022 - Springer
Abstract 3D point-clouds and 2D images are different visual representations of the physical
world. While human vision can understand both representations, computer vision models …

Detmatch: Two teachers are better than one for joint 2d and 3d semi-supervised object detection

J Park, C Xu, Y Zhou, M Tomizuka, W Zhan - European Conference on …, 2022 - Springer
While numerous 3D detection works leverage the complementary relationship between RGB
images and point clouds, developments in the broader framework of semi-supervised object …

3D recognition based on sensor modalities for robotic systems: A survey

S Manzoor, SH Joo, EJ Kim, SH Bae, GG In, JW Pyo… - Sensors, 2021 - mdpi.com
3D visual recognition is a prerequisite for most autonomous robotic systems operating in the
real world. It empowers robots to perform a variety of tasks, such as tracking, understanding …

FULLER: Unified Multi-modality Multi-task 3D Perception via Multi-level Gradient Calibration

Z Huang, S Lin, G Liu, M Luo, C Ye… - Proceedings of the …, 2023 - openaccess.thecvf.com
Multi-modality fusion and multi-task learning are becoming trendy in 3D autonomous driving
scenario, considering robust prediction and computation budget. However, naively …

SA-BEV: Generating Semantic-Aware Bird's-Eye-View Feature for Multi-view 3D Object Detection

J Zhang, Y Zhang, Q Liu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Recently, the pure camera-based Bird's-Eye-View (BEV) perception provides a feasible
solution for economical autonomous driving. However, the existing BEV-based multi-view …

You only group once: Efficient point-cloud processing with token representation and relation inference module

C Xu, B Zhai, B Wu, T Li, W Zhan… - 2021 IEEE/RSJ …, 2021 - ieeexplore.ieee.org
3D perception on point-cloud is a challenging and crucial computer vision task. A point-
cloud consists of a sparse, unstructured, and unordered set of points. To understand a point …

Lidarmultinet: Unifying lidar semantic segmentation, 3d object detection, and panoptic segmentation in a single multi-task network

D Ye, W Chen, Z Zhou, Y Xie, Y Wang, P Wang… - arXiv preprint arXiv …, 2022 - arxiv.org
This technical report presents the 1st place winning solution for the Waymo Open Dataset
3D semantic segmentation challenge 2022. Our network, termed LidarMultiNet, unifies the …

HFT: Lifting Perspective Representations via Hybrid Feature Transformation for BEV Perception

J Zou, Z Zhu, J Huang, T Yang… - … on Robotics and …, 2023 - ieeexplore.ieee.org
Restoring an accurate Bird's Eye View (BEV) map plays a crucial role in the perception of
autonomous driving. The existing works of lifting representations from frontal view to BEV …