Attention mechanisms in computer vision: A survey

MH Guo, TX Xu, JJ Liu, ZN Liu, PT Jiang, TJ Mu… - Computational visual …, 2022 - Springer
Humans can naturally and effectively find salient regions in complex scenes. Motivated by
this observation, attention mechanisms were introduced into computer vision with the aim of …

Deep learning for 3d point clouds: A survey

Y Guo, H Wang, Q Hu, H Liu, L Liu… - IEEE transactions on …, 2020 - ieeexplore.ieee.org
Point cloud learning has lately attracted increasing attention due to its wide applications in
many areas, such as computer vision, autonomous driving, and robotics. As a dominating …

Visual attention network

MH Guo, CZ Lu, ZN Liu, MM Cheng, SM Hu - Computational Visual Media, 2023 - Springer
While originally designed for natural language processing tasks, the self-attention
mechanism has recently taken various computer vision areas by storm. However, the 2D …

Point transformer

H Zhao, L Jiang, J Jia, PHS Torr… - Proceedings of the …, 2021 - openaccess.thecvf.com
Self-attention networks have revolutionized natural language processing and are making
impressive strides in image analysis tasks such as image classification and object detection …

Pct: Point cloud transformer

MH Guo, JX Cai, ZN Liu, TJ Mu, RR Martin… - Computational Visual …, 2021 - Springer
The irregular domain and lack of ordering make it challenging to design deep neural
networks for point cloud processing. This paper presents a novel framework named Point …

Co-scale conv-attentional image transformers

W Xu, Y Xu, T Chang, Z Tu - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
In this paper, we present Co-scale conv-attentional image Transformers (CoaT), a
Transformer-based image classifier equipped with co-scale and conv-attentional …

Pointcontrast: Unsupervised pre-training for 3d point cloud understanding

S Xie, J Gu, D Guo, CR Qi, L Guibas… - Computer Vision–ECCV …, 2020 - Springer
Arguably one of the top success stories of deep learning is transfer learning. The finding that
pre-training a network on a rich source set (eg, ImageNet) can help boost performance once …

Se (3)-transformers: 3d roto-translation equivariant attention networks

F Fuchs, D Worrall, V Fischer… - Advances in neural …, 2020 - proceedings.neurips.cc
Abstract We introduce the SE (3)-Transformer, a variant of the self-attention module for 3D
point-clouds, which is equivariant under continuous 3D roto-translations. Equivariance is …

Randla-net: Efficient semantic segmentation of large-scale point clouds

Q Hu, B Yang, L Xie, S Rosa, Y Guo… - Proceedings of the …, 2020 - openaccess.thecvf.com
We study the problem of efficient semantic segmentation for large-scale 3D point clouds. By
relying on expensive sampling techniques or computationally heavy pre/post-processing …

Pointasnl: Robust point clouds processing using nonlocal neural networks with adaptive sampling

X Yan, C Zheng, Z Li, S Wang… - Proceedings of the IEEE …, 2020 - openaccess.thecvf.com
Raw point clouds data inevitably contains outliers or noise through acquisition from 3D
sensors or reconstruction algorithms. In this paper, we present a novel end-to-end network …