Openscene: 3d scene understanding with open vocabularies

S Peng, K Genova, C Jiang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Traditional 3D scene understanding approaches rely on labeled 3D datasets to train a
model for a single task with supervision. We propose OpenScene, an alternative approach …

Pointclip: Point cloud understanding by clip

R Zhang, Z Guo, W Zhang, K Li… - Proceedings of the …, 2022 - openaccess.thecvf.com
Recently, zero-shot and few-shot learning via Contrastive Vision-Language Pre-training
(CLIP) have shown inspirational performance on 2D visual recognition, which learns to …

Clip2scene: Towards label-efficient 3d scene understanding by clip

R Chen, Y Liu, L Kong, X Zhu, Y Ma… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Contrastive Language-Image Pre-training (CLIP) achieves promising results in 2D
zero-shot and few-shot learning. Despite the impressive performance in 2D, applying CLIP …

Clip2point: Transfer clip to point cloud classification with image-depth pre-training

T Huang, B Dong, Y Yang, X Huang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Pre-training across 3D vision and language remains under development because of limited
training data. Recent works attempt to transfer vision-language (VL) pre-training methods to …

CLIP2: Contrastive language-image-point pretraining from real-world point cloud data

Y Zeng, C Jiang, J Mao, J Han, C Ye… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Contrastive Language-Image Pre-training, benefiting from large-scale unlabeled
text-image pairs, has demonstrated great performance in open-world vision understanding …

Language-grounded indoor 3d semantic segmentation in the wild

D Rozenberszki, O Litany, A Dai - European Conference on Computer …, 2022 - Springer
Recent advances in 3D semantic segmentation with deep neural networks have shown
remarkable success, with rapid performance increase on available datasets. However …

Pointclip v2: Prompting clip and gpt for powerful 3d open-world learning

X Zhu, R Zhang, B He, Z Guo, Z Zeng… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale pre-trained models have shown promising open-world performance for both
vision and language tasks. However, their transferred capacity on 3D point clouds is still …

Clip-fo3d: Learning free open-world 3d scene representations from 2d dense clip

J Zhang, R Dong, K Ma - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Training a 3D scene understanding model requires complicated human annotations, which
are laborious to collect and result in a model only encoding close-set object semantics. In …

See more and know more: Zero-shot point cloud segmentation via multi-modal visual data

Y Lu, Q Jiang, R Chen, Y Hou… - Proceedings of the …, 2023 - openaccess.thecvf.com
Zero-shot point cloud segmentation aims to make deep models capable of recognizing
novel objects in point cloud that are unseen in the training phase. Recent trends favor the …

Synthesized feature based few-shot class-incremental learning on a mixture of subspaces

A Cheraghian, S Rahman… - Proceedings of the …, 2021 - openaccess.thecvf.com
Few-shot class incremental learning (FSCIL) aims to incrementally add sets of novel classes
to a well-trained base model in multiple training sessions with the restriction that only a few …