Rethinking range view representation for lidar segmentation

L Kong, Y Liu, R Chen, Y Ma, X Zhu… - Proceedings of the …, 2023 - openaccess.thecvf.com
LiDAR segmentation is crucial for autonomous driving perception. Recent trends favor point-
or voxel-based methods as they often yield better performance than the traditional range …

Openshape: Scaling up 3d shape representation towards open-world understanding

M Liu, R Shi, K Kuang, Y Zhu, X Li… - Advances in neural …, 2024 - proceedings.neurips.cc
We introduce OpenShape, a method for learning multi-modal joint representations of text,
image, and point clouds. We adopt the commonly used multi-modal contrastive learning …

Towards open vocabulary learning: A survey

J Wu, X Li, S Xu, H Yuan, H Ding… - … on Pattern Analysis …, 2024 - ieeexplore.ieee.org
In the field of visual scene understanding, deep neural networks have made impressive
advancements in various core tasks like segmentation, tracking, and detection. However …

Unsupervised 3d perception with 2d vision-language distillation for autonomous driving

M Najibi, J Ji, Y Zhou, CR Qi, X Yan… - Proceedings of the …, 2023 - openaccess.thecvf.com
Closed-set 3D perception models trained on only a pre-defined set of object categories can
be inadequate for safety critical applications such as autonomous driving where new object …

Multi-modal data-efficient 3d scene understanding for autonomous driving

L Kong, X Xu, J Ren, W Zhang, L Pan… - … on Pattern Analysis …, 2025 - ieeexplore.ieee.org
Efficient data utilization is crucial for advancing 3D scene understanding in autonomous
driving, where reliance on heavily human-annotated LiDAR point clouds challenges fully …

Clip-fo3d: Learning free open-world 3d scene representations from 2d dense clip

J Zhang, R Dong, K Ma - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Training a 3D scene understanding model requires complicated human annotations, which
are laborious to collect and result in a model only encoding close-set object semantics. In …

See more and know more: Zero-shot point cloud segmentation via multi-modal visual data

Y Lu, Q Jiang, R Chen, Y Hou… - Proceedings of the …, 2023 - openaccess.thecvf.com
Zero-shot point cloud segmentation aims to make deep models capable of recognizing
novel objects in point cloud that are unseen in the training phase. Recent trends favor the …

Weakly supervised 3d open-vocabulary segmentation

K Liu, F Zhan, J Zhang, M Xu, Y Yu… - Advances in …, 2023 - proceedings.neurips.cc
Open-vocabulary segmentation of 3D scenes is a fundamental function of human perception
and thus a crucial objective in computer vision research. However, this task is heavily …

A survey on deep learning based segmentation, detection and classification for 3D point clouds

PK Vinodkumar, D Karabulut, E Avots, C Ozcinar… - Entropy, 2023 - mdpi.com
The computer vision, graphics, and machine learning research groups have given a
significant amount of focus to 3D object recognition (segmentation, detection, and …

Openins3d: Snap and lookup for 3d open-vocabulary instance segmentation

Z Huang, X Wu, X Chen, H Zhao, L Zhu… - European Conference on …, 2024 - Springer
In this work, we introduce OpenIns3D, a new 3D-input-only framework for 3D open-
vocabulary scene understanding. The OpenIns3D framework employs a “Mask-Snap …