ULIP: Learning a Unified Representation of Language, Images, and Point Clouds for 3D Understanding

The recognition capabilities of current state-of-the-art 3D models are limited by datasets with a small number of annotated data and a pre-defined set of categories. In its 2D counterpart, recent advances have shown that similar problems can be significantly alleviated by employing knowledge from other modalities, such as language. Inspired by this, leveraging multimodal information for 3D modality could be promising to improve 3D understanding under the restricted data regime, but this line of research is not well studied. Therefore, we introduce ULIP to learn a unified representation of images, texts, and 3D point clouds by pre-training with object triplets from the three modalities. To overcome the shortage of training triplets, ULIP leverages a pre-trained vision-language model that has already learned a common visual and textual space by training with massive image-text pairs. Then, ULIP learns a 3D representation space aligned with the common image-text space, using a small number of automatically synthesized triplets. ULIP is agnostic to 3D backbone networks and can easily be integrated into any 3D architecture. Experiments show that ULIP effectively improves the performance of multiple recent 3D backbones by simply pre-training them on ShapeNet55 using our framework, achieving state-of-the-art performance in both standard 3D classification and zero-shot 3D classification on ModelNet40 and ScanObjectNN. ULIP also improves the performance of PointMLP by around 3% in 3D classification on ScanObjectNN, and outperforms PointCLIP by 28.8% on top-1 accuracy for zero-shot 3D classification on ModelNet40. Our code and pre-trained models are released at https://github.com/salesforce/ULIP.

PDF Abstract CVPR 2023 PDF CVPR 2023 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Training-free 3D Point Cloud Classification ModelNet40 ULIP Accuracy (%) 60.4 # 3
Need 3D Data? Yes # 1
3D Point Cloud Classification ModelNet40 ULIP + PointMLP Overall Accuracy 94.7 # 7
Mean Accuracy 92.4 # 2
Zero-Shot Transfer 3D Point Cloud Classification ModelNet40 ULIP + PointMLP Accuracy (%) 61.5 # 8
Zero-Shot Transfer 3D Point Cloud Classification ModelNet40 ULIP + PointBERT Accuracy (%) 60.4 # 9
3D Point Cloud Classification ModelNet40 ULIP + PointNet++(ssg) Overall Accuracy 93.4 # 53
Mean Accuracy 91.2 # 15
3D Point Cloud Classification ModelNet40 ULIP + PointBERT Overall Accuracy 94.1 # 19
3D Point Cloud Classification ScanObjectNN ULIP + PointMLP Overall Accuracy 89.4 # 17
Mean Accuracy 88.5 # 6
3D Point Cloud Classification ScanObjectNN ULIP + PointNeXt Overall Accuracy 89.7 # 15
Mean Accuracy 88.6 # 5
Number of params 1.4M # 51
3D Point Cloud Classification ScanObjectNN ULIP + PointBERT Overall Accuracy 86.4 # 36

Methods


No methods listed for this paper. Add relevant methods here