[TMM 2023] Self-Supervised Intra-Modal and Cross-Modal Contrastive Learning for Point Cloud Understanding
This paper proposes a self-supervised point cloud understanding method called CrossNet. CrossNet is simple and efficient, developing the intra-modal contrastive loss between the point clouds and the cross-modal contrastive loss between the point clouds and images. Finally, we combine the overall training objectives.
If you entrust our work with value, please consider giving a star ⭐ and citation.
@article{wu2023self,
title={Self-Supervised Intra-Modal and Cross-Modal Contrastive Learning for Point Cloud Understanding},
author={Wu, Yue and Liu, Jiaming and Gong, Maoguo and Gong, Peiran and Fan, Xiaolong and Qin, AK and Miao, Qiguang and Ma, Wenping},
journal={IEEE Transactions on Multimedia},
year={2023},
publisher={IEEE}
}
Refer requirements.txt
for the required packages.
Datasets are available here. Run the command below to download all the datasets (ShapeNetRender, ModelNet40, ScanObjectNN, ShapeNetPart) to reproduce the results. Additional S3DIS is optional.
cd data
source download_data.sh
Refer python train_crossnet_con.py
for the command to train CrossNet.
Run downstream/classification/main.py
to perform linear SVM object classification in both ModelNet40 and ScanObjectNN datasets.
Refer downstream/segmentation/main_partseg.py
for fine-tuning experiment for part segmentation in ShapeNetPart dataset.
Refer downstream/segmentation/main_semseg.py
for fine-tuning experiment for semantic segmentation in S3DIS dataset.
Our code borrows heavily from CrossPoint repository. We thank the authors of CrossPoint for releasing their code.