Skip to content

Latest commit

 

History

History
executable file
·
151 lines (113 loc) · 5.23 KB

README.md

File metadata and controls

executable file
·
151 lines (113 loc) · 5.23 KB
  • To change data encoding method: (fixed, linear, quadratic, exponential)

    dataloader/dataset_semantickitti.py, line 258 - 288, comment and uncomment the corresponding part

  • To change feature shape: ([480, 360, 32], [360, 360, 32],.....)

    config/semantickitti.yaml, line 10 - 12

  • To change network architecture: (AsymCNN, vanillaCNN)

    network/segmentator_3d_asymm_spconv.py

    line 46 - 109 class ResContextBlock(nn.Module):

    comment and uncomment the corresponding part in both init and forward function

  • To change loss function (unweighted, weighted)

    builder/loss_builder.py

    line 11 - 21

    comment and uncomment the corresponding part

Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR Segmentation

The source code of our work "Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR Segmentation img|center

News

  • 2021-03 [NEW:fire:] Cylinder3D is accepted to CVPR 2021 as an Oral presentation
  • 2021-01 [NEW:fire:] Cylinder3D achieves the 1st place in the leaderboard of SemanticKITTI multiscan semantic segmentation

  • 2020-12 [NEW:fire:] Cylinder3D achieves the 2nd place in the challenge of nuScenes LiDAR segmentation, with mIoU=0.779, fwIoU=0.899 and FPS=10Hz.
  • 2020-12 We release the new version of Cylinder3D with nuScenes dataset support.
  • 2020-11 We preliminarily release the Cylinder3D--v0.1, supporting the LiDAR semantic segmentation on SemanticKITTI and nuScenes.
  • 2020-11 Our work achieves the 1st place in the leaderboard of SemanticKITTI semantic segmentation (until CVPR2021 DDL, still rank 1st in term of Accuracy now), and based on the proposed method, we also achieve the 1st place in the leaderboard of SemanticKITTI panoptic segmentation.

Installation

Requirements

Data Preparation

SemanticKITTI

./
├── 
├── ...
└── path_to_data_shown_in_config/
    ├──sequences
        ├── 00/           
        │   ├── velodyne/	
        |   |	├── 000000.bin
        |   |	├── 000001.bin
        |   |	└── ...
        │   └── labels/ 
        |       ├── 000000.label
        |       ├── 000001.label
        |       └── ...
        ├── 08/ # for validation
        ├── 11/ # 11-21 for testing
        └── 21/
	    └── ...

nuScenes

./
├── 
├── ...
└── path_to_data_shown_in_config/
		├──v1.0-trainval
		├──v1.0-test
		├──samples
		├──sweeps
		├──maps

Training

  1. modify the config/semantickitti.yaml with your custom settings. We provide a sample yaml for SemanticKITTI
  2. train the network by running "sh train.sh"

Training for nuScenes

Please refer to NUSCENES-GUIDE

Pretrained Models

-- We provide a pretrained model for SemanticKITTI LINK1 or LINK2 (access code: xqmi)

-- For nuScenes dataset, please refer to NUSCENES-GUIDE

Semantic segmentation demo for a folder of lidar scans

python demo_folder.py --demo-folder YOUR_FOLDER --save-folder YOUR_SAVE_FOLDER

If you want to validate with your own datasets, you need to provide labels. --demo-label-folder is optional

python demo_folder.py --demo-folder YOUR_FOLDER --save-folder YOUR_SAVE_FOLDER --demo-label-folder YOUR_LABEL_FOLDER

TODO List

  • Release pretrained model for nuScenes.
  • Support multiscan semantic segmentation.
  • Support more models, including PolarNet, RandLA, SequeezeV3 and etc.
  • Integrate LiDAR Panotic Segmentation into the codebase.

Reference

If you find our work useful in your research, please consider citing our paper:

@article{zhu2020cylindrical,
  title={Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR Segmentation},
  author={Zhu, Xinge and Zhou, Hui and Wang, Tai and Hong, Fangzhou and Ma, Yuexin and Li, Wei and Li, Hongsheng and Lin, Dahua},
  journal={arXiv preprint arXiv:2011.10033},
  year={2020}
}

#for LiDAR panoptic segmentation
@article{hong2020lidar,
  title={LiDAR-based Panoptic Segmentation via Dynamic Shifting Network},
  author={Hong, Fangzhou and Zhou, Hui and Zhu, Xinge and Li, Hongsheng and Liu, Ziwei},
  journal={arXiv preprint arXiv:2011.11964},
  year={2020}
}

Acknowledgments

We thanks for the opensource codebases, PolarSeg and spconv

Hiring from Zhouhui

The SenseTime-LidarSegmentation is now hiring. If you are interested in internship, researcher and software engineer positions related to lidar segmentation or deep learning, feel free to send email: [email protected].