Skip to content

SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud, CVPR 2021.

Notifications You must be signed in to change notification settings

laiyingxin2/SE-SSD

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud (CVPR 2021) [Paper]

An accurate and fast single-stage 3D object detection framework on KITTI dataset.

Authors: Wu Zheng, Weiliang Tang, Li Jiang, Chi-Wing Fu.

AP on KITTI Dataset

Val Split (11 recall points):

AP_11: car AP(Average Precision)@0.70, 0.70, 0.70:
bbox AP:98.72, 90.10, 89.57
bev  AP:90.61, 88.76, 88.18
3d   AP:90.21, 86.25, 79.22
aos  AP:98.67, 89.86, 89.16


AP_40: car AP(Average Precision)@0.70, 0.70, 0.70:
bbox AP:99.57, 95.58, 93.16
bev  AP:96.70, 92.15, 89.74
3d   AP:93.75, 86.18, 83.50
aos  AP:99.52, 95.28, 92.69

Test Split: Submission link

You may download the trained student model here, which is trained on the train split (3712 samples).

Pipeline

pipeline The framework of our Self-Ensembling Single-Stage object Detector (SE-SSD) with a teacher SSD and a student SSD. To start, we feed the input point cloud to the teacher to produce relatively precise bounding boxes and confidence, and take these predictions (after global transformations) as soft targets to supervise the student with our consistency loss. On the top branch, we apply the same global transformations to the input, then perform our shape-aware data augmentation to generate augmented samples as inputs to the student. Further, we formulate the Orientation-aware Distance-IoU loss to supervise the student with hard targets, and update the teacher parameters based on the student parameters with the exponential moving average (EMA) strategy. In this way, the framework can boost the precisions of the detector significantly without incurring extra computation during the inference.

Installation

$ git clone https://github.com/Vegeta2020/SE-SSD.git
$ cd ./SE-SSD/det3d/core/iou3d
$ python setup.py install
$ cd ./SE-SSD
$ python setup.py build develop

$ git clone https://github.com/jackd/ifp-sample.git
$ pip install -e ifp-sample

Please follow Det3D for installation of other related packages and data preparation.

Train and Eval

Configure the model in

$ /SE-SSD/examples/second/configs/config.py

Please use our code to generate ground truth data:

$ python ./SE-SSD/tools/create_data.py

Train the SE-SSD:

$ cd ./SE-SSD/tools
$ python train.py  # Single GPU
$ python -m torch.distributed.launch --nproc_per_node=4 train.py   # Multiple GPU

Evaluate the SE-SSD:

$ cd ./SE-SSD/tools
$ python test.py

Citation

If you find this work useful in your research, please star our repository and consider citing:

@inproceedings{zheng2021se,
  title={SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud},
  author={Wu Zheng, Weiliang Tang, Li Jiang, Chi-Wing Fu},
  booktitle={CVPR},
  pages={14494--14503},
  year={2021}
}

Acknowledgement

Thanks for previous works Det3D and CIA-SSD, as our codebase are mainly based on them.

Special thanks for Dingfu Zhou for his shared code.

Thanks for the reviewers's valuable comments on this paper.

Contact

If you have any question or suggestion about this repo, please feel free to contact me ([email protected])

About

SE-SSD: Self-Ensembling Single-Stage Object Detector From Point Cloud, CVPR 2021.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 84.5%
  • Cuda 7.9%
  • C++ 7.2%
  • Other 0.4%