README.md 5.47 KB
Newer Older
Shaoshuai Shi's avatar
Shaoshuai Shi committed
1
# OpenLidarPerceptron
Shaoshuai Shi's avatar
Shaoshuai Shi committed
2
3

## Introduction
Shaoshuai Shi's avatar
Shaoshuai Shi committed
4
5
6
`OpenLidarPerceptron` is an open source project for LiDAR-based 3D scene perception. 
As of now, it mainly consists of `PCDet` toolbox for 3D object detection from point cloud, 
and we are going to support LiDAR segmentation and odometry in the next phase.
Shaoshuai Shi's avatar
Shaoshuai Shi committed
7
8
9
10


### What does `PCDet` toolbox do?

Gus-Guo's avatar
Gus-Guo committed
11
Note that we have upgrated `PCDet` from `v0.1` to `v0.2` with pretty new structures to support various datasets and models.
12

Shaoshuai Shi's avatar
Shaoshuai Shi committed
13
`PCDet` is a general PyTorch-based codebase for 3D object detection from point cloud. 
Shaoshuai Shi's avatar
Shaoshuai Shi committed
14
It currently supports multiple state-of-the-art 3D object detection methods with highly refactored codes for both one-stage and two-stage 3D detection frameworks.
Shaoshuai Shi's avatar
Shaoshuai Shi committed
15

Shaoshuai Shi's avatar
Shaoshuai Shi committed
16
17
18
19
Based on `PCDet` toolbox, we win the Waymo Open Dataset challenge in [3D Detection](https://waymo.com/open/challenges/3d-detection/), 
[3D Tracking](https://waymo.com/open/challenges/3d-tracking/), [Domain Adaptation](https://waymo.com/open/challenges/domain-adaptation/) 
three tracks among all LiDAR-only methods, and the Waymo related models will be released to `PCDet` soon.    

Shaoshuai Shi's avatar
Shaoshuai Shi committed
20
21
It is also the official code release of [`[Part-A^2 net]`](https://arxiv.org/abs/1907.03670) and [`[PV-RCNN]`](https://arxiv.org/abs/1912.13192). 

Shaoshuai Shi's avatar
Shaoshuai Shi committed
22
23
24
We are actively updating this repo currently, and more datasets and models will be supported soon. 
Contributions are also welcomed. 

Shaoshuai Shi's avatar
Shaoshuai Shi committed
25
26

### Currently Supported Features
Shaoshuai Shi's avatar
Shaoshuai Shi committed
27
- [x] Unified point cloud coordinate and clear code structure and for supporting lots of datasets and approaches
Shaoshuai Shi's avatar
Shaoshuai Shi committed
28
- [x] Support both one-stage and two-stage 3D object detection frameworks
Shaoshuai Shi's avatar
Shaoshuai Shi committed
29
30
31
32
33
34
35
36
37
38
39
40
41
- [x] Support distributed training & testing with multiple GPUs and multiple machines
- [x] Support multiple heads on different scales to detect different classes
- [x] Support stacked version set abstraction to encode various number of points in different scenes
- [x] Support Adaptive Training Sample Selection (ATSS) for target assignment
- [x] Support RoI-aware point cloud pooling & RoI-grid point cloud pooling
- [x] Support GPU version 3D IoU calculation and rotated NMS 

## Model Zoo

### KITTI 3D Object Detection Baselines
Selected supported methods are shown in the below table. The results are the 3D detection performance of car class on the *val* set of KITTI dataset.
All models are trained with 8 GPUs and are available for download.

42
|                                             | Batch Size | AP_Easy | **AP_Moderate** | AP_Hard | download  |
Shaoshuai Shi's avatar
Shaoshuai Shi committed
43
|---------------------------------------------|:----------:|:-------:|:-------:|:-------:|:---------:|
44
45
46
47
| [PointPillar](tools/cfgs/kitti_models/pointpillar.yaml) | 32 | 86.46 | 77.28 | 74.65 | [model-18M](https://drive.google.com/file/d/1wMxWTpU1qUoY3DsCH31WJmvJxcjFXKlm/view?usp=sharing) | 
| [SECOND](tools/cfgs/kitti_models/second.yaml)           | 32  | 88.61 | 78.62| 77.22 | [model-20M](https://drive.google.com/file/d/1-01zsPOsqanZQqIIyy7FpNXStL3y4jdR/view?usp=sharing) |
| [Part-A^2](tools/cfgs/kitti_models/PartA2.yaml)    | 32 | 89.55 | 79.40 | 78.84 | [model-244M](https://drive.google.com/file/d/10GK1aCkLqxGNeX3lVu8cLZyE0G8002hY/view?usp=sharing) |
| [PV-RCNN](tools/cfgs/kitti_models/pv_rcnn.yaml) | 16 | 89.34 | 83.69 | 78.70 | [model-50M](https://drive.google.com/file/d/1lIOq4Hxr0W3qsX83ilQv0nk1Cls6KAr-/view?usp=sharing) |
Shaoshuai Shi's avatar
Shaoshuai Shi committed
48
49
| [SECOND-MultiHead](tools/cfgs/kitti_models/second_multihead.yaml) | 32 | - | - | - | ongoing |
| PointRCNN | 32 | - | - | - | ongoing|
Shaoshuai Shi's avatar
Shaoshuai Shi committed
50
51
52
53
54
55
56
57
58
59
60
61

## Installation

Please refer to [INSTALL.md](docs/INSTALL.md) for installation and dataset preparation.


## Get Started

Please refer to [GETTING_STARTED.md](docs/GETTING_STARTED.md) to learn more usage about this project.

## License

Shaoshuai Shi's avatar
Shaoshuai Shi committed
62
`OpenLidarPerceptron` is released under the [Apache 2.0 license](LICENSE).
Shaoshuai Shi's avatar
Shaoshuai Shi committed
63
64

## Acknowledgement
Shaoshuai Shi's avatar
Shaoshuai Shi committed
65
`OpenLidarPerceptron` is an open source project for LiDAR-based 3D scene perception that supports multiple
Shaoshuai Shi's avatar
Shaoshuai Shi committed
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
LiDAR-based perception models as shown above. Some parts of `PCDet` are learned from the official released codes of the above supported methods. 
We would like to thank for their proposed methods and the official implementation.   

We hope that this repo could serve as a strong and flexible codebase to benefit the research community by speeding up the process of reimplementing previous works and/or developing new methods.


## Citation 
If you find this project useful in your research, please consider cite:


```
@inproceedings{shi2020pv,
  title={Pv-rcnn: Point-voxel feature set abstraction for 3d object detection},
  author={Shi, Shaoshuai and Guo, Chaoxu and Jiang, Li and Wang, Zhe and Shi, Jianping and Wang, Xiaogang and Li, Hongsheng},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={10529--10538},
  year={2020}
}


@article{shi2020points,
  title={From Points to Parts: 3D Object Detection from Point Cloud with Part-aware and Part-aggregation Network},
  author={Shi, Shaoshuai and Wang, Zhe and Shi, Jianping and Wang, Xiaogang and Li, Hongsheng},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence},
  year={2020},
  publisher={IEEE}
}

@inproceedings{shi2019pointrcnn,
  title={PointRCNN: 3d Object Progposal Generation and Detection from Point Cloud},
  author={Shi, Shaoshuai and Wang, Xiaogang and Li, Hongsheng},
  booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
  pages={770--779},
  year={2019}
}
```
Shaoshuai Shi's avatar
Shaoshuai Shi committed
102

Shaoshuai Shi's avatar
Shaoshuai Shi committed
103
104
## Contact
This project is currently maintained by Shaoshuai Shi ([@sshaoshuai](http://github.com/sshaoshuai)) and Chaoxu Guo ([@Gus-Guo](https://github.com/Gus-Guo)).