README.md 10.5 KB
Newer Older
zhangwenwei's avatar
zhangwenwei committed
1
<div align="center">
zhangwenwei's avatar
zhangwenwei committed
2
  <img src="resources/mmdet3d-logo.png" width="600"/>
zhangwenwei's avatar
zhangwenwei committed
3
</div>
zhangwenwei's avatar
zhangwenwei committed
4

Wenwei Zhang's avatar
Wenwei Zhang committed
5
6
7
8
9
10
[![docs](https://img.shields.io/badge/docs-latest-blue)](https://mmdetection3d.readthedocs.io/en/latest/)
[![badge](https://github.com/open-mmlab/mmdetection3d/workflows/build/badge.svg)](https://github.com/open-mmlab/mmdetection3d/actions)
[![codecov](https://codecov.io/gh/open-mmlab/mmdetection3d/branch/master/graph/badge.svg)](https://codecov.io/gh/open-mmlab/mmdetection3d)
[![license](https://img.shields.io/github/license/open-mmlab/mmdetection3d.svg)](https://github.com/open-mmlab/mmdetection3d/blob/master/LICENSE)


Tai-Wang's avatar
Tai-Wang committed
11
**News**: We released the codebase v0.17.0.
twang's avatar
twang committed
12

Tai-Wang's avatar
Tai-Wang committed
13
In the [nuScenes 3D detection challenge](https://www.nuscenes.org/object-detection?externalData=all&mapData=all&modalities=Any) of the 5th AI Driving Olympics in NeurIPS 2020, we obtained the best PKL award and the second runner-up by multi-modality entry, and the best vision-only results.
14
15

Code and models for the best vision-only method, [FCOS3D](https://arxiv.org/abs/2104.10956), have been released. Please stay tuned for [MoCa](https://arxiv.org/abs/2012.12741).
zhangwenwei's avatar
zhangwenwei committed
16
17

Documentation: https://mmdetection3d.readthedocs.io/
zhangwenwei's avatar
zhangwenwei committed
18
19
20

## Introduction

21
22
23
English | [简体中文](README_zh-CN.md)

The master branch works with **PyTorch 1.3+**.
zhangwenwei's avatar
zhangwenwei committed
24

25
MMDetection3D is an open source object detection toolbox based on PyTorch, towards the next-generation platform for general 3D detection. It is
zhangwenwei's avatar
zhangwenwei committed
26
a part of the OpenMMLab project developed by [MMLab](http://mmlab.ie.cuhk.edu.hk/).
zhangwenwei's avatar
zhangwenwei committed
27

zhangwenwei's avatar
zhangwenwei committed
28
![demo image](resources/mmdet3d_outdoor_demo.gif)
zhangwenwei's avatar
zhangwenwei committed
29
30
31

### Major features

zhangwenwei's avatar
zhangwenwei committed
32
- **Support multi-modality/single-modality detectors out of box**
zhangwenwei's avatar
zhangwenwei committed
33

34
  It directly supports multi-modality/single-modality detectors including MVXNet, VoteNet, PointPillars, etc.
zhangwenwei's avatar
zhangwenwei committed
35

zhangwenwei's avatar
zhangwenwei committed
36
- **Support indoor/outdoor 3D detection out of box**
zhangwenwei's avatar
zhangwenwei committed
37

Wenwei Zhang's avatar
Wenwei Zhang committed
38
  It directly supports popular indoor and outdoor 3D detection datasets, including ScanNet, SUNRGB-D, Waymo, nuScenes, Lyft, and KITTI.
39
  For nuScenes dataset, we also support [nuImages dataset](https://github.com/open-mmlab/mmdetection3d/tree/master/configs/nuimages).
zhangwenwei's avatar
zhangwenwei committed
40

zhangwenwei's avatar
zhangwenwei committed
41
- **Natural integration with 2D detection**
42

43
  All the about **300+ models, methods of 40+ papers**, and modules supported in [MMDetection](https://github.com/open-mmlab/mmdetection/blob/master/docs/model_zoo.md) can be trained or used in this codebase.
zhangwenwei's avatar
zhangwenwei committed
44

zhangwenwei's avatar
zhangwenwei committed
45
- **High efficiency**
zhangwenwei's avatar
zhangwenwei committed
46

Wenwei Zhang's avatar
Wenwei Zhang committed
47
  It trains faster than other codebases. The main results are as below. Details can be found in [benchmark.md](./docs/benchmarks.md). We compare the number of samples trained per second (the higher, the better). The models that are not supported by other codebases are marked by `×`.
zhangwenwei's avatar
zhangwenwei committed
48

Wenwei Zhang's avatar
Wenwei Zhang committed
49
50
51
52
53
54
55
56
57
  | Methods | MMDetection3D | [OpenPCDet](https://github.com/open-mmlab/OpenPCDet) |[votenet](https://github.com/facebookresearch/votenet)| [Det3D](https://github.com/poodarchu/Det3D) |
  |:-------:|:-------------:|:---------:|:-----:|:-----:|
  | VoteNet | 358           | ×         |   77  | ×     |
  | PointPillars-car| 141           | ×         |   ×  | 140     |
  | PointPillars-3class| 107           |44     |   ×      | ×    |
  | SECOND| 40           |30     |   ×      | ×    |
  | Part-A2| 17           |14     |   ×      | ×    |

Like [MMDetection](https://github.com/open-mmlab/mmdetection) and [MMCV](https://github.com/open-mmlab/mmcv), MMDetection3D can also be used as a library to support different projects on top of it.
zhangwenwei's avatar
zhangwenwei committed
58
59
60
61
62

## License

This project is released under the [Apache 2.0 license](LICENSE).

zhangwenwei's avatar
zhangwenwei committed
63
## Changelog
zhangwenwei's avatar
zhangwenwei committed
64

Tai-Wang's avatar
Tai-Wang committed
65
v0.17.0 was released in 1/9/2021.
zhangwenwei's avatar
zhangwenwei committed
66
Please refer to [changelog.md](docs/changelog.md) for details and release history.
zhangwenwei's avatar
zhangwenwei committed
67
68
69
70

## Benchmark and model zoo

Supported methods and backbones are shown in the below table.
zhangwenwei's avatar
zhangwenwei committed
71
Results and models are available in the [model zoo](docs/model_zoo.md).
zhangwenwei's avatar
zhangwenwei committed
72

xiliu8006's avatar
xiliu8006 committed
73
74
75
76
77
Support backbones:

- [x] PointNet (CVPR'2017)
- [x] PointNet++ (NeurIPS'2017)
- [x] RegNet (CVPR'2020)
78
- [x] DGCNN (TOG'2019)
xiliu8006's avatar
xiliu8006 committed
79
80
81
82
83
84
85
86
87
88
89

Support methods

- [x] [SECOND (Sensor'2018)](configs/second/README.md)
- [x] [PointPillars (CVPR'2019)](configs/pointpillars/README.md)
- [x] [FreeAnchor (NeurIPS'2019)](configs/free_anchor/README.md)
- [x] [VoteNet (ICCV'2019)](configs/votenet/README.md)
- [x] [H3DNet (ECCV'2020)](configs/h3dnet/README.md)
- [x] [3DSSD (CVPR'2020)](configs/3dssd/README.md)
- [x] [Part-A2 (TPAMI'2020)](configs/parta2/README.md)
- [x] [MVXNet (ICRA'2019)](configs/mvxnet/README.md)
Tianwei Yin's avatar
Tianwei Yin committed
90
- [x] [CenterPoint (CVPR'2021)](configs/centerpoint/README.md)
xiliu8006's avatar
xiliu8006 committed
91
- [x] [SSN (ECCV'2020)](configs/ssn/README.md)
92
- [x] [ImVoteNet (CVPR'2020)](configs/imvotenet/README.md)
93
- [x] [FCOS3D (Arxiv'2021)](configs/fcos3d/README.md)
94
- [x] [PointNet++ (NeurIPS'2017)](configs/pointnet2/README.md)
95
- [x] [Group-Free-3D (Arxiv'2021)](configs/groupfree3d/README.md)
96
97
- [x] [ImVoxelNet (Arxiv'2021)](configs/imvoxelnet/README.md)
- [x] [PAConv (CVPR'2021)](configs/paconv/README.md)
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
- [x] [DGCNN (TOG'2019)](configs/dgcnn/README.md)

|                    | ResNet   | ResNeXt  | SENet    |PointNet++ |DGCNN     | HRNet | RegNetX | Res2Net |
|--------------------|:--------:|:--------:|:--------:|:---------:|:---------:|:-----:|:--------:|:-----:|
| SECOND             | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| PointPillars       | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| FreeAnchor         | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| VoteNet            | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| H3DNet            | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| 3DSSD            | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| Part-A2            | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| MVXNet             | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| CenterPoint        | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| SSN                | ☐        | ☐        | ☐        | ✗         | ✗         | ☐     | ✓        | ☐     |
| ImVoteNet            | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| FCOS3D               | ✓        | ☐        | ☐        | ✗         | ✗         | ☐     | ☐        | ☐     |
| PointNet++           | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| Group-Free-3D        | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| ImVoxelNet           | ✓         | ✗        | ✗        | ✗        | ✗         | ✗     | ✗        | ✗     |
| PAConv               | ✗        | ✗        | ✗        | ✓         | ✗         | ✗     | ✗        | ✗     |
| DGCNN                | ✗        | ✗        | ✗        | ✗         | ✓         | ✗     | ✗        | ✗     |
zhangwenwei's avatar
zhangwenwei committed
119
120

Other features
121
- [x] [Dynamic Voxelization](configs/dynamic_voxelization/README.md)
zhangwenwei's avatar
zhangwenwei committed
122

123
**Note:** All the about **300+ models, methods of 40+ papers** in 2D detection supported by [MMDetection](https://github.com/open-mmlab/mmdetection/blob/master/docs/model_zoo.md) can be trained or used in this codebase.
zhangwenwei's avatar
zhangwenwei committed
124
125
126

## Installation

twang's avatar
twang committed
127
Please refer to [getting_started.md](docs/getting_started.md) for installation.
zhangwenwei's avatar
zhangwenwei committed
128
129
130

## Get Started

131
Please see [getting_started.md](docs/getting_started.md) for the basic usage of MMDetection3D. We provide guidance for quick run [with existing dataset](docs/1_exist_data_model.md) and [with customized dataset](docs/2_new_data_model.md) for beginners. There are also tutorials for [learning configuration systems](docs/tutorials/config.md), [adding new dataset](docs/tutorials/customize_dataset.md), [designing data pipeline](docs/tutorials/data_pipeline.md), [customizing models](docs/tutorials/customize_models.md), [customizing runtime settings](docs/tutorials/customize_runtime.md) and [Waymo dataset](docs/datasets/waymo_det.md).
zhangwenwei's avatar
zhangwenwei committed
132

133
134
Please refer to [FAQ](docs/faq.md) for frequently asked questions. When updating the version of MMDetection3D, please also check the [compatibility doc](docs/compatibility.md) to be aware of the BC-breaking updates introduced in each version.

135
136
137
138
139
140
## Citation

If you find this project useful in your research, please consider cite:

```latex
@misc{mmdet3d2020,
Ziyi Wu's avatar
Ziyi Wu committed
141
    title={{MMDetection3D: OpenMMLab} next-generation platform for general {3D} object detection},
142
143
144
145
146
147
    author={MMDetection3D Contributors},
    howpublished = {\url{https://github.com/open-mmlab/mmdetection3d}},
    year={2020}
}
```

zhangwenwei's avatar
zhangwenwei committed
148
149
## Contributing

zhangwenwei's avatar
zhangwenwei committed
150
We appreciate all contributions to improve MMDetection3D. Please refer to [CONTRIBUTING.md](.github/CONTRIBUTING.md) for the contributing guideline.
zhangwenwei's avatar
zhangwenwei committed
151
152
153

## Acknowledgement

zhangwenwei's avatar
zhangwenwei committed
154
MMDetection3D is an open source project that is contributed by researchers and engineers from various colleges and companies. We appreciate all the contributors as well as users who give valuable feedbacks.
zhangwenwei's avatar
zhangwenwei committed
155
We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new 3D detectors.
156
157
158
159

## Projects in OpenMMLab

- [MMCV](https://github.com/open-mmlab/mmcv): OpenMMLab foundational library for computer vision.
Wenhao Wu's avatar
Wenhao Wu committed
160
- [MIM](https://github.com/open-mmlab/mim): MIM Installs OpenMMLab Packages.
161
162
163
164
165
166
167
168
- [MMClassification](https://github.com/open-mmlab/mmclassification): OpenMMLab image classification toolbox and benchmark.
- [MMDetection](https://github.com/open-mmlab/mmdetection): OpenMMLab detection toolbox and benchmark.
- [MMDetection3D](https://github.com/open-mmlab/mmdetection3d): OpenMMLab next-generation platform for general 3D object detection.
- [MMSegmentation](https://github.com/open-mmlab/mmsegmentation): OpenMMLab semantic segmentation toolbox and benchmark.
- [MMAction2](https://github.com/open-mmlab/mmaction2): OpenMMLab's next-generation action understanding toolbox and benchmark.
- [MMTracking](https://github.com/open-mmlab/mmtracking): OpenMMLab video perception toolbox and benchmark.
- [MMPose](https://github.com/open-mmlab/mmpose): OpenMMLab pose estimation toolbox and benchmark.
- [MMEditing](https://github.com/open-mmlab/mmediting): OpenMMLab image and video editing toolbox.
169
170
- [MMOCR](https://github.com/open-mmlab/mmocr): OpenMMLab text detection, recognition and understanding toolbox.
- [MMGeneration](https://github.com/open-mmlab/mmgeneration): OpenMMLab image and video generative models toolbox.