README.md 12.4 KB
Newer Older
zhangwenwei's avatar
zhangwenwei committed
1
<div align="center">
zhangwenwei's avatar
zhangwenwei committed
2
  <img src="resources/mmdet3d-logo.png" width="600"/>
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
  <div>&nbsp;</div>
  <div align="center">
    <b><font size="5">OpenMMLab website</font></b>
    <sup>
      <a href="https://openmmlab.com">
        <i><font size="4">HOT</font></i>
      </a>
    </sup>
    &nbsp;&nbsp;&nbsp;&nbsp;
    <b><font size="5">OpenMMLab platform</font></b>
    <sup>
      <a href="https://platform.openmmlab.com">
        <i><font size="4">TRY IT OUT</font></i>
      </a>
    </sup>
  </div>
  <div>&nbsp;</div>
zhangwenwei's avatar
zhangwenwei committed
20
</div>
zhangwenwei's avatar
zhangwenwei committed
21

Wenwei Zhang's avatar
Wenwei Zhang committed
22
23
24
25
26
27
[![docs](https://img.shields.io/badge/docs-latest-blue)](https://mmdetection3d.readthedocs.io/en/latest/)
[![badge](https://github.com/open-mmlab/mmdetection3d/workflows/build/badge.svg)](https://github.com/open-mmlab/mmdetection3d/actions)
[![codecov](https://codecov.io/gh/open-mmlab/mmdetection3d/branch/master/graph/badge.svg)](https://codecov.io/gh/open-mmlab/mmdetection3d)
[![license](https://img.shields.io/github/license/open-mmlab/mmdetection3d.svg)](https://github.com/open-mmlab/mmdetection3d/blob/master/LICENSE)


Wenhao Wu's avatar
Wenhao Wu committed
28
**News**: We released the codebase v0.18.0.
Tai-Wang's avatar
Tai-Wang committed
29

Tai-Wang's avatar
Tai-Wang committed
30
31
32
In addition, we have preliminarily supported several new models on the [v1.0.0.dev0](https://github.com/open-mmlab/mmdetection3d/tree/v1.0.0.dev0) branch, including [DGCNN](https://github.com/open-mmlab/mmdetection3d/blob/v1.0.0.dev0/configs/dgcnn/README.md), [SMOKE](https://github.com/open-mmlab/mmdetection3d/blob/v1.0.0.dev0/configs/smoke/README.md) and [PGD](https://github.com/open-mmlab/mmdetection3d/blob/v1.0.0.dev0/configs/pgd/README.md).

Note: We are going through large refactoring to provide simpler and more unified usage of many modules. Thus, few features will be added to the master branch in the following months.
Tai-Wang's avatar
Tai-Wang committed
33
34
35

The compatibilities of models are broken due to the unification and simplification of coordinate systems. For now, most models are benchmarked with similar performance, though few models are still being benchmarked.

Tai-Wang's avatar
Tai-Wang committed
36
You can start experiments with [v1.0.0.dev0](https://github.com/open-mmlab/mmdetection3d/tree/v1.0.0.dev0) if you are interested. Please note that our new features will only be supported in v1.0.0 branch afterward.
twang's avatar
twang committed
37

Tai-Wang's avatar
Tai-Wang committed
38
In the [nuScenes 3D detection challenge](https://www.nuscenes.org/object-detection?externalData=all&mapData=all&modalities=Any) of the 5th AI Driving Olympics in NeurIPS 2020, we obtained the best PKL award and the second runner-up by multi-modality entry, and the best vision-only results.
39
40

Code and models for the best vision-only method, [FCOS3D](https://arxiv.org/abs/2104.10956), have been released. Please stay tuned for [MoCa](https://arxiv.org/abs/2012.12741).
zhangwenwei's avatar
zhangwenwei committed
41
42

Documentation: https://mmdetection3d.readthedocs.io/
zhangwenwei's avatar
zhangwenwei committed
43
44
45

## Introduction

46
47
48
English | [简体中文](README_zh-CN.md)

The master branch works with **PyTorch 1.3+**.
zhangwenwei's avatar
zhangwenwei committed
49

50
MMDetection3D is an open source object detection toolbox based on PyTorch, towards the next-generation platform for general 3D detection. It is
zhangwenwei's avatar
zhangwenwei committed
51
a part of the OpenMMLab project developed by [MMLab](http://mmlab.ie.cuhk.edu.hk/).
zhangwenwei's avatar
zhangwenwei committed
52

zhangwenwei's avatar
zhangwenwei committed
53
![demo image](resources/mmdet3d_outdoor_demo.gif)
zhangwenwei's avatar
zhangwenwei committed
54
55
56

### Major features

zhangwenwei's avatar
zhangwenwei committed
57
- **Support multi-modality/single-modality detectors out of box**
zhangwenwei's avatar
zhangwenwei committed
58

59
  It directly supports multi-modality/single-modality detectors including MVXNet, VoteNet, PointPillars, etc.
zhangwenwei's avatar
zhangwenwei committed
60

zhangwenwei's avatar
zhangwenwei committed
61
- **Support indoor/outdoor 3D detection out of box**
zhangwenwei's avatar
zhangwenwei committed
62

Wenwei Zhang's avatar
Wenwei Zhang committed
63
  It directly supports popular indoor and outdoor 3D detection datasets, including ScanNet, SUNRGB-D, Waymo, nuScenes, Lyft, and KITTI.
64
  For nuScenes dataset, we also support [nuImages dataset](https://github.com/open-mmlab/mmdetection3d/tree/master/configs/nuimages).
zhangwenwei's avatar
zhangwenwei committed
65

zhangwenwei's avatar
zhangwenwei committed
66
- **Natural integration with 2D detection**
67

Wenhao Wu's avatar
Wenhao Wu committed
68
  All the about **300+ models, methods of 40+ papers**, and modules supported in [MMDetection](https://github.com/open-mmlab/mmdetection/blob/master/docs/en/model_zoo.md) can be trained or used in this codebase.
zhangwenwei's avatar
zhangwenwei committed
69

zhangwenwei's avatar
zhangwenwei committed
70
- **High efficiency**
zhangwenwei's avatar
zhangwenwei committed
71

Wenhao Wu's avatar
Wenhao Wu committed
72
  It trains faster than other codebases. The main results are as below. Details can be found in [benchmark.md](./docs/en/benchmarks.md). We compare the number of samples trained per second (the higher, the better). The models that are not supported by other codebases are marked by `×`.
zhangwenwei's avatar
zhangwenwei committed
73

Wenwei Zhang's avatar
Wenwei Zhang committed
74
75
76
77
78
79
80
81
82
  | Methods | MMDetection3D | [OpenPCDet](https://github.com/open-mmlab/OpenPCDet) |[votenet](https://github.com/facebookresearch/votenet)| [Det3D](https://github.com/poodarchu/Det3D) |
  |:-------:|:-------------:|:---------:|:-----:|:-----:|
  | VoteNet | 358           | ×         |   77  | ×     |
  | PointPillars-car| 141           | ×         |   ×  | 140     |
  | PointPillars-3class| 107           |44     |   ×      | ×    |
  | SECOND| 40           |30     |   ×      | ×    |
  | Part-A2| 17           |14     |   ×      | ×    |

Like [MMDetection](https://github.com/open-mmlab/mmdetection) and [MMCV](https://github.com/open-mmlab/mmcv), MMDetection3D can also be used as a library to support different projects on top of it.
zhangwenwei's avatar
zhangwenwei committed
83
84
85
86
87

## License

This project is released under the [Apache 2.0 license](LICENSE).

zhangwenwei's avatar
zhangwenwei committed
88
## Changelog
zhangwenwei's avatar
zhangwenwei committed
89

Wenhao Wu's avatar
Wenhao Wu committed
90
v0.18.0 was released in 1/1/2022.
Wenhao Wu's avatar
Wenhao Wu committed
91
Please refer to [changelog.md](docs/en/changelog.md) for details and release history.
zhangwenwei's avatar
zhangwenwei committed
92

Tai-Wang's avatar
Tai-Wang committed
93
94
For branch v1.0.0.dev0, please refer to [changelog_v1.0.md](https://github.com/Tai-Wang/mmdetection3d/blob/v1.0.0.dev0-changelog/docs/changelog_v1.0.md) for our latest features and more details.

zhangwenwei's avatar
zhangwenwei committed
95
96
97
## Benchmark and model zoo

Supported methods and backbones are shown in the below table.
Wenhao Wu's avatar
Wenhao Wu committed
98
Results and models are available in the [model zoo](docs/en/model_zoo.md).
zhangwenwei's avatar
zhangwenwei committed
99

xiliu8006's avatar
xiliu8006 committed
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
Support backbones:

- [x] PointNet (CVPR'2017)
- [x] PointNet++ (NeurIPS'2017)
- [x] RegNet (CVPR'2020)

Support methods

- [x] [SECOND (Sensor'2018)](configs/second/README.md)
- [x] [PointPillars (CVPR'2019)](configs/pointpillars/README.md)
- [x] [FreeAnchor (NeurIPS'2019)](configs/free_anchor/README.md)
- [x] [VoteNet (ICCV'2019)](configs/votenet/README.md)
- [x] [H3DNet (ECCV'2020)](configs/h3dnet/README.md)
- [x] [3DSSD (CVPR'2020)](configs/3dssd/README.md)
- [x] [Part-A2 (TPAMI'2020)](configs/parta2/README.md)
- [x] [MVXNet (ICRA'2019)](configs/mvxnet/README.md)
Tianwei Yin's avatar
Tianwei Yin committed
116
- [x] [CenterPoint (CVPR'2021)](configs/centerpoint/README.md)
xiliu8006's avatar
xiliu8006 committed
117
- [x] [SSN (ECCV'2020)](configs/ssn/README.md)
118
- [x] [ImVoteNet (CVPR'2020)](configs/imvotenet/README.md)
119
- [x] [FCOS3D (ICCVW'2021)](configs/fcos3d/README.md)
120
- [x] [PointNet++ (NeurIPS'2017)](configs/pointnet2/README.md)
121
- [x] [Group-Free-3D (ICCV'2021)](configs/groupfree3d/README.md)
122
- [x] [ImVoxelNet (WACV'2022)](configs/imvoxelnet/README.md)
123
- [x] [PAConv (CVPR'2021)](configs/paconv/README.md)
xiliu8006's avatar
xiliu8006 committed
124

zhangwenwei's avatar
zhangwenwei committed
125
126
|                    | ResNet   | ResNeXt  | SENet    |PointNet++ | HRNet | RegNetX | Res2Net |
|--------------------|:--------:|:--------:|:--------:|:---------:|:-----:|:--------:|:-----:|
127
128
129
| SECOND             | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
| PointPillars       | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
| FreeAnchor         | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
zhangwenwei's avatar
zhangwenwei committed
130
| VoteNet            | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
131
132
| H3DNet            | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
| 3DSSD            | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
133
134
| Part-A2            | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
| MVXNet             | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
Wenwei Zhang's avatar
Wenwei Zhang committed
135
| CenterPoint        | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
136
| SSN                | ☐        | ☐        | ☐        | ✗         | ☐     | ✓        | ☐     |
137
| ImVoteNet            | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
138
| FCOS3D               | ✓        | ☐        | ☐        | ✗         | ☐     | ☐        | ☐     |
139
| PointNet++           | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
140
| Group-Free-3D        | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
141
142
| ImVoxelNet           | ✓         | ✗        | ✗        | ✗        | ✗     | ✗        | ✗     |
| PAConv               | ✗        | ✗        | ✗        | ✓         | ✗     | ✗        | ✗     |
zhangwenwei's avatar
zhangwenwei committed
143
144

Other features
145
- [x] [Dynamic Voxelization](configs/dynamic_voxelization/README.md)
zhangwenwei's avatar
zhangwenwei committed
146

Wenhao Wu's avatar
Wenhao Wu committed
147
**Note:** All the about **300+ models, methods of 40+ papers** in 2D detection supported by [MMDetection](https://github.com/open-mmlab/mmdetection/blob/master/docs/en/model_zoo.md) can be trained or used in this codebase.
zhangwenwei's avatar
zhangwenwei committed
148
149
150

## Installation

Wenhao Wu's avatar
Wenhao Wu committed
151
Please refer to [getting_started.md](docs/en/getting_started.md) for installation.
zhangwenwei's avatar
zhangwenwei committed
152
153
154

## Get Started

Wenhao Wu's avatar
Wenhao Wu committed
155
Please see [getting_started.md](docs/en/getting_started.md) for the basic usage of MMDetection3D. We provide guidance for quick run [with existing dataset](docs/en/1_exist_data_model.md) and [with customized dataset](docs/en/2_new_data_model.md) for beginners. There are also tutorials for [learning configuration systems](docs/en/tutorials/config.md), [adding new dataset](docs/en/tutorials/customize_dataset.md), [designing data pipeline](docs/en/tutorials/data_pipeline.md), [customizing models](docs/en/tutorials/customize_models.md), [customizing runtime settings](docs/en/tutorials/customize_runtime.md) and [Waymo dataset](docs/en/datasets/waymo_det.md).
zhangwenwei's avatar
zhangwenwei committed
156

Wenhao Wu's avatar
Wenhao Wu committed
157
Please refer to [FAQ](docs/en/faq.md) for frequently asked questions. When updating the version of MMDetection3D, please also check the [compatibility doc](docs/en/compatibility.md) to be aware of the BC-breaking updates introduced in each version.
158

159
160
161
162
163
164
## Citation

If you find this project useful in your research, please consider cite:

```latex
@misc{mmdet3d2020,
Ziyi Wu's avatar
Ziyi Wu committed
165
    title={{MMDetection3D: OpenMMLab} next-generation platform for general {3D} object detection},
166
167
168
169
170
171
    author={MMDetection3D Contributors},
    howpublished = {\url{https://github.com/open-mmlab/mmdetection3d}},
    year={2020}
}
```

zhangwenwei's avatar
zhangwenwei committed
172
173
## Contributing

zhangwenwei's avatar
zhangwenwei committed
174
We appreciate all contributions to improve MMDetection3D. Please refer to [CONTRIBUTING.md](.github/CONTRIBUTING.md) for the contributing guideline.
zhangwenwei's avatar
zhangwenwei committed
175
176
177

## Acknowledgement

zhangwenwei's avatar
zhangwenwei committed
178
MMDetection3D is an open source project that is contributed by researchers and engineers from various colleges and companies. We appreciate all the contributors as well as users who give valuable feedbacks.
zhangwenwei's avatar
zhangwenwei committed
179
We wish that the toolbox and benchmark could serve the growing research community by providing a flexible toolkit to reimplement existing methods and develop their own new 3D detectors.
180
181
182
183

## Projects in OpenMMLab

- [MMCV](https://github.com/open-mmlab/mmcv): OpenMMLab foundational library for computer vision.
Wenhao Wu's avatar
Wenhao Wu committed
184
- [MIM](https://github.com/open-mmlab/mim): MIM Installs OpenMMLab Packages.
185
186
187
188
189
190
191
192
- [MMClassification](https://github.com/open-mmlab/mmclassification): OpenMMLab image classification toolbox and benchmark.
- [MMDetection](https://github.com/open-mmlab/mmdetection): OpenMMLab detection toolbox and benchmark.
- [MMDetection3D](https://github.com/open-mmlab/mmdetection3d): OpenMMLab next-generation platform for general 3D object detection.
- [MMSegmentation](https://github.com/open-mmlab/mmsegmentation): OpenMMLab semantic segmentation toolbox and benchmark.
- [MMAction2](https://github.com/open-mmlab/mmaction2): OpenMMLab's next-generation action understanding toolbox and benchmark.
- [MMTracking](https://github.com/open-mmlab/mmtracking): OpenMMLab video perception toolbox and benchmark.
- [MMPose](https://github.com/open-mmlab/mmpose): OpenMMLab pose estimation toolbox and benchmark.
- [MMEditing](https://github.com/open-mmlab/mmediting): OpenMMLab image and video editing toolbox.
193
194
- [MMOCR](https://github.com/open-mmlab/mmocr): OpenMMLab text detection, recognition and understanding toolbox.
- [MMGeneration](https://github.com/open-mmlab/mmgeneration): OpenMMLab image and video generative models toolbox.
ChaimZhu's avatar
ChaimZhu committed
195
- [MMFlow](https://github.com/open-mmlab/mmflow): OpenMMLab optical flow toolbox and benchmark.
ChaimZhu's avatar
ChaimZhu committed
196
197
- [MMFewShot](https://github.com/open-mmlab/mmfewshot): OpenMMLab fewshot learning toolbox and benchmark.
- [MMHuman3D](https://github.com/open-mmlab/mmhuman3d): OpenMMLab 3D human parametric model toolbox and benchmark.
198
199
200
- [MMSelfSup](https://github.com/open-mmlab/mmselfsup): OpenMMLab self-supervised learning toolbox and benchmark.
- [MMRazor](https://github.com/open-mmlab/mmrazor): OpenMMLab Model Compression Toolbox and Benchmark.
- [MMDeploy](https://github.com/open-mmlab/mmdeploy): OpenMMLab Model Deployment Framework.