[Feature] Support PyTorch backend on MLU (#1770)
* feat(MLU): Support PyTorch backend on MLU * MMCV support PyTorch backend on MLU * Add MLUDataParallel and MLUDistributedDataParallel * Add MLU operator support * [Fix]: Fix PR comments and add IS_MLU to get device available check * [Fix]: fix PR comments of dist_utils.py * [Doc] Rewrite annotations of functions. * [Docs] Rewrite annotation in distributed.py Co-authored-by:Zaida Zhou <58739961+zhouzaida@users.noreply.github.com> * [Docs] Fix lint Co-authored-by:
Zaida Zhou <58739961+zhouzaida@users.noreply.github.com>
Showing
mmcv/device/__init__.py
0 → 100644
mmcv/device/mlu/__init__.py
0 → 100644
mmcv/device/mlu/utils.py
0 → 100644
Please register or sign in to comment