Commit 0129ab02 authored by Rick Ho's avatar Rick Ho
Browse files

update readme

parent 91a5e794
...@@ -99,8 +99,20 @@ FastMoE's model parallel requires sophiscated parallel strategies that neither P ...@@ -99,8 +99,20 @@ FastMoE's model parallel requires sophiscated parallel strategies that neither P
Megatron-LM provides. The `fmoe.DistributedGroupedDataParallel` module is Megatron-LM provides. The `fmoe.DistributedGroupedDataParallel` module is
introduced to replace PyTorch's DDP module. introduced to replace PyTorch's DDP module.
#### Faster Performance Features
From a PPoPP'22 paper, _FasterMoE: modeling and optimizing training of
large-scale dynamic pre-trained models_, we have adopted techniques to make
FastMoE's model parallel much more efficient.
These optimizations are named as **Faster Performance Features**, and can be
enabled via several environment variables. Their usage and constraints are
detailed in [a separate document](doc/fastermoe).
## Citation ## Citation
For the core FastMoE system.
``` ```
@article{he2021fastmoe, @article{he2021fastmoe,
title={FastMoE: A Fast Mixture-of-Expert Training System}, title={FastMoE: A Fast Mixture-of-Expert Training System},
...@@ -110,6 +122,27 @@ introduced to replace PyTorch's DDP module. ...@@ -110,6 +122,27 @@ introduced to replace PyTorch's DDP module.
} }
``` ```
For the [faster performance features](doc/fastermoe).
```
@inproceedings{he2022fastermoe,
author = {He, Jiaao and Zhai, Jidong and Antunes, Tiago and Wang, Haojie and Luo, Fuwen and Shi, Shangfeng and Li, Qin},
title = {FasterMoE: Modeling and Optimizing Training of Large-Scale Dynamic Pre-Trained Models},
year = {2022},
isbn = {9781450392044},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3503221.3508418},
doi = {10.1145/3503221.3508418},
booktitle = {Proceedings of the 27th ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming},
pages = {120–134},
numpages = {15},
keywords = {parallelism, distributed deep learning, performance modeling},
location = {Seoul, Republic of Korea},
series = {PPoPP '22}
}
```
## Troubleshootings / Discussion ## Troubleshootings / Discussion
If you have any problem using FastMoE, or you are interested in getting involved in developing FastMoE, feel free to join the [our slack channel](https://join.slack.com/t/fastmoe/shared_invite/zt-mz0ai6ol-ggov75D62YsgHfzShw8KYw). If you have any problem using FastMoE, or you are interested in getting involved in developing FastMoE, feel free to join the [our slack channel](https://join.slack.com/t/fastmoe/shared_invite/zt-mz0ai6ol-ggov75D62YsgHfzShw8KYw).
...@@ -13,6 +13,7 @@ authors = [ ...@@ -13,6 +13,7 @@ authors = [
'Tiago Antunes', 'Tiago Antunes',
'Jinjun Peng', 'Jinjun Peng',
'Qin Li', 'Qin Li',
'Mingshu Zhai'
] ]
is_rocm_pytorch = False is_rocm_pytorch = False
...@@ -37,7 +38,7 @@ else: ...@@ -37,7 +38,7 @@ else:
if __name__ == '__main__': if __name__ == '__main__':
setuptools.setup( setuptools.setup(
name='fastmoe', name='fastmoe',
version='0.3.0', version='1.0.0',
description='An efficient Mixture-of-Experts system for PyTorch', description='An efficient Mixture-of-Experts system for PyTorch',
author=', '.join(authors), author=', '.join(authors),
author_email='hja20@mails.tsinghua.edu.cn', author_email='hja20@mails.tsinghua.edu.cn',
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment