Commit 0c3aa2cd authored by Rick Ho's avatar Rick Ho
Browse files

checkout to version v0.1.1

parent 53b5b8c3
## v0.1.1
### Distributed
- Broadcast data-parallel parameters before training.
### Megatron adaption
- Initialize `FMoELinear` parameters using different seed in model parallel even using the same random seed in megatron.
- Use proper comm for mp and dp.
### Transformer-XL example
- Improve scripts.
### Misc
- Logo and slack workspace link.
- Document in Chinese.
- Figures to explain how FastMoE works.
## v0.1.0 ## v0.1.0
### Functions ### Functions
......
...@@ -15,9 +15,9 @@ if os.environ.get('USE_NCCL', '0') == '1': ...@@ -15,9 +15,9 @@ if os.environ.get('USE_NCCL', '0') == '1':
if __name__ == '__main__': if __name__ == '__main__':
setuptools.setup( setuptools.setup(
name='fmoe', name='fastmoe',
version='0.1.0', version='0.1.1',
description='An efficient Mixture-of-Experts impl. for PyTorch', description='An efficient Mixture-of-Experts system for PyTorch',
author='Jiaao He, Jiezhong Qiu and Aohan Zeng', author='Jiaao He, Jiezhong Qiu and Aohan Zeng',
author_email='hja20@mails.tsinghua.edu.cn', author_email='hja20@mails.tsinghua.edu.cn',
license='Apache-2', license='Apache-2',
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment