- 28 Jan, 2022 1 commit
-
-
Frank Lee authored
-
- 25 Jan, 2022 1 commit
-
-
Jiarui Fang authored
* add pytorch hooks fix #175 * remove licenses in src code * add gpu memory tracer * replacing print with logger in ophooks.
-
- 21 Jan, 2022 3 commits
- 20 Jan, 2022 2 commits
- 19 Jan, 2022 6 commits
- 18 Jan, 2022 6 commits
- 17 Jan, 2022 2 commits
- 13 Jan, 2022 1 commit
-
-
ver217 authored
-
- 10 Jan, 2022 2 commits
-
-
BoxiangW authored
Update the documentations of layer integration Update _log_hook.py Update _operation.py
-
binmakeswell authored
-
- 07 Jan, 2022 5 commits
- 06 Jan, 2022 2 commits
-
-
Frank Lee authored
* enable CI after PR sync * Fixed github action
-
binmakeswell authored
-
- 05 Jan, 2022 1 commit
-
-
Jiarui Fang authored
-
- 04 Jan, 2022 4 commits
- 30 Dec, 2021 2 commits
-
-
ver217 authored
* add pipeline shared module wrapper and update load batch * added model parallel process group for amp and clip grad (#86) * added model parallel process group for amp and clip grad * update amp and clip with model parallel process group * remove pipeline_prev/next group (#88) * micro batch offload * optimize pipeline gpu memory usage * pipeline can receive tensor shape (#93) * optimize pipeline gpu memory usage * fix grad accumulation step counter * rename classes and functions Co-authored-by:Frank Lee <somerlee.9@gmail.com>
-
アマデウス authored
-
- 29 Dec, 2021 1 commit
-
-
アマデウス authored
* optimized 1d layer apis; reorganized nn.layer modules; fixed tests * fixed 2.5d runtime issue * reworked split batch, now called in trainer.schedule.load_batch Co-authored-by:BoxiangW <45734921+BoxiangW@users.noreply.github.com>
-
- 27 Dec, 2021 1 commit
-
-
アマデウス authored
* integrated parallel layers for ease of building models * integrated 2.5d layers * cleaned codes and unit tests * added log metric by step hook; updated imagenet benchmark; fixed some bugs * reworked initialization; cleaned codes Co-authored-by:BoxiangW <45734921+BoxiangW@users.noreply.github.com>
-