- 29 Mar, 2022 7 commits
-
-
Liang Bowen authored
-
Jiarui Fang authored
-
Jie Zhu authored
* add memory trainer hook * fix bug * add memory trainer hook * fix import bug * fix import bug * add trainer hook * fix #370 git log bug * modify `to_tensorboard` function to support better output * remove useless output * change the name of `MemProfiler` * complete memory profiler * replace error with warning * finish trainer hook * modify interface of MemProfiler * modify `__init__.py` in profiler * remove unnecessary pass statement * add usage to doc string * add usage to trainer hook * new location to store temp data file
-
ver217 authored
* optimize grad offload * polish code * polish code
-
Jiarui Fang authored
-
ver217 authored
-
Jiarui Fang authored
-
- 28 Mar, 2022 4 commits
-
-
HELSON authored
only process module's own parameters in Zero context add zero hooks for all modules that contrain parameters gather parameters only belonging to module itself
-
Jiarui Fang authored
-
Jiarui Fang authored
-
Jiarui Fang authored
-
- 26 Mar, 2022 1 commit
-
-
ver217 authored
-
- 25 Mar, 2022 12 commits
-
-
ver217 authored
-
Jiarui Fang authored
* [zero] fix grad offload * polish code
-
LuGY authored
* [zero]added hybrid adam, removed loss scale of adam * remove useless code
-
Jiarui Fang authored
-
Frank Lee authored
-
Jiarui Fang authored
-
LuGY authored
-
Jiarui Fang authored
-
ver217 authored
-
Jiarui Fang authored
-
fastalgo authored
-
Jiarui Fang authored
-
- 24 Mar, 2022 10 commits
-
-
ver217 authored
* enable rm_torch_payload_on_the_fly * polish docstr
-
Jiarui Fang authored
-
HELSON authored
-
Jiarui Fang authored
-
Jiarui Fang authored
-
Frank Lee authored
-
Jiarui Fang authored
-
Frank Lee authored
-
Frank Lee authored
-
binmakeswell authored
-
- 23 Mar, 2022 2 commits
-
-
Jiarui Fang authored
-
ver217 authored
* sharded model supports reuse fp16 shard * rename variable * polish code * polish code * polish code
-
- 22 Mar, 2022 4 commits