- 06 Apr, 2023 1 commit
-
-
ver217 authored
-
- 04 Apr, 2023 1 commit
-
-
ver217 authored
* [zero] refactor low-level zero folder structure * [zero] fix legacy zero import path * [zero] fix legacy zero import path * [zero] remove useless import * [zero] refactor gemini folder structure * [zero] refactor gemini folder structure * [zero] refactor legacy zero import path * [zero] refactor gemini folder structure * [zero] refactor gemini folder structure * [zero] refactor gemini folder structure * [zero] refactor legacy zero import path * [zero] fix test import path * [zero] fix test * [zero] fix circular import * [zero] update import
-
- 28 Jan, 2023 1 commit
-
-
HELSON authored
* [zero] add strict ddp mode for chunk init * [gemini] update gpt example
-
- 09 Jan, 2023 1 commit
-
-
HELSON authored
* [gemini] polish code * [testing] remove code * [gemini] make more robust
-
- 26 Dec, 2022 1 commit
-
-
HELSON authored
* [testing] add beit model * [beit] fix bugs * [beit] fix bugs * [testing] fix bugs
-
- 12 Dec, 2022 1 commit
-
-
Jiarui Fang authored
-
- 09 Dec, 2022 1 commit
-
-
HELSON authored
* [zero] add L2 gradient clipping * [testing] add MlpModel * [zero] add unit test for grad clipping * fix atol
-
- 05 Dec, 2022 1 commit
-
-
Jiarui Fang authored
-
- 30 Nov, 2022 5 commits
-
-
HELSON authored
* [gemini] fix init bugs for modules * fix bugs
-
Jiarui Fang authored
-
Jiarui Fang authored
-
HELSON authored
-
HELSON authored
-
- 29 Nov, 2022 1 commit
-
-
Jiarui Fang authored
* [Gemini] more tests for Gemini * polish code
-
- 24 Nov, 2022 1 commit
-
-
Jiarui Fang authored
-
- 16 Nov, 2022 1 commit
-
-
Jiarui Fang authored
-
- 02 Nov, 2022 1 commit
-
-
HELSON authored
* [hotfix] fix zero's incompatibility with checkpoint in torch-1.12 * [zero] add cpu shard init * [zero] add tiny example test * [colo_tensor] fix bugs for torch-1.11
-
- 18 Oct, 2022 1 commit
-
-
HELSON authored
* add chunk manager init function * fix unit tests * add comment * add flush=True
-
- 14 Oct, 2022 1 commit
-
-
HELSON authored
* fixes memory leak when paramter is in fp16 in ZeroDDP init. * bans chunk releasement in CUDA. Only when a chunk is about to offload, it is allowed to release. * adds a constant placement policy. With it, users can allocate a reserved caching memory space for parameters.
-
- 09 Oct, 2022 1 commit
-
-
HELSON authored
-
- 26 Sep, 2022 1 commit
-
-
Jiarui Fang authored
This reverts commit 5be118f4.
-
- 24 Sep, 2022 1 commit
-
-
HELSON authored
-