- 24 Sep, 2022 1 commit
-
-
HELSON authored
-
- 23 Sep, 2022 14 commits
-
-
Boyuan Yao authored
* [autoparallel] annotate pofo sequence * [autoparallel] remove unused print * [autoparallel] fix some code
-
Super Daniel authored
* [fx/profiling] provide summary for MetaInfoProp. * [fx/profiler] provide a table of summary. * [fx] optimize table repr.
-
HELSON authored
-
Jiarui Fang authored
-
HELSON authored
-
YuliangLiu0306 authored
* [tensor] use communication autograd func * change all to all comm spec info * rename pattern and distinguish fwd/bwd * polish code
-
YuliangLiu0306 authored
* [autoparallel] add elementwise handler * polish code * polish code * reduce skipped strategies range * polish code
-
YuliangLiu0306 authored
-
YuliangLiu0306 authored
-
YuliangLiu0306 authored
-
YuliangLiu0306 authored
-
Boyuan Yao authored
* [fx] modify offload codegen * [fx] remove repeated hook definitions * [fx] modify offload test
-
YuliangLiu0306 authored
-
Super Daniel authored
* [fx] tuned the meta info and rotor solver. * [fx] remove import. * [fx] remove import. * [fx] remove import. * [fx] tune the meta calculations. * [fx] polish comments. * [fx] remove assertions. * [fx] modify test cases. * [fx] modify test cases. * [fx] optimize import. * [fx
-
- 22 Sep, 2022 2 commits
-
-
HELSON authored
* remove forced FP32 modules * correct no_shard-contexts' positions
-
Jiarui Fang authored
-
- 21 Sep, 2022 1 commit
-
-
Frank Lee authored
-
- 20 Sep, 2022 9 commits
-
-
Kirigaya Kazuto authored
* [pipeline/tuning] improve dispatch performance both time and space cost * [pipeline/converge] add interface for testing convergence * [NFC] polish colossalai/utils/multi_tensor_apply/multi_tensor_apply.py code style * Update PipelineBase.py * [pipeline/chimera] reconstruct PipelineBase and Worker to support more feasible custom schedule | finish Chimera * [pipeline/chimera] test chimera | fix bug of initializing
-
Jiarui Fang authored
-
YuliangLiu0306 authored
-
Frank Lee authored
-
YuliangLiu0306 authored
* [fx] PoC of runtime shape consistency application * polish code
-
YuliangLiu0306 authored
-
Frank Lee authored
-
Boyuan Yao authored
* [fx] add pofo algorithm * [fx] Add pofo solver * [fx] code refactor * [fx] fix test_linearize import
-
github-actions[bot] authored
Co-authored-by:github-actions <github-actions@github.com>
-
- 19 Sep, 2022 2 commits
-
-
Frank Lee authored
-
Kirigaya Kazuto authored
[pipeline/chimera] reconstruct PipelineBase and Worker to support more feasible custom schedule | finish Chimera (#1595) * [pipeline/tuning] improve dispatch performance both time and space cost * [pipeline/converge] add interface for testing convergence * [NFC] polish colossalai/utils/multi_tensor_apply/multi_tensor_apply.py code style * Update PipelineBase.py * [pipeline/chimera] reconstruct PipelineBase and Worker to support more feasible custom schedule | finish Chimera
-
- 16 Sep, 2022 3 commits
-
-
ver217 authored
-
YuliangLiu0306 authored
* [autoparallel] add bcast op handler * polish code * add more BCAST FUNC OP * polish code * add exception handler * polish
-
Frank Lee authored
-
- 15 Sep, 2022 3 commits
-
-
Frank Lee authored
-
Jiarui Fang authored
-
Super Daniel authored
-
- 14 Sep, 2022 5 commits
-
-
Boyuan Yao authored
* [fx] add input activation offload to codegen * [fx] modify unit test * [fx] remove two skips in torch11 * [fx] use all_input_nodes instead of _input_nodes
-
Super Daniel authored
* [fx] add some comment and docstrings. * [fx] add dataflow analysis for an autograd graph. * add intepretation for graph analysis. * [fx] before doing save_tensor_hooks. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] a very accurate version on GPT-2. * [fx] refactor code. * [fx] remove redundant inplace=True. * [fx] refactor code. * [fx] refactor code. * [fx] refactor code. * [fx] dive into backward memory. * [fx] fix variable names in ckpt_solvers and unskip tests. * [fx] commit my changes. * [fx] restore skips. * [fx] restore skips. * [fx] chaange stage into phase. * [fx] chaange stage into phase. * [fx] chaange stage into phase.
-
YuliangLiu0306 authored
* [autoparallel] add reshape handler * polish code
-
github-actions[bot] authored
Co-authored-by:github-actions <github-actions@github.com>
-
Super Daniel authored
* [fx] add some comment and docstrings. * [fx] add dataflow analysis for an autograd graph. * add intepretation for graph analysis. * [fx] before doing save_tensor_hooks. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] provide an accurate estimation of memory except for GPT-2. * [fx] a very accurate version on GPT-2. * [fx] refactor code. * [fx] remove redundant inplace=True. * [fx] refactor code. * [fx] refactor code. * [fx] refactor code. * [fx] dive into backward memory.
-