- 10 Oct, 2023 1 commit
-
-
flybird11111 authored
* [doc]update advanced tutorials, training gpt with hybrid parallelism * [doc]update advanced tutorials, training gpt with hybrid parallelism * update vit tutorials * update vit tutorials * update vit tutorials * update vit tutorials * update en/train_vit_with_hybrid_parallel.py * fix * resolve comments * fix
-
- 07 Oct, 2023 5 commits
-
-
Blagoy Simandoff authored
-
Camille Zhong authored
-
Michelle authored
-
littsk authored
-
Hongxin Liu authored
-
- 06 Oct, 2023 2 commits
- 05 Oct, 2023 2 commits
-
-
shaoyuw authored
-
Zhongkai Zhao authored
-
- 04 Oct, 2023 2 commits
- 02 Oct, 2023 2 commits
-
-
Yuanheng Zhao authored
* fix imports * add ray-serve with Colossal-Infer tp * trivial: send requests script * add README * fix worker port * fix readme * use app builder and autoscaling * trivial: input args * clean code; revise readme * testci (skip example test) * use auto model/tokenizer * revert imports fix (fixed in other PRs)
-
Yuanheng Zhao authored
* add Colossal-Inference serving example w/ TorchServe * add dockerfile * fix dockerfile * fix dockerfile: fix commit hash, install curl * refactor file structure * revise readme * trivial * trivial: dockerfile format * clean dir; revise readme * fix comments: fix imports and configs * fix formats * remove unused requirements
-
- 28 Sep, 2023 2 commits
- 27 Sep, 2023 8 commits
-
-
binmakeswell authored
-
Yuanchen authored
Co-authored-by:Xu Yuanchen <yuanchen.xu00@gmail.com>
-
flybird11111 authored
* [chat] fix gemini strategy * [chat] fix gemini strategy * [chat] fix gemini strategy * [chat] fix gemini strategy * g# This is a combination of 2 commits. [chat] fix gemini strategy fox * [chat] fix gemini strategy update llama2 example [chat] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * [fix] fix gemini strategy * fix * fix * fix * fix * fix * Update train_prompts.py
-
Tong Li authored
-
github-actions[bot] authored
Co-authored-by:github-actions <github-actions@github.com>
-
littsk authored
-
littsk authored
-
Hongxin Liu authored
-
- 26 Sep, 2023 8 commits
-
-
Yan haixu authored
-
Chandler-Bing authored
change filename: pretraining.py -> trainin.py there is no file named pretraing.py. wrong writing
-
Desperado-Jia authored
[doc] Update TODO in README of Colossal-LLaMA-2
-
Tong Li authored
-
Hongxin Liu authored
* [lazy] patch from pretrained * [lazy] fix from pretrained and add tests * [devops] update ci
-
Tong Li authored
-
Baizhou Zhang authored
* support unsharded saving/loading for model * support optimizer unsharded saving * update doc * support unsharded loading for optimizer * small fix
-
Baizhou Zhang authored
* fix example format in docstring * polish shardformer doc
-
- 25 Sep, 2023 2 commits
-
-
flybird11111 authored
* [fix] fix weekly runing example * [fix] fix weekly runing example
-
binmakeswell authored
* [doc] add llama2 domain-specific solution news
-
- 24 Sep, 2023 2 commits
-
-
Yuanchen authored
* Add ColossalEval * Delete evaluate in Chat --------- Co-authored-by:
Xu Yuanchen <yuanchen.xu00@gmail.com> Co-authored-by:
Tong Li <tong.li352711588@gmail.com>
-
Tong Li authored
-
- 22 Sep, 2023 4 commits
-
-
Hongxin Liu authored
* [release] update version * [doc] revert versions
-
Jianghai authored
* add chatglm2 * add * gather needed kernels * fix some bugs * finish context forward * finish context stage * fix * add * pause * add * fix bugs * finish chatglm * fix bug * change some logic * fix bugs * change some logics * add * add * add * fix * fix tests * fix
-
Xu Kai authored
* [gptq] add gptq kernel (#4416) * add gptq * refactor code * fix tests * replace auto-gptq * rname inferance/quant * refactor test * add auto-gptq as an option * reset requirements * change assert and check auto-gptq * add import warnings * change test flash attn version * remove example * change requirements of flash_attn * modify tests * [skip ci] change requirements-test * [gptq] faster gptq cuda kernel (#4494) * [skip ci] add cuda kernels * add license * [skip ci] fix max_input_len * format files & change test size * [skip ci] * [gptq] add gptq tensor parallel (#4538) * add gptq tensor parallel * add gptq tp * delete print * add test gptq check * add test auto gptq check * [gptq] combine gptq and kv cache manager (#4706) * combine gptq and kv cache manager * add init bits * delete useless code * add model path * delete usless print and update test * delete usless import * move option gptq to shard config * change replace linear to shardformer * update bloom policy * delete useless code * fix import bug and delete uselss code * change colossalai/gptq to colossalai/quant/gptq * update import linear for tests * delete useless code and mv gptq_kernel to kernel directory * fix triton kernel * add triton import
-
littsk authored
* Fix the version check bug in colossalai run when generating the cmd. * polish code
-