- 28 Jul, 2023 1 commit
-
-
lvhan028 authored
* bump version to v0.0.2 * fix command * update installation and inference section
-
- 27 Jul, 2023 1 commit
-
-
MaxMatthew authored
-
- 26 Jul, 2023 1 commit
-
-
Chen Xin authored
* defer symlink * fix lint
-
- 25 Jul, 2023 2 commits
-
-
q.yao authored
Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
lvhan028 authored
-
- 24 Jul, 2023 1 commit
-
-
Li Zhang authored
* decode only forward pass * fix lint * batch embedding
-
- 23 Jul, 2023 1 commit
-
-
lvhan028 authored
* refactor model.py and support baichuan-7b * remove model_name * remove hard session_len * export tokenizer.py to target dir * remove model_name from client * remove model_name * update * correct throughput equation * fix session.response * update serving.md * update readme * update according to review comments * update * update * update * update
-
- 22 Jul, 2023 1 commit
-
-
q.yao authored
* add profile throughput benchmark * add output only throughput * update req/min * update benckmark readme * fix lint --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
- 21 Jul, 2023 3 commits
-
-
MaxMatthew authored
* Fix lmdeploy.serve.turbomind bug * add __init__.py for turbomind * add resume function * fix the assignment for session.response * Fix code style
-
Li Zhang authored
* add GQA for llama2 * fix model conversion * fix lint & remove dev log * update news * minor * fix allocation size * fix split_dim for w_qkv.bias
-
Kevin Wang authored
* [Fix] fix issue 127 * 优化防止接口更改 * 如果没有deepspeed用python启动需要手动加载到GPU上 * rollback the changes about max_out_tokens and delelte torch > 2.0 if statement * support kernel injection with customized deepspeed * spelling error * Update chat.py --------- Co-authored-by:wangruohui <12756472+wangruohui@users.noreply.github.com>
-
- 20 Jul, 2023 3 commits
-
-
q.yao authored
* add llama2 template * update readme and fix lint * update readme * add bos * add bos * remove bos * Update model.py --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
WRH authored
-
humu789 authored
* fix get_dataset error * fix lint * add datasets to requirements.txt * update some msci
-
- 19 Jul, 2023 2 commits
- 18 Jul, 2023 3 commits
-
-
AllentDan authored
* update requirements * update transformers version * lint * comments * lint * update requirements * remove setup_requires --------- Co-authored-by:dongchunyu <dongchunyu@pjlab.org.cn>
-
Kevin Wang authored
-
q.yao authored
* wip * profile disable tp * fix profile * lint * fix dlpack * remove comment * add tp flag * add session len check * add eos * remove tp and session len inputs * warp tokenizer * multithread load weight * update profile * refactor tokenizer * remove pre/post process * remove mpi4py requirement * remove * remove bind * remove mpi requirement * check backend_tokenizer
-
- 17 Jul, 2023 1 commit
-
-
Kevin Wang authored
* [Fix] fix attempted_relative_import * use try...except...else
-
- 14 Jul, 2023 3 commits
- 12 Jul, 2023 2 commits
- 11 Jul, 2023 3 commits
-
-
lvhan028 authored
-
tpoisonooo authored
* feat(deploy.py): support w pack qkv
-
WRH authored
* previous merged * add chinese * support torch<2 * add a docstring * fix typo * rename torch submodule * rename to pytorch * rename in readme
-
- 06 Jul, 2023 4 commits
-
-
WRH authored
* draft torch client * deal with space of tokenizer * support tensor parallel * fix * fix * move folder * move instruction to readme * move to torch/ * rename client to chat * very bad response * stash * rename streamer * support internlm * change default args * remove test * improve instructions * remove module docstring * decrease header level of torch model
-
q.yao authored
* streaming-output * fix end * fix profile * support chinese streaming * lint * update chat * lint * fix benchmark --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
tpoisonooo authored
-
pppppM authored
-
- 05 Jul, 2023 5 commits
-
-
lvhan028 authored
* update internlm model * update * update * update * update * update temperature, topk and top_p * update * update * loosen log level
-
tpoisonooo authored
* fix(kv_qparams.py): zp use min * revert(qparams.py): revert format * fix(kv_qparams.py): update formula
-
q.yao authored
-
pppppM authored
* add cal qparams * support offload inference * add collect funtions (mod,weight) * stats kv scales * update init * add user guide * fix hints * fix comments & support turbomind format * update user guide * fix slice kv cache error & support pileval dataset (used in llm-awq) * fix wrong num heads slice * update default dataset * fix conflict * fix hints * fix hints * add gitignore
-
q.yao authored
* wip * wip * example finish * fix include and namespace * wtf * install lib * batchize * update cmake install * multithread * fix comment * fix * add mmengine * bind llamamodel --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
- 04 Jul, 2023 2 commits
- 03 Jul, 2023 1 commit
-
-
lvhan028 authored
-