"vscode:/vscode.git/clone" did not exist on "86cb1e09fe88b1316698dc64727b136f03e75623"
- 10 Aug, 2023 1 commit
-
-
Lyu Han authored
* add release note template * add 'impovement' label to the excluding list
-
- 07 Aug, 2023 5 commits
-
-
WRH authored
* add some dist utils * add model utils * add termio and basicstreamer * typo * fix world size * refactor chat and tested llama1 * add internlm adapter and support stoping criteria * concat with id for internlm * update docstring * update and support llama2 * typo * move docs to docs * update docstring of session manager * update docstring * update docs * fix accel none in model * fix and add test for tensor broadcast * fix session using typing to check type * add docstrings and comprehensive condition test * unit test for dist * fix session * split unittests of utils * typo * update control flow of accel * move test model * remove main in unittest * remove some log * remove some comments
-
lvhan028 authored
-
lvhan028 authored
* add non-stream inference api for chatbot * update according to reviewer's comments
-
LZHgrla authored
* add get_small_sharded_hf.py * fix pre-commit
-
lvhan028 authored
* change to incremental decoding * update
-
- 04 Aug, 2023 1 commit
-
-
AllentDan authored
* use local model for webui * local model for app.py * lint * remove print * add seed * comments * fixed seesion_id * support turbomind batch inference * update app.py * lint and docstring * move webui to serve/gradio * update doc * update doc * update docstring and rmeove print conversition * log * Update docs/zh_cn/build.md Co-authored-by:
Chen Xin <xinchen.tju@gmail.com> * Update docs/en/build.md Co-authored-by:
Chen Xin <xinchen.tju@gmail.com> * use latest gradio * fix * replace partial with InterFace * use host ip instead of coolie --------- Co-authored-by:
Chen Xin <xinchen.tju@gmail.com>
-
- 03 Aug, 2023 3 commits
- 01 Aug, 2023 1 commit
-
-
tpoisonooo authored
-
- 31 Jul, 2023 4 commits
-
-
q.yao authored
* works on interlm and vicuna * support GQA * remove comment * update readme, add logger, default tp=1 * remove log
-
Li Zhang authored
* clean-up * fix lint * fix lint
-
lvhan028 authored
-
del-zhenwu authored
-
- 28 Jul, 2023 1 commit
-
-
lvhan028 authored
* bump version to v0.0.2 * fix command * update installation and inference section
-
- 27 Jul, 2023 4 commits
-
-
Chen Xin authored
* add pypi ci * fix build
-
vansin authored
* Doc: add Twitter link * Doc: add a space
-
MaxMatthew authored
-
Chen Xin authored
* update builder * remove root permission * update readme * update setup.py * add install cuda 12.1 script * use generate.sh * add nccl to install_requires * update README.md * fix lint * update setup.py --------- Co-authored-by:chenxin <chenxin@pjlab.org.cn>
-
- 26 Jul, 2023 3 commits
-
-
Xin Li authored
* translate quantization doc * revise
-
tpoisonooo authored
* Update README_zh-CN.md * Update README.md * Update README_zh-CN.md * Update README.md * Update README_zh-CN.md
-
Chen Xin authored
* defer symlink * fix lint
-
- 25 Jul, 2023 2 commits
-
-
q.yao authored
Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
lvhan028 authored
-
- 24 Jul, 2023 2 commits
- 23 Jul, 2023 1 commit
-
-
lvhan028 authored
* refactor model.py and support baichuan-7b * remove model_name * remove hard session_len * export tokenizer.py to target dir * remove model_name from client * remove model_name * update * correct throughput equation * fix session.response * update serving.md * update readme * update according to review comments * update * update * update * update
-
- 22 Jul, 2023 1 commit
-
-
q.yao authored
* add profile throughput benchmark * add output only throughput * update req/min * update benckmark readme * fix lint --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
- 21 Jul, 2023 4 commits
-
-
MaxMatthew authored
* Fix lmdeploy.serve.turbomind bug * add __init__.py for turbomind * add resume function * fix the assignment for session.response * Fix code style
-
Li Zhang authored
* add GQA for llama2 * fix model conversion * fix lint & remove dev log * update news * minor * fix allocation size * fix split_dim for w_qkv.bias
-
Kevin Wang authored
* [Fix] fix issue 127 * 优化防止接口更改 * 如果没有deepspeed用python启动需要手动加载到GPU上 * rollback the changes about max_out_tokens and delelte torch > 2.0 if statement * support kernel injection with customized deepspeed * spelling error * Update chat.py --------- Co-authored-by:wangruohui <12756472+wangruohui@users.noreply.github.com>
-
RunningLeon authored
* add docker action * update * fix * update
-
- 20 Jul, 2023 3 commits
-
-
q.yao authored
* add llama2 template * update readme and fix lint * update readme * add bos * add bos * remove bos * Update model.py --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
WRH authored
-
humu789 authored
* fix get_dataset error * fix lint * add datasets to requirements.txt * update some msci
-
- 19 Jul, 2023 3 commits
-
-
lvhan028 authored
-
q.yao authored
* remove copy * repetition_penalty=1 * add repetition_penalty to chat args * update readme * update readme
-
rollroll90 authored
-
- 18 Jul, 2023 1 commit
-
-
AllentDan authored
* update requirements * update transformers version * lint * comments * lint * update requirements * remove setup_requires --------- Co-authored-by:dongchunyu <dongchunyu@pjlab.org.cn>
-