- 17 Oct, 2023 1 commit
-
-
Lyu Han authored
-
- 16 Oct, 2023 2 commits
- 13 Oct, 2023 3 commits
-
-
del-zhenwu authored
* [doc] Update benchmark command in w4a16.md * Update w4a16.md * Update w4a16.md add pip install nvidia-ml-py * [doc] Update w4a16.md * fix lint error Signed-off-by:
del-zhenwu <dele.zhenwu@gmail.com> * [doc] update model_path & prompt_tokens Signed-off-by:
del-zhenwu <dele.zhenwu@gmail.com> --------- Signed-off-by:
del-zhenwu <dele.zhenwu@gmail.com>
-
Chen Xin authored
* add tp hint for deploy * fix lint * assert tp in turbomind * fix lint
-
YiiSh authored
-
- 12 Oct, 2023 2 commits
- 11 Oct, 2023 3 commits
-
-
akhoroshev authored
-
Shahrukh Khan authored
-
AllentDan authored
* make IPv6 compatible, safe run for coroutine interrupting * instance_id -> session_id and fix api_client.py * update doc * remove useless faq * safe ip mapping * update app.py * remove print * update doc
-
- 09 Oct, 2023 3 commits
- 26 Sep, 2023 7 commits
-
-
Lyu Han authored
-
Lyu Han authored
* Fix memory leak * modern c++
-
AllentDan authored
* fix benchmark serving cannot use Qwen tokenizer * update benchmark readme
-
aisensiy authored
-
akhoroshev authored
-
AllentDan authored
* expose stop words * support string * fix * remove eoa from chatbot * remove eoa of turbomind * fix ut * suffix wheel and fix InternLM no system bug
-
akhoroshev authored
* cuda allocator fix * graceful termination * lint and compilation fix
-
- 25 Sep, 2023 3 commits
-
-
Lyu Han authored
-
Lyu Han authored
Fix side effect brought by supporting codellama: `sequence_start` is always true when calling `model.get_prompt` (#466)
-
Ikko Eltociear Ashimine authored
quantilized -> quantized
-
- 20 Sep, 2023 2 commits
-
-
Lyu Han authored
-
Lyu Han authored
* better profiler * wait for releasing mem * remove fire * remove support for multiple model benchmark * comments * support actual seqlen * change chat template * update * fix ut * int->size_t * output more details * correct tp * rollback * update * update readme * add 'internlm-chat' as the default tag for internlm chat models * rollback tokenizer --------- Co-authored-by:
AllentDan <AllentDan@yeah.net> Co-authored-by:
grimoire <yaoqian@pjlab.org.cn>
-
- 19 Sep, 2023 1 commit
-
-
RunningLeon authored
-
- 18 Sep, 2023 4 commits
-
-
AllentDan authored
* better profiler * wait for releasing mem * remove fire * remove support for multiple model benchmark * comments * output more details * correct tp
-
q.yao authored
* support actual seqlen * fix lint * update variable types * lint * update type * fix lint ---------
-
AllentDan authored
* fix token count bug * fix error response
-
Chen Xin authored
* reduce gil switching * ffi lock func * remove unused * remove unused * remove unused
-
- 14 Sep, 2023 2 commits
- 13 Sep, 2023 2 commits
- 12 Sep, 2023 1 commit
-
-
RunningLeon authored
This reverts commit 7368b88692ecca3f5b39f92a8cc41cf21e3fd71e.
-
- 11 Sep, 2023 3 commits
-
-
Lyu Han authored
-
liukuikun authored
-
Lyu Han authored
* tmp * add demo for codellama inference * update * update * update * update codellama.md * export rope_theta * update * update doc * fix client.py * define SamplingParam * rollback 'end' * rotary_emb_base to rotary_embedding_base * change to baichuan2-7b
-
- 08 Sep, 2023 1 commit
-
-
WRH authored
* support baichuan2-chat * update args from generation config * update deploy.py * update readme * tested with tp * step-1 when last id is eos * add news --------- Co-authored-by:chenxin <chenxin@pjlab.org.cn>
-