1. 06 Dec, 2023 1 commit
    • Lyu Han's avatar
      Report the inference benchmark of models with different size (#794) · ebe90bc9
      Lyu Han authored
      * update test scripts for models with different sizes
      
      * update
      
      * only test after tunning gemm
      
      * chmod +x
      
      * fix typo
      
      * benchmark on a100
      
      * fix typo
      
      * fix typo
      
      * per-token latency percentile in profile_throughput
      
      * fix
      
      * fix
      
      * rename
      
      * make the script accept parameters
      
      * minor fix
      
      * indent
      
      * reformat table
      
      * change to 3000
      
      * minor fix
      ebe90bc9
  2. 29 Nov, 2023 1 commit
    • Lyu Han's avatar
      Report first-token-latency and token-latency percentiles (#736) · 5c9e1e28
      Lyu Han authored
      * update profile scripts
      
      * add top_p, top_k and temperature as input arguments
      
      * fix input_ids
      
      * update profile_throughput
      
      * update profile_restful_api
      
      * update profile_serving
      
      * update
      
      * update
      
      * add progress bar
      
      * remove TODO comments
      
      * update
      
      * remove useless profile_* argument
      
      * remove log level
      
      * change concurrency default value to 64
      
      * update restful_api.md
      
      * update according to review comments
      
      * fix docstring
      5c9e1e28
  3. 08 Nov, 2023 1 commit
    • AllentDan's avatar
      fix benchmark serving computation mistake (#630) · 529e56bd
      AllentDan authored
      * fix benchmark serving computation mistake
      
      * fix timestamps computations
      
      * remove speed up
      
      * no mp
      
      * mp seems faster?
      
      * remove
      
      * update
      
      * remove
      
      * fix
      
      * update
      
      * update print log
      
      * typo
      
      * print fist token latency only stream==True
      
      * remove renew_session
      
      * update AsyncEngine
      529e56bd
  4. 01 Nov, 2023 1 commit
    • AllentDan's avatar
      Improve api_server and webui usage (#544) · 373bd013
      AllentDan authored
      * make IPv6 compatible, safe run for coroutine interrupting
      
      * instance_id -> session_id and fix api_client.py
      
      * update doc
      
      * remove useless faq
      
      * safe ip mapping
      
      * update app.py
      
      * WIP completion
      
      * completion
      
      * update doc
      
      * disable interactive mode for /v1/chat/completions
      
      * docstring
      
      * docstring
      
      * refactor gradio
      
      * update gradio
      
      * udpate
      
      * update doc
      
      * rename
      
      * session_id default -1
      
      * missed two files
      
      * add a APIClient
      
      * add chat func for APIClient
      
      * refine
      
      * add concurrent function
      
      * sequence_start, sequence_end --> interactive_mode
      
      * update doc
      
      * comments
      
      * doc
      
      * better text completion
      
      * remove /v1/embeddings
      
      * comments
      
      * deprecate generate and use /v1/interactive/completions
      
      * /v1/interactive/completion -> /v1/chat/interactive
      
      * embeddings
      
      * rename
      
      * remove wrong arg description
      
      * docstring
      
      * fix
      
      * update cli
      
      * update doc
      
      * strict session_len limit condition
      
      * pass model args to api_server
      373bd013
  5. 16 Oct, 2023 1 commit
  6. 11 Oct, 2023 1 commit
  7. 26 Sep, 2023 1 commit
  8. 22 Aug, 2023 1 commit
    • AllentDan's avatar
      Add Restful API (#223) · d5c10e7a
      AllentDan authored
      * add restful api
      
      * refine
      
      * add simple doc
      
      * lint
      
      * add uvicorn requirement
      
      * more args
      
      * add llama2
      
      * docstring
      
      * update doc
      
      * save
      
      * refine
      
      * lint
      
      * better decode
      
      * add v1/embedding
      
      * add GenerateRequest
      
      * add llama2 chat template
      
      * correct profiling
      
      * update documents
      
      * add length judge
      
      * add faq
      
      * update doc and rename req_que to req_queue
      
      * fix md link, use get_logger, fix sequence_end bug
      
      * use another doc link for go to avoid lint error
      
      * add api_client.py
      
      * update doc
      
      * update doc
      
      * update function interface
      
      * update FAQ
      
      * resolve comments
      d5c10e7a
  9. 07 Aug, 2023 1 commit
  10. 31 Jul, 2023 1 commit
  11. 23 Jul, 2023 1 commit
    • lvhan028's avatar
      Refactor the chat template of supported models using factory pattern (#144) · 7b470f07
      lvhan028 authored
      * refactor model.py and support baichuan-7b
      
      * remove model_name
      
      * remove hard session_len
      
      * export tokenizer.py to target dir
      
      * remove model_name from client
      
      * remove model_name
      
      * update
      
      * correct throughput equation
      
      * fix session.response
      
      * update serving.md
      
      * update readme
      
      * update according to review comments
      
      * update
      
      * update
      
      * update
      
      * update
      7b470f07
  12. 19 Jul, 2023 1 commit
  13. 30 Jun, 2023 2 commits
  14. 25 Jun, 2023 1 commit
    • lvhan028's avatar
      Add profile (#15) · 23c05372
      lvhan028 authored
      * remove constraints on model name
      
      * remove duplicate model converter
      
      * add profile
      
      * get eos and bos from server
      
      * update stop_words
      
      * update sequence_length when the last generated token is eos_id
      
      * fix
      
      * fix
      
      * check-in models
      
      * valicate model_name
      
      * make stop_words as property
      
      * debug profiling
      
      * better stats
      
      * fix assistant reponse
      
      * update profile serving
      
      * update
      
      * update
      23c05372