1. 22 Nov, 2023 1 commit
    • Chen Xin's avatar
      Support loading hf model directly (#685) · 6b00f623
      Chen Xin authored
      * turbomind support export model params
      
      * fix overflow
      
      * support turbomind.from_pretrained
      
      * fix tp
      
      * support AutoModel
      
      * support load kv qparams
      
      * update auto_awq
      
      * udpate docstring
      
      * export lmdeploy version
      
      * update doc
      
      * remove download_hf_repo
      
      * LmdeployForCausalLM -> LmdeployForCausalLM
      
      * refactor turbomind.py
      
      * update comment
      
      * add bfloat16 convert back
      
      * support gradio run_locl load hf
      
      * support resuful api server load hf
      
      * add docs
      
      * support loading previous quantized model
      
      * adapt pr 690
      
      * udpate docs
      
      * not export turbomind config when quantize a model
      
      * check model_name when can not get it from config.json
      
      * update readme
      
      * remove model_name in auto_awq
      
      * update
      
      * update
      
      * udpate
      
      * fix build
      
      * absolute import
      6b00f623
  2. 10 Nov, 2023 1 commit
    • Li Zhang's avatar
      TurboMind 2 (#590) · ab1767cf
      Li Zhang authored
      * refresh decoder attention kernel
      
      * block-level kv cache
      
      * `BlockManager` & `SequenceManager`
      
      * update
      
      * update
      
      * update
      
      * update
      
      * rename
      
      * GQA support
      
      * fix context length
      
      * GQA dispatch
      
      * kv8
      
      * tune
      
      * async stream cb
      
      * nvtx
      
      * config parsing
      
      * debug
      
      * optimize output cost
      
      * split-k decoding
      
      * minor
      
      * truncate `session_len` by available blocks
      
      * minor
      
      * license
      
      * fix
      
      * dispatch `cp.async`
      
      * fix linking
      
      * fix
      
      * fix deadlock
      
      * guard input length
      
      * correct start offset
      
      * fix prefill chunking
      
      * fix `cache_block_seq_len` param passing
      
      * fix `block_size` fmtstr
      
      * fix output tokens
      
      * fix batch resizing
      
      * fix masking of finished sequences
      
      * add debug util
      
      * free unused block early
      
      * add ntk scaling and logn scaling
      
      * cmake flags
      
      * fix typo
      
      * w4a16 for sm75
      
      * fix msvc build
      
      * fix msvc build
      
      * fix block verification
      
      * fix msvc build
      
      * use `std::shuffle`
      
      * fix lint
      
      * fix lint
      
      * fix lint
      
      * clear incoming buffer
      
      * clear finished requests
      
      * fix batch initialization
      
      * fix typo
      
      * fix typo
      
      * fix comparison
      ab1767cf
  3. 09 Oct, 2023 1 commit
  4. 26 Sep, 2023 1 commit
  5. 18 Sep, 2023 1 commit
  6. 11 Sep, 2023 1 commit
    • Lyu Han's avatar
      Support codellama (#359) · 65c662f9
      Lyu Han authored
      * tmp
      
      * add demo for codellama inference
      
      * update
      
      * update
      
      * update
      
      * update codellama.md
      
      * export rope_theta
      
      * update
      
      * update doc
      
      * fix client.py
      
      * define SamplingParam
      
      * rollback 'end'
      
      * rotary_emb_base to rotary_embedding_base
      
      * change to baichuan2-7b
      65c662f9
  7. 18 Aug, 2023 1 commit
  8. 14 Aug, 2023 1 commit
    • Li Zhang's avatar
      [Feature] Blazing fast W4A16 inference (#202) · c3290cad
      Li Zhang authored
      * add w4a16
      
      * fix `deploy.py`
      
      * add doc
      
      * add w4a16 kernels
      
      * fuse w1/w3 & bugfixes
      
      * fix typo
      
      * python
      
      * guard sm75/80 features
      
      * add missing header
      
      * refactor
      
      * qkvo bias
      
      * update cost model
      
      * fix lint
      
      * update `deploy.py`
      c3290cad
  9. 21 Jul, 2023 1 commit
  10. 01 Jul, 2023 3 commits
  11. 28 Jun, 2023 1 commit
    • tpoisonooo's avatar
      feat(src): add kv cache int8 quantization (#22) · cc93136e
      tpoisonooo authored
      * feat(src): add int8 and compile passed
      
      * feat(kernels): fix
      
      * feat(llama): update kernel
      
      * feat(src): add debug
      
      * fix(kernel): k_cache use int8_t pointer
      
      * style(llama): clean code
      
      * feat(deploy.py): revert to enable fmha
      
      * style(LlamaV2): clean code
      
      * feat(deploy.py): add default quant policy
      cc93136e
  12. 24 Jun, 2023 1 commit
  13. 20 Jun, 2023 1 commit