- 15 Dec, 2023 2 commits
-
-
q.yao authored
* Add bf16 template sp * prepare merge * add enable bf * add bf16 decode attention support * fix python lint * fix yapf * fix c format * c format11 * fix cast * fix on sm<80 * fix linux bf162 cast * fix type cast * fix lint * support from hf pretrained * fix pybind * fix converter * add trust remote code * fix comment * fix convert qwen * fix lint * fix baichuan * update weight map
-
Chen Xin authored
* support image_embs input * add some checks * update interactive/config.pbtxt && TurbomindModelConfig * update docstring * refactor * support convert embeddings to bf16 * update interactive/config.pbtxt * embeddings -> input_embeddings * use input_embedding_ranges * remove embedding_begins/ends
-
- 06 Dec, 2023 1 commit
-
-
Lyu Han authored
-
- 04 Dec, 2023 1 commit
-
-
Li Zhang authored
* Unify prefill and decode passes * dynamic split-fuse * refactor * correct input count calculation * remove unused * lint * lint * fix msvc build * fix msvc build * fix msvc build * fix msvc build * fix msvc build * fix msvc build * fix msvc build * fix msvc build * fix msvc build
-
- 28 Nov, 2023 1 commit
-
-
q.yao authored
-
- 20 Nov, 2023 1 commit
-
-
Li Zhang authored
* tmp * update * update * optimize for throughput * update * fix eos * clean up * fix serving * fix indexed copy * minor * minor --------- Co-authored-by:lvhan028 <lvhan_028@163.com>
-
- 10 Nov, 2023 1 commit
-
-
Li Zhang authored
* refresh decoder attention kernel * block-level kv cache * `BlockManager` & `SequenceManager` * update * update * update * update * rename * GQA support * fix context length * GQA dispatch * kv8 * tune * async stream cb * nvtx * config parsing * debug * optimize output cost * split-k decoding * minor * truncate `session_len` by available blocks * minor * license * fix * dispatch `cp.async` * fix linking * fix * fix deadlock * guard input length * correct start offset * fix prefill chunking * fix `cache_block_seq_len` param passing * fix `block_size` fmtstr * fix output tokens * fix batch resizing * fix masking of finished sequences * add debug util * free unused block early * add ntk scaling and logn scaling * cmake flags * fix typo * w4a16 for sm75 * fix msvc build * fix msvc build * fix block verification * fix msvc build * use `std::shuffle` * fix lint * fix lint * fix lint * clear incoming buffer * clear finished requests * fix batch initialization * fix typo * fix typo * fix comparison
-
- 11 Oct, 2023 1 commit
-
-
akhoroshev authored
-
- 26 Sep, 2023 1 commit
-
-
akhoroshev authored
* cuda allocator fix * graceful termination * lint and compilation fix
-
- 18 Sep, 2023 1 commit
-
-
q.yao authored
* support actual seqlen * fix lint * update variable types * lint * update type * fix lint ---------
-
- 24 Aug, 2023 1 commit
-
-
Lyu Han authored
* Pad tok_embedding and output weights to make their shape divisible by TP * update * update * update * update * update llamaBatch
-
- 18 Aug, 2023 1 commit
-
-
Li Zhang authored
* qwen support * dynamic ntk & logn attn * fix ntk & add chat template * fix ntk scaling & stop words * fix lint * add tiktoken to requirements.txt * fix tokenizer, set model format automatically * update model.py * update readme * fix lint
-
- 17 Aug, 2023 1 commit
-
-
Chen Xin authored
* __PRETTY_FUNCTION__ * CASE_K * uint * remove not * HALF_FLT_MAX * struct init * port utils * better build pthread-win32 * port kernels * port utils/gemm_test * hide windows header * port models * port examples && triton_backend && unittests * update build readme * fix lint * fix lint * fix lint * fix lint * fix lint * fix build * fix build * cmake version * fix typos * update ci * port kernels/gemm_s_f16 * update ci * fix ci * use cudaStreamSynchronize instead of volatile check * remove gettimeofday * remove pthread-win32 * remove dirent.h * update pre-commit * update * remove todo * fix include * fix build * fix build * fix build ci * fix github action trigger * update README * fix linux-build ci * remove windows folder * fix lint * update readme
-
- 21 Jul, 2023 1 commit
-
-
Li Zhang authored
* add GQA for llama2 * fix model conversion * fix lint & remove dev log * update news * minor * fix allocation size * fix split_dim for w_qkv.bias
-
- 01 Jul, 2023 3 commits
- 28 Jun, 2023 1 commit
-
-
tpoisonooo authored
* feat(src): add int8 and compile passed * feat(kernels): fix * feat(llama): update kernel * feat(src): add debug * fix(kernel): k_cache use int8_t pointer * style(llama): clean code * feat(deploy.py): revert to enable fmha * style(LlamaV2): clean code * feat(deploy.py): add default quant policy
-
- 20 Jun, 2023 1 commit
-
-
Li Zhang authored
* add ft code * gitignore * fix lint * revert fmha
-