- 27 Nov, 2023 1 commit
-
-
zhouxiang authored
-
- 18 Nov, 2023 2 commits
- 15 Nov, 2023 1 commit
-
-
xiabo authored
-
- 14 Nov, 2023 1 commit
-
-
xiabo authored
-
- 11 Oct, 2023 1 commit
-
-
akhoroshev authored
-
- 09 Oct, 2023 1 commit
-
-
Lyu Han authored
* change shared_instances_ from weakptr to sharedptr * update
-
- 26 Sep, 2023 3 commits
-
-
Lyu Han authored
* Fix memory leak * modern c++
-
akhoroshev authored
-
akhoroshev authored
* cuda allocator fix * graceful termination * lint and compilation fix
-
- 18 Sep, 2023 2 commits
- 14 Sep, 2023 1 commit
-
-
Chen Xin authored
-
- 11 Sep, 2023 1 commit
-
-
Lyu Han authored
* tmp * add demo for codellama inference * update * update * update * update codellama.md * export rope_theta * update * update doc * fix client.py * define SamplingParam * rollback 'end' * rotary_emb_base to rotary_embedding_base * change to baichuan2-7b
-
- 07 Sep, 2023 1 commit
-
-
Lyu Han authored
* Fix crash when context window size is large by setting max dynamic smem size * fix linting
-
- 01 Sep, 2023 1 commit
-
-
Chen Xin authored
* pack llama_gemm * update CMakeLists.txt * remove candidate * update MANIFEST.in
-
- 29 Aug, 2023 1 commit
-
-
q.yao authored
* first * fix causal mask * disable flash attention2 on sm70 * fix 2 * update readme * clang-format * disable ft2 on windows * fix lint * fix build * fix build * fix long kv seq * fix lint * sync copy output --------- Co-authored-by:
grimoire <yaoqian@pjlab.org.cn> Co-authored-by:
irexyc <irexyc@gmail.com>
-
- 24 Aug, 2023 1 commit
-
-
Lyu Han authored
* Pad tok_embedding and output weights to make their shape divisible by TP * update * update * update * update * update llamaBatch
-
- 22 Aug, 2023 1 commit
-
-
Li Zhang authored
* disable cache hint for CUDA < 11.4 * fix lint * fix lint * fix cuda-11.3 build
-
- 18 Aug, 2023 1 commit
-
-
Li Zhang authored
* qwen support * dynamic ntk & logn attn * fix ntk & add chat template * fix ntk scaling & stop words * fix lint * add tiktoken to requirements.txt * fix tokenizer, set model format automatically * update model.py * update readme * fix lint
-
- 17 Aug, 2023 2 commits
-
-
Li Zhang authored
-
Chen Xin authored
* __PRETTY_FUNCTION__ * CASE_K * uint * remove not * HALF_FLT_MAX * struct init * port utils * better build pthread-win32 * port kernels * port utils/gemm_test * hide windows header * port models * port examples && triton_backend && unittests * update build readme * fix lint * fix lint * fix lint * fix lint * fix lint * fix build * fix build * cmake version * fix typos * update ci * port kernels/gemm_s_f16 * update ci * fix ci * use cudaStreamSynchronize instead of volatile check * remove gettimeofday * remove pthread-win32 * remove dirent.h * update pre-commit * update * remove todo * fix include * fix build * fix build * fix build ci * fix github action trigger * update README * fix linux-build ci * remove windows folder * fix lint * update readme
-
- 15 Aug, 2023 1 commit
-
-
Chen Xin authored
-
- 14 Aug, 2023 2 commits
-
-
tpoisonooo authored
* feat(quantization): kv cache use asymmetric
-
Li Zhang authored
* add w4a16 * fix `deploy.py` * add doc * add w4a16 kernels * fuse w1/w3 & bugfixes * fix typo * python * guard sm75/80 features * add missing header * refactor * qkvo bias * update cost model * fix lint * update `deploy.py`
-
- 03 Aug, 2023 1 commit
-
-
lvhan028 authored
* fix build tests failure * move src test cases to tests/csrc
-
- 31 Jul, 2023 2 commits
- 27 Jul, 2023 1 commit
-
-
Chen Xin authored
* update builder * remove root permission * update readme * update setup.py * add install cuda 12.1 script * use generate.sh * add nccl to install_requires * update README.md * fix lint * update setup.py --------- Co-authored-by:chenxin <chenxin@pjlab.org.cn>
-
- 25 Jul, 2023 1 commit
-
-
q.yao authored
Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
- 24 Jul, 2023 1 commit
-
-
Li Zhang authored
* decode only forward pass * fix lint * batch embedding
-
- 21 Jul, 2023 1 commit
-
-
Li Zhang authored
* add GQA for llama2 * fix model conversion * fix lint & remove dev log * update news * minor * fix allocation size * fix split_dim for w_qkv.bias
-
- 18 Jul, 2023 1 commit
-
-
q.yao authored
* wip * profile disable tp * fix profile * lint * fix dlpack * remove comment * add tp flag * add session len check * add eos * remove tp and session len inputs * warp tokenizer * multithread load weight * update profile * refactor tokenizer * remove pre/post process * remove mpi4py requirement * remove * remove bind * remove mpi requirement * check backend_tokenizer
-
- 17 Jul, 2023 1 commit
-
-
q.yao authored
* update log info * format cuda utils
-
- 06 Jul, 2023 2 commits
-
-
q.yao authored
* streaming-output * fix end * fix profile * support chinese streaming * lint * update chat * lint * fix benchmark --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
AllentDan authored
-
- 05 Jul, 2023 2 commits
-
-
pppppM authored
* add cal qparams * support offload inference * add collect funtions (mod,weight) * stats kv scales * update init * add user guide * fix hints * fix comments & support turbomind format * update user guide * fix slice kv cache error & support pileval dataset (used in llm-awq) * fix wrong num heads slice * update default dataset * fix conflict * fix hints * fix hints * add gitignore
-
q.yao authored
* wip * wip * example finish * fix include and namespace * wtf * install lib * batchize * update cmake install * multithread * fix comment * fix * add mmengine * bind llamamodel --------- Co-authored-by:grimoire <yaoqian@pjlab.org.cn>
-
- 04 Jul, 2023 1 commit
-
-
AllentDan authored
* format-11.1 * md-link-config
-
- 03 Jul, 2023 1 commit
-
-
lvhan028 authored
-