"thinking/template_test.go" did not exist on "c6a6d7294dd50b9216918fe72fd92bc4ae572ac0"
- 15 Oct, 2025 5 commits
- 14 Oct, 2025 1 commit
-
-
Mac Misiura authored
*
✨ added an approach to use tokenizer_info endpoint from vllm Signed-off-by:m-misiura <mmisiura@redhat.com> *
🚧 removed all auto-detection and tokenization logic from `LocalChatCompletion` * pacify pre-commit --------- Signed-off-by:m-misiura <mmisiura@redhat.com> Co-authored-by:
Baber <baber@hey.com>
-
- 02 Oct, 2025 1 commit
-
-
Vineeth authored
-
- 21 Sep, 2025 1 commit
-
-
kaixuanliu authored
Signed-off-by:Liu, Kaixuan <kaixuan.liu@intel.com>
-
- 12 Sep, 2025 1 commit
-
-
fxmarty-amd authored
-
- 08 Sep, 2025 2 commits
-
-
Slim Frikha authored
* feat(vllm_causallms): make collator ignore seed when splitting batch into chunks * fix(collator): revert PR changes * fix(vllm-causallm): update collator call with groupby None * feat(sglang-causallms): make generation accept a list of sampling params --------- Co-authored-by:Baber <baber@hey.com>
-
Lucia Quirke authored
-
- 27 Aug, 2025 1 commit
-
-
Baber Abbasi authored
-
- 26 Aug, 2025 1 commit
-
-
Janna authored
* add AIME tasks * standardize the repeats * fix task naming * aime25 only has test set * edit readme * add utils * standardize * fix case sensitivity * repeat once * lint * more linting * lint huggingface.py
-
- 25 Aug, 2025 1 commit
-
-
Nikita Savelyev authored
* Add support for OVModelForSeq2SeqLM * Add test
-
- 21 Aug, 2025 2 commits
-
-
Kurt Yang authored
Adding support for OpenAI GPT-5 model; Models only support hardcoded tempeature=1 and stop=None (#3247)
-
Cyrus Leung authored
Signed-off-by:DarkLight1337 <tlleungac@connect.ust.hk>
-
- 13 Aug, 2025 1 commit
-
-
Xinhe Shi authored
-
- 02 Aug, 2025 1 commit
-
-
Cyrus Leung authored
* Update vLLM compatibility Signed-off-by:
DarkLight1337 <tlleungac@connect.ust.hk> * add TokensPrompt to all generate calls --------- Signed-off-by:
DarkLight1337 <tlleungac@connect.ust.hk> Co-authored-by:
Baber <baber@hey.com>
-
- 24 Jul, 2025 2 commits
-
-
Baber Abbasi authored
-
weiliang authored
-
- 23 Jul, 2025 3 commits
-
-
Michael Goin authored
Device has been a deprecated arg for a few releases of vLLM and is now removed in 0.10.0 https://github.com/vllm-project/vllm/pull/21349
-
Baber Abbasi authored
* Fix: pin datasets < 4.0 * fix * update type hints in HF * fix hellaswag path
-
Avelina Asada Hadji-Kyriacou authored
* added support for additional chat template arguments * use `enable_thinking` * add wrap logging function * add `chat_template_args` back to HF --------- Co-authored-by:Baber <baber@hey.com>
-
- 18 Jul, 2025 2 commits
-
-
Ramiro R. C. authored
* added headers and custom model name | fixed bug with trust_remote_code param * linting * removed custom model name | changed headers override * add `header` to base TemplateAPI * nit --------- Co-authored-by:Baber <baber@hey.com>
-
mans authored
* fix request hanging when request api * pre commit --------- Co-authored-by:qinyidao <qinyidao@moonshot.cn>
-
- 16 Jul, 2025 1 commit
-
-
Baber Abbasi authored
* feat: add postprocessing for generated text to strip stop sequences and thinking tokens * nit * fix: trim leading whitespace after stripping thinking tokens from generation * feat: add think_end_token to model_args * nit * nit * nit * add to readme * nit
-
- 15 Jul, 2025 1 commit
-
-
MaYongQing authored
-
- 14 Jul, 2025 1 commit
-
-
Avelina Asada Hadji-Kyriacou authored
-
- 06 Jul, 2025 1 commit
-
-
Baber Abbasi authored
-
- 03 Jul, 2025 1 commit
-
-
Ankush authored
* fix(hf-gguf): skip gguf_file if external tokenizer is provided * docs(readme): add instructions for evaluating GGUF models with Hugging Face backend
-
- 30 Jun, 2025 1 commit
-
-
Baber Abbasi authored
* Try fixing issue 3026 which is caused by the quantization_config argument introduced in Commit 758c5ed8 . The argument is in Dict type, but for a GPTQ quantized model, it has a conflict with the huggingface interface which expects QuantizationConfigMixin type. Current solution is removing quantization_config argument in HFLM._create_model() of lm_eval/models/huggingface.py. Require further modification to restore the functionality provided by the previous commit. * wrap quantization_config in AutoQuantizationConfig * handle quantization config not dict * wrap quantization_config in AutoQuantizationConfig if dict --------- Co-authored-by:
shanhx2000 <hs359@duke.edu>
-
- 25 Jun, 2025 2 commits
-
-
Younes B authored
* add subfolder * lint * change it to empty string * fix typehints --------- Co-authored-by:Baber <baber@hey.com>
-
Baber Abbasi authored
-
- 23 Jun, 2025 1 commit
-
-
NourFahmy authored
* Fix Anthropic API compatibility issues in chat completions solves two important compatibility issues between the LM Eval Harness and Anthropic's API: 1) The type field issue - Anthropic's Messages API doesn't accept the type field that other APIs might expect, that was previously included 2) The stop sequences issue - Anthropic requires stop sequences to contain non-whitespace characters tested with most recent models from anthopic; claude-sonnet-4-0, claude-opus-4-0, resolved my local api errors * pacufy pre-commit * add type --------- Co-authored-by:Baber <baber@hey.com>
-
- 08 Jun, 2025 1 commit
-
-
Baber Abbasi authored
* use all answers * use middle truncation * maybe fix classification score * strip classification preds * [vllm] remove stop tokens post-hoc * strip all preds * pacify pre-commit * start on truncation utility * add to readme * add a footgun doc * fix newline in yaml templates * do not strip code_sim preds! * fix pre-commit config * fix instruction warning * add not to longbench readme
-
- 03 Jun, 2025 1 commit
-
-
Younes B authored
-
- 02 Jun, 2025 1 commit
-
-
Yury Sulsky authored
-
- 26 May, 2025 1 commit
-
-
Baber Abbasi authored
* add data_parallel for V1 * use Process instead of Queue * ray used if V0 DP * better error handling * fix truncation warning comparison
-
- 23 May, 2025 2 commits
-
-
Ameya Godbole authored
* FIX error due to grouping queries with different continuation length Make Collator choose query with the longest continuation as the candidate for generation * use max for key selection * added comments explaining variable cont length (identical ctx+cont[:-1]) --------- Co-authored-by:Baber <baber@hey.com>
-
fxmarty-amd authored
* fix arguments * pacify pre-commit --------- Co-authored-by:Baber <baber@hey.com>
-