- 22 Jul, 2025 3 commits
- 21 Jul, 2025 18 commits
-
-
Baber authored
feat: implement check_gold_index_error utility and refactor process_results for improved error handling. remove generate_until multiple-choice
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
# Conflicts: # lm_eval/api/filter.py # lm_eval/api/metrics.py # lm_eval/api/task.py # lm_eval/filters/extraction.py
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
Baber authored
-
- 19 Jul, 2025 5 commits
-
-
Avelina Asada Hadji-Kyriacou authored
* Added missing fixture in test_unitxt_tasks.py * pacify pre-commit --------- Co-authored-by:Baber Abbasi <92168766+baberabb@users.noreply.github.com>
-
Baber Abbasi authored
-
James A. Michaelov authored
* add multiblimp * run linter
-
Avelina Asada Hadji-Kyriacou authored
* Update default.yaml
-
Baber authored
-
- 18 Jul, 2025 6 commits
-
-
Baber authored
-
Baber authored
-
Baber authored
# Conflicts: # tests/test_tasks.py
-
Ramiro R. C. authored
* added headers and custom model name | fixed bug with trust_remote_code param * linting * removed custom model name | changed headers override * add `header` to base TemplateAPI * nit --------- Co-authored-by:Baber <baber@hey.com>
-
mans authored
* fix request hanging when request api * pre commit --------- Co-authored-by:qinyidao <qinyidao@moonshot.cn>
-
Idan Tene authored
* Update utils.py
-
- 16 Jul, 2025 2 commits
-
-
philipdoldo authored
* Removed the 'Let''s think step by step.' text from the start of the target entry in each of the samples to prevent this phrase from being repeated twice in the few-shot prompts and to match the behavior from the original bbh repository. Worth noting that this applied to only 26 out of 27 subtasks, the only one it did not apply to is boolean_expressions.yaml. When it comes to boolean_expressions.yaml, in my opinion there is an error in that it doesn't say the 'Remember that (i) ...' text after the final 'A: Let's think step by step.' in the prompt. Models like EleutherAI/gpt-neo-125m seem to always begin answers with this string anyway (copying what was done in the few-shot prompts), but I think it really should've been part of the prompt, much like how 'A: Let's think step by step.' is included in the prompt for all of the cot tasks. However, the original bbh repo also has this issue, so I think it is fine to keep it this way for consistency, but just thought I'd point it out anyway. * feat: remove extra space from answers; add changelog --------- Co-authored-by:Baber <baber@hey.com>
-
Baber Abbasi authored
* feat: add postprocessing for generated text to strip stop sequences and thinking tokens * nit * fix: trim leading whitespace after stripping thinking tokens from generation * feat: add think_end_token to model_args * nit * nit * nit * add to readme * nit
-
- 15 Jul, 2025 1 commit
-
-
MaYongQing authored
-
- 14 Jul, 2025 3 commits
-
-
Ankit Gola authored
-
Avelina Asada Hadji-Kyriacou authored
-
Atou Houdaifa authored
* add egy mmlu hellaswag * add egymmlu egyhellaswag to tasks readme * fix egymmlu config generation * fix _generate_configs formating
-
- 10 Jul, 2025 2 commits
-
-
Baber Abbasi authored
-
Baber Abbasi authored
* check for chat for warning * add test * remove yaml extension from some evalita configs * move unitxt to own test script * fix CI test
-