- 15 Jan, 2025 1 commit
-
-
Baber Abbasi authored
* add assistant prefix * add arc_challenge from llama * nit * nit * nit * add assistant prefix * add mmlu_llama * nit * nit * Revert "nit" This reverts commit 6a97f8356237305e375212b966b30e8de59dd4bc. * fix regex bug * add assistant_prefix to vllm * add `Question:` * add mmlu_pro * add fewshot assistant_prefix * use `assistant_prefill` * typehints * nits * nits * add to docs * add readme
-
- 07 Jan, 2025 1 commit
-
-
CL-ModelCloud authored
* hf support load gguf file * code review * code review * code clean up * note about use_fast compat with gguf --------- Co-authored-by:Qubitium-ModelCloud <qubitium@modelcloud.ai>
-
- 25 Dec, 2024 1 commit
-
-
Wang, Yi authored
* fix extra_match low if batch_size > 1 Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> * add sorting to logprobs * nit --------- Signed-off-by:
Wang, Yi A <yi.a.wang@intel.com> Co-authored-by:
Baber <baber@hey.com>
-
- 19 Dec, 2024 1 commit
-
-
Baber Abbasi authored
* add warning for truncation
-
- 16 Dec, 2024 1 commit
-
-
Baber Abbasi authored
* batch all rolling token windows * nit * copy to vllm * fix max_length for `get_rolling_token_windows` * bugfix * bugfix * add type hints
-
- 13 Dec, 2024 1 commit
-
-
Yao Matrix authored
* initial support for optimum-intel ipex model. LM model as first step * format Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * pass dtype Signed-off-by:
Yao Matrix <matrix.yao@intel.com> * update README Signed-off-by:
Yao, Matrix <matrix.yao@intel.com> --------- Signed-off-by:
Yao Matrix <matrix.yao@intel.com>
-
- 09 Dec, 2024 2 commits
-
-
Maanu Grover authored
* update import Signed-off-by:
Maanu Grover <maanug@nvidia.com> * run formatting --------- Signed-off-by:
Maanu Grover <maanug@nvidia.com>
-
Baber Abbasi authored
* left truncate for generate_until * pre-commit
-
- 04 Dec, 2024 1 commit
-
-
Slawomir Strehlke authored
* Handle pipeline_parallel parameter * Add description of pipeline parallelism with OV models
-
- 03 Dec, 2024 1 commit
-
-
Trawinski, Dariusz authored
* avoid timeout errors with high concurrency in api_model * style * add timeout * add docs --------- Co-authored-by:Baber <baber@hey.com>
-
- 01 Dec, 2024 1 commit
-
-
Yoav Katz authored
Update Unitxt task to use locally installed unitxt and not download Unitxt code from Huggingface (#2514) * Moved to require unitxt installation and not download unitxt from HF hub. This has performance benefits and simplifies the code. Signed-off-by:
Yoav Katz <katz@il.ibm.com> * Updated watsonx documentation * Updated installation instructions * Removed redundant comman * Allowed unitxt tasks to generate chat APIs Modified WatsonXI model to support chat apis * Removed print * Run precommit formatting --------- Signed-off-by:
Yoav Katz <katz@il.ibm.com>
-
- 30 Nov, 2024 1 commit
-
-
Baber Abbasi authored
* make utility function to handle `until` * fix text
-
- 22 Nov, 2024 1 commit
-
-
Baber Abbasi authored
-
- 18 Nov, 2024 1 commit
-
-
Baber Abbasi authored
* add hf mamba to mamba_lm * fix _model_generate for hf
-
- 16 Nov, 2024 1 commit
-
-
Baber Abbasi authored
* pre-commit update * update github actions * make logging less verbose * fix artifacts
-
- 15 Nov, 2024 2 commits
-
-
Oyvind Tafjord authored
-
Nikodem Szwast authored
* refactor code, fix config path bug * update types to be from typing lib * add pre-commit formatting * specify version of ibm_watsonx_ai package * adjust get_watsonx_credentials() function, add minor refactor to adress PR review comments * change missing installation hint from ibm_watsonx_ai to lm_eval[ibm_watsonx_ai]
-
- 11 Nov, 2024 2 commits
-
-
Baber Abbasi authored
-
Baber Abbasi authored
* batch commit * :Revert "batch commit" This reverts commit d859d1ca . * batch commit * checkout from main * checkout from main * checkout from main * checkout from main * checkout from main * cleanup * cleanup * cleanup * cleanup * cleanup * cleanup * cleanup * cleanup * cleanup * Chat template fix (#7) * cleanup * cleanup * cleanup * linting * fix tests * add ifeval install to new_task CI * Revert "add ifeval install to new_task CI" This reverts commit 1d19449bb7fbfa05d51e7cd20950475eae533bf1. * adds leaderboard tasks (#1) * adds leaderboard tasks * Delete lm_eval/tasks/leaderboard/leaderboard_chat_template.yaml * add readme * Delete lm_eval/tasks/leaderboard/mmlu_pro/mmlu_pro_chat_template.yaml * modify readme * fix bbh task * fix bbh salient task * modify the readme * Delete lm_eval/tasks/leaderboard/ifeval/README.md * Delete lm_eval/tasks/leaderboard/math/README.md * add leaderboard to the tasks repertory * add anouncment about new leaderbaord tasks * linting * Update README.md Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> * installs ifeval dependency in new_task github workflow --------- Co-authored-by:
Nathan Habib <nathan.habib@huggingface.com> Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> * fix math parser * fix math parser * fix version * add warning about chat template --------- Co-authored-by:
Nathan Habib <nathan.habib@huggingface.co> Co-authored-by:
Nathan Habib <30601243+NathanHB@users.noreply.github.com> Co-authored-by:
Nathan Habib <nathan.habib@huggingface.com> Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> Co-authored-by:
Nathan Habib <nathan.habib19@gmail.com>
-
- 09 Nov, 2024 1 commit
-
-
Baber Abbasi authored
* switch `max_tokens` for `max_completion_tokens`. OpenAI ChatCompletions * remove stop, temp=1 for o1 * add chat assertion * HF_DATASETS_TRUST_REMOTE_CODE = True for task tests * move warning
-
- 07 Nov, 2024 1 commit
-
-
Baber Abbasi authored
* pass device_map other than auto for parallelize
-
- 06 Nov, 2024 1 commit
-
-
Rob Geada authored
-
- 31 Oct, 2024 1 commit
-
-
Qubitium-ModelCloud authored
* support gptqmodel * code opt * add gptqmodel option * Update huggingface.py * Update pyproject.toml * gptqmodel version upgraded to 1.0.6 * GPTQModel version upgraded to 1.0.8 * Update pyproject.toml * fix ruff-format error * add gptqmodel test * Update gptqmodel test model * skip cuda * python3.8 compatible * Update README.md * Update README.md --------- Co-authored-by:CL-ModelCloud <cl@modelcloud.ai>
-
- 30 Oct, 2024 2 commits
-
-
Samuel Monson authored
-
Chris Kerwell Gresla authored
* fix: use lora_request for data parallel vllm evals * fix(docs): include type hint * chore: lint, et pre-commit al --------- Co-authored-by:Chris Kerwell Gresla <chris@wafer.systems>
-
- 25 Oct, 2024 1 commit
-
-
Kiersten Stokes authored
* Update pyproject.toml with watsonx package extra Signed-off-by:
kiersten-stokes <kierstenstokes@gmail.com> * Remove unused function Signed-off-by:
kiersten-stokes <kierstenstokes@gmail.com> --------- Signed-off-by:
kiersten-stokes <kierstenstokes@gmail.com>
-
- 23 Oct, 2024 1 commit
-
-
Nikodem Szwast authored
* add support for IBM watsonx_llm * add ibm_watsonx_ai package to optional-dependencies * move global scope imports to inner scope * change cache to lru_cache * fix circular import * use 3.8 typing * use 3.8 typing --------- Co-authored-by:Baber <baber@hey.com>
-
- 22 Oct, 2024 1 commit
-
-
Leonid Sinev authored
* Replace generic exception classes with a more specific ones * rerun pre-commit to pass linter tests * Revert "rerun pre-commit to pass linter tests" This reverts commit 67f88ccf144469853217704520e613196042d859. * reduce repetitions in errors or so * Replace generic exception class with a more specific one
-
- 08 Oct, 2024 3 commits
-
-
Hailey Schoelkopf authored
-
Baber Abbasi authored
* max_images are passed on to vllms `limit_mm_per_prompt` * replace max image placeholders in string * handle chat_template error * move `fewshot_random_seed` to global
-
Baber Abbasi authored
* switch conditional checks to `self.backend` * nit * nit * commit feedback * fix test; update precommit hooks * add escape hatch for custom self.AUTO_MODEL_CLASS * add escape hatch for custom self.AUTO_MODEL_CLASS * fix * move assertion * add logging messages * update AUTO_MODEL_CLASS behavior in _get_backend --------- Co-authored-by:haileyschoelkopf <hailey@eleuther.ai>
-
- 07 Oct, 2024 1 commit
-
-
Baber Abbasi authored
* tokenizer: trust-remote-code * pre-commit --------- Co-authored-by:Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
-
- 26 Sep, 2024 1 commit
-
-
Baber Abbasi authored
* better error message; fix greedy matching * Update lm_eval/models/openai_completions.py Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> * Update lm_eval/models/openai_completions.py Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> * pre-commit --------- Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
-
- 24 Sep, 2024 1 commit
-
-
Amine Elhattami authored
-
- 18 Sep, 2024 1 commit
-
-
David Corvoysier authored
* feat(neuron): align with latest optimum-neuron * feat(neuron): support pre-exported neuron models * fix(neuron): correctly use max_length * fix(neuron): adapt loglikelihood The evaluation of log likelihood was not working for neuron models using continuous batching, such as all cached neuron LLama models. * refactor(neuron): remove dead code
-
- 13 Sep, 2024 1 commit
-
-
Lintang Sutawika authored
* add WIP hf vlm class * add doc_to_image * add mmmu tasks * fix merge conflicts * add lintang's changes to hf_vlms.py * fix doc_to_image * added yaml_path for config-loading * revert * add line to process str type v * update * modeling cleanup * add aggregation for mmmu * rewrite MMMU processing code based on only MMMU authors' repo (doc_to_image still WIP) * implemented doc_to_image * update doc_to_image to accept list of features * update functions * readd image processed * update args process * bugfix for repeated images fed to model * push WIP loglikelihood code * commit most recent code (generative ; qwen2-vl testing) * preliminary image_token_id handling * small mmmu update: some qs have >4 mcqa options * push updated modeling code * use processor.apply_chat_template * add mathvista draft * nit * nit * ensure no footguns in text<>multimodal LM<>task incompatibility * add notification to readme regarding launch of prototype! * fix compatibility check * reorganize mmmu configs * chat_template=None * add interleave chat_template * add condition * add max_images; interleave=true * nit * testmini_mcq * nit * pass image string; convert img * add vllm * add init * vlm add multi attr * fixup * pass max images to vllm model init * nit * encoding to device * fix HFMultimodalLM.chat_template ? * add mmmu readme * remove erroneous prints * use HFMultimodalLM.chat_template ; restore tasks/__init__.py * add docstring for replace_placeholders in utils * fix `replace_placeholders`; set image_string=None * fix typo * cleanup + fix merge conflicts * update MMMU readme * del mathvista * add some sample scores * Update README.md * add log msg for image_string value --------- Co-authored-by:
haileyschoelkopf <hailey@eleuther.ai> Co-authored-by:
Baber Abbasi <baber@eleuther.ai> Co-authored-by:
Baber <baber@hey.com> Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
-
- 04 Sep, 2024 1 commit
-
-
Baber Abbasi authored
* default chat template method fix * move chat_template to TemplateLM * remove hotfix * handle openai `chat_template` * Update lm_eval/api/model.py Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com> * add 'max_tokens' to gen_kwargs * pre-commit --------- Co-authored-by:
KonradSzafer <szafer.konrad@gmail.com> Co-authored-by:
Hailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
-
- 30 Aug, 2024 2 commits
-
-
Baber Abbasi authored
* max_length - 1 (generation always >= 1) * vllm: fix rolling prefix_token * nit: add comment * fixup! max_length should be handled for logliklihoods * Revert "fixup! max_length should be handled for logliklihoods" This reverts commit 432d1a3b754c117c3a54ea2fe792ab3a1bd09ed3.
-
Baber Abbasi authored
* max_length - 1 (generation always >= 1) * vllm: fix rolling prefix_token * nit: add comment * fixup! max_length should be handled for logliklihoods
-
- 28 Aug, 2024 1 commit
-
-
Hailey Schoelkopf authored
* fix revision type * allow for None-input loglikelihood reqs to be cached * handle no remaining cache items * pre-commit * change cache_hook.add_partial(loglikelihood_rolling...) convention --------- Co-authored-by:Baber Abbasi <baber@eleuther.ai>
-