1. 15 Jan, 2025 1 commit
    • Baber Abbasi's avatar
      assistant prefill (#2615) · 703fbffd
      Baber Abbasi authored
      * add assistant prefix
      
      * add arc_challenge from llama
      
      * nit
      
      * nit
      
      * nit
      
      * add assistant prefix
      
      * add mmlu_llama
      
      * nit
      
      * nit
      
      * Revert "nit"
      
      This reverts commit 6a97f8356237305e375212b966b30e8de59dd4bc.
      
      * fix regex bug
      
      * add assistant_prefix to vllm
      
      * add `Question:`
      
      * add mmlu_pro
      
      * add fewshot assistant_prefix
      
      * use `assistant_prefill`
      
      * typehints
      
      * nits
      
      * nits
      
      * add to docs
      
      * add readme
      703fbffd
  2. 07 Jan, 2025 1 commit
  3. 25 Dec, 2024 1 commit
  4. 19 Dec, 2024 1 commit
  5. 16 Dec, 2024 1 commit
  6. 13 Dec, 2024 1 commit
  7. 09 Dec, 2024 2 commits
  8. 04 Dec, 2024 1 commit
  9. 03 Dec, 2024 1 commit
  10. 01 Dec, 2024 1 commit
    • Yoav Katz's avatar
      Update Unitxt task to use locally installed unitxt and not download Unitxt... · 1170ef9e
      Yoav Katz authored
      
      Update Unitxt task to  use locally installed unitxt and not download Unitxt code from Huggingface (#2514)
      
      * Moved to require unitxt installation and not download unitxt from HF hub.
      
      This has performance benefits and simplifies the code.
      Signed-off-by: default avatarYoav Katz <katz@il.ibm.com>
      
      * Updated watsonx documentation
      
      * Updated installation instructions
      
      * Removed redundant comman
      
      * Allowed unitxt tasks to generate chat APIs
      
      Modified WatsonXI model to support chat apis
      
      * Removed print
      
      * Run precommit formatting
      
      ---------
      Signed-off-by: default avatarYoav Katz <katz@il.ibm.com>
      1170ef9e
  11. 30 Nov, 2024 1 commit
  12. 22 Nov, 2024 1 commit
  13. 18 Nov, 2024 1 commit
  14. 16 Nov, 2024 1 commit
  15. 15 Nov, 2024 2 commits
  16. 11 Nov, 2024 2 commits
  17. 09 Nov, 2024 1 commit
  18. 07 Nov, 2024 1 commit
  19. 06 Nov, 2024 1 commit
  20. 31 Oct, 2024 1 commit
    • Qubitium-ModelCloud's avatar
      Add GPTQModel support for evaluating GPTQ models (#2217) · 4f8e479e
      Qubitium-ModelCloud authored
      
      
      * support gptqmodel
      
      * code opt
      
      * add gptqmodel option
      
      * Update huggingface.py
      
      * Update pyproject.toml
      
      * gptqmodel version upgraded to 1.0.6
      
      * GPTQModel version upgraded to 1.0.8
      
      * Update pyproject.toml
      
      * fix ruff-format error
      
      * add gptqmodel test
      
      * Update gptqmodel test model
      
      * skip cuda
      
      * python3.8 compatible
      
      * Update README.md
      
      * Update README.md
      
      ---------
      Co-authored-by: default avatarCL-ModelCloud <cl@modelcloud.ai>
      4f8e479e
  21. 30 Oct, 2024 2 commits
  22. 25 Oct, 2024 1 commit
  23. 23 Oct, 2024 1 commit
    • Nikodem Szwast's avatar
      Support for IBM watsonx_llm (#2397) · 1185e89a
      Nikodem Szwast authored
      
      
      * add support for IBM watsonx_llm
      
      * add ibm_watsonx_ai package to optional-dependencies
      
      * move global scope imports to inner scope
      
      * change cache to lru_cache
      
      * fix circular import
      
      * use 3.8 typing
      
      * use 3.8 typing
      
      ---------
      Co-authored-by: default avatarBaber <baber@hey.com>
      1185e89a
  24. 22 Oct, 2024 1 commit
    • Leonid Sinev's avatar
      [Fix] Replace generic exception classes with a more specific ones (#1989) · d4ae9635
      Leonid Sinev authored
      * Replace generic exception classes with a more specific ones
      
      * rerun pre-commit to pass linter tests
      
      * Revert "rerun pre-commit to pass linter tests"
      
      This reverts commit 67f88ccf144469853217704520e613196042d859.
      
      * reduce repetitions in errors or so
      
      * Replace generic exception class with a more specific one
      d4ae9635
  25. 08 Oct, 2024 3 commits
  26. 07 Oct, 2024 1 commit
  27. 26 Sep, 2024 1 commit
  28. 24 Sep, 2024 1 commit
  29. 18 Sep, 2024 1 commit
    • David Corvoysier's avatar
      Update neuron backend (#2314) · 9a092f37
      David Corvoysier authored
      * feat(neuron): align with latest optimum-neuron
      
      * feat(neuron): support pre-exported neuron models
      
      * fix(neuron): correctly use max_length
      
      * fix(neuron): adapt loglikelihood
      
      The evaluation of log likelihood was not working for neuron models
      using continuous batching, such as all cached neuron LLama models.
      
      * refactor(neuron): remove dead code
      9a092f37
  30. 13 Sep, 2024 1 commit
    • Lintang Sutawika's avatar
      Multimodal prototyping (#2243) · fb963f0f
      Lintang Sutawika authored
      
      
      * add WIP hf vlm class
      
      * add doc_to_image
      
      * add mmmu tasks
      
      * fix merge conflicts
      
      * add lintang's changes to hf_vlms.py
      
      * fix doc_to_image
      
      * added yaml_path for config-loading
      
      * revert
      
      * add line to process str type v
      
      * update
      
      * modeling cleanup
      
      * add aggregation for mmmu
      
      * rewrite MMMU processing code based on only MMMU authors' repo (doc_to_image still WIP)
      
      * implemented doc_to_image
      
      * update doc_to_image to accept list of features
      
      * update functions
      
      * readd image processed
      
      * update args process
      
      * bugfix for repeated images fed to model
      
      * push WIP loglikelihood code
      
      * commit most recent code (generative ; qwen2-vl testing)
      
      * preliminary image_token_id handling
      
      * small mmmu update: some qs have >4 mcqa options
      
      * push updated modeling code
      
      * use processor.apply_chat_template
      
      * add mathvista draft
      
      * nit
      
      * nit
      
      * ensure no footguns in text<>multimodal LM<>task incompatibility
      
      * add notification to readme regarding launch of prototype!
      
      * fix compatibility check
      
      * reorganize mmmu configs
      
      * chat_template=None
      
      * add interleave chat_template
      
      * add condition
      
      * add max_images; interleave=true
      
      * nit
      
      * testmini_mcq
      
      * nit
      
      * pass image string; convert img
      
      * add vllm
      
      * add init
      
      * vlm add multi attr
      
      * fixup
      
      * pass max images to vllm model init
      
      * nit
      
      * encoding to device
      
      * fix HFMultimodalLM.chat_template ?
      
      * add mmmu readme
      
      * remove erroneous prints
      
      * use HFMultimodalLM.chat_template ; restore tasks/__init__.py
      
      * add docstring for replace_placeholders in utils
      
      * fix `replace_placeholders`; set image_string=None
      
      * fix typo
      
      * cleanup + fix merge conflicts
      
      * update MMMU readme
      
      * del mathvista
      
      * add some sample scores
      
      * Update README.md
      
      * add log msg for image_string value
      
      ---------
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      Co-authored-by: default avatarBaber Abbasi <baber@eleuther.ai>
      Co-authored-by: default avatarBaber <baber@hey.com>
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      fb963f0f
  31. 04 Sep, 2024 1 commit
  32. 30 Aug, 2024 2 commits
    • Baber Abbasi's avatar
      hotfix #2262 (#2264) · 928e8bb6
      Baber Abbasi authored
      * max_length - 1 (generation always >= 1)
      
      * vllm: fix rolling prefix_token
      
      * nit: add comment
      
      * fixup! max_length should be handled for logliklihoods
      
      * Revert "fixup! max_length should be handled for logliklihoods"
      
      This reverts commit 432d1a3b754c117c3a54ea2fe792ab3a1bd09ed3.
      928e8bb6
    • Baber Abbasi's avatar
      API: fix maxlen; vllm: prefix_token_id bug (#2262) · b31f92e8
      Baber Abbasi authored
      * max_length - 1 (generation always >= 1)
      
      * vllm: fix rolling prefix_token
      
      * nit: add comment
      
      * fixup! max_length should be handled for logliklihoods
      b31f92e8
  33. 28 Aug, 2024 1 commit