1. 23 Oct, 2024 1 commit
    • Nikodem Szwast's avatar
      Support for IBM watsonx_llm (#2397) · 1185e89a
      Nikodem Szwast authored
      
      
      * add support for IBM watsonx_llm
      
      * add ibm_watsonx_ai package to optional-dependencies
      
      * move global scope imports to inner scope
      
      * change cache to lru_cache
      
      * fix circular import
      
      * use 3.8 typing
      
      * use 3.8 typing
      
      ---------
      Co-authored-by: default avatarBaber <baber@hey.com>
      1185e89a
  2. 22 Oct, 2024 2 commits
  3. 20 Oct, 2024 1 commit
  4. 17 Oct, 2024 2 commits
  5. 16 Oct, 2024 1 commit
  6. 14 Oct, 2024 1 commit
  7. 08 Oct, 2024 4 commits
  8. 07 Oct, 2024 5 commits
  9. 04 Oct, 2024 3 commits
  10. 03 Oct, 2024 2 commits
    • zxcvuser's avatar
      Add new benchmark: Galician bench (#2155) · 0e763862
      zxcvuser authored
      * Add galician_bench
      
      * Update xnli_gl path
      
      * Add flores_gl group
      
      * Update _flores_common_yaml
      
      * Updated some task groupings and readme
      
      ---------
      0e763862
    • zxcvuser's avatar
      Add new benchmark: Spanish bench (#2157) · ea17b98e
      zxcvuser authored
      * Add spanish_bench
      
      * Add flores_es group
      
      * Update _flores_common_yaml
      
      * Delete lm_eval/tasks/spanish_bench/escola.yaml
      
      * Delete escola from spanish_bench.yaml
      
      * Delete escola from README.md
      
      * pre-commit run --all-files
      
      * Updated some task groupings and readme
      
      ---------
      ea17b98e
  11. 30 Sep, 2024 2 commits
  12. 28 Sep, 2024 1 commit
  13. 26 Sep, 2024 9 commits
  14. 24 Sep, 2024 2 commits
  15. 18 Sep, 2024 1 commit
    • David Corvoysier's avatar
      Update neuron backend (#2314) · 9a092f37
      David Corvoysier authored
      * feat(neuron): align with latest optimum-neuron
      
      * feat(neuron): support pre-exported neuron models
      
      * fix(neuron): correctly use max_length
      
      * fix(neuron): adapt loglikelihood
      
      The evaluation of log likelihood was not working for neuron models
      using continuous batching, such as all cached neuron LLama models.
      
      * refactor(neuron): remove dead code
      9a092f37
  16. 17 Sep, 2024 2 commits
    • Baber Abbasi's avatar
      repr bug (#2315) · 88ea85b4
      Baber Abbasi authored
      88ea85b4
    • SYusupov's avatar
      Update README.md (#2297) · a5e0adcb
      SYusupov authored
      * Update README.md
      
      I encounter some Git buffer size limits when trying to download all commits history of the repository, such as:
      ```error: RPC failed; curl 18 transfer closed with outstanding read data remaining
      error: 5815 bytes of body are still expected
      fetch-pack: unexpected disconnect while reading sideband packet
      fatal: early EOF```
      
      therefore the installation is faster and there are not errors when I download only the last version of the repository
      
      * Fix linting issue
      a5e0adcb
  17. 13 Sep, 2024 1 commit
    • Lintang Sutawika's avatar
      Multimodal prototyping (#2243) · fb963f0f
      Lintang Sutawika authored
      
      
      * add WIP hf vlm class
      
      * add doc_to_image
      
      * add mmmu tasks
      
      * fix merge conflicts
      
      * add lintang's changes to hf_vlms.py
      
      * fix doc_to_image
      
      * added yaml_path for config-loading
      
      * revert
      
      * add line to process str type v
      
      * update
      
      * modeling cleanup
      
      * add aggregation for mmmu
      
      * rewrite MMMU processing code based on only MMMU authors' repo (doc_to_image still WIP)
      
      * implemented doc_to_image
      
      * update doc_to_image to accept list of features
      
      * update functions
      
      * readd image processed
      
      * update args process
      
      * bugfix for repeated images fed to model
      
      * push WIP loglikelihood code
      
      * commit most recent code (generative ; qwen2-vl testing)
      
      * preliminary image_token_id handling
      
      * small mmmu update: some qs have >4 mcqa options
      
      * push updated modeling code
      
      * use processor.apply_chat_template
      
      * add mathvista draft
      
      * nit
      
      * nit
      
      * ensure no footguns in text<>multimodal LM<>task incompatibility
      
      * add notification to readme regarding launch of prototype!
      
      * fix compatibility check
      
      * reorganize mmmu configs
      
      * chat_template=None
      
      * add interleave chat_template
      
      * add condition
      
      * add max_images; interleave=true
      
      * nit
      
      * testmini_mcq
      
      * nit
      
      * pass image string; convert img
      
      * add vllm
      
      * add init
      
      * vlm add multi attr
      
      * fixup
      
      * pass max images to vllm model init
      
      * nit
      
      * encoding to device
      
      * fix HFMultimodalLM.chat_template ?
      
      * add mmmu readme
      
      * remove erroneous prints
      
      * use HFMultimodalLM.chat_template ; restore tasks/__init__.py
      
      * add docstring for replace_placeholders in utils
      
      * fix `replace_placeholders`; set image_string=None
      
      * fix typo
      
      * cleanup + fix merge conflicts
      
      * update MMMU readme
      
      * del mathvista
      
      * add some sample scores
      
      * Update README.md
      
      * add log msg for image_string value
      
      ---------
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      Co-authored-by: default avatarBaber Abbasi <baber@eleuther.ai>
      Co-authored-by: default avatarBaber <baber@hey.com>
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      fb963f0f