1. 13 Sep, 2024 1 commit
    • Lintang Sutawika's avatar
      Multimodal prototyping (#2243) · fb963f0f
      Lintang Sutawika authored
      
      
      * add WIP hf vlm class
      
      * add doc_to_image
      
      * add mmmu tasks
      
      * fix merge conflicts
      
      * add lintang's changes to hf_vlms.py
      
      * fix doc_to_image
      
      * added yaml_path for config-loading
      
      * revert
      
      * add line to process str type v
      
      * update
      
      * modeling cleanup
      
      * add aggregation for mmmu
      
      * rewrite MMMU processing code based on only MMMU authors' repo (doc_to_image still WIP)
      
      * implemented doc_to_image
      
      * update doc_to_image to accept list of features
      
      * update functions
      
      * readd image processed
      
      * update args process
      
      * bugfix for repeated images fed to model
      
      * push WIP loglikelihood code
      
      * commit most recent code (generative ; qwen2-vl testing)
      
      * preliminary image_token_id handling
      
      * small mmmu update: some qs have >4 mcqa options
      
      * push updated modeling code
      
      * use processor.apply_chat_template
      
      * add mathvista draft
      
      * nit
      
      * nit
      
      * ensure no footguns in text<>multimodal LM<>task incompatibility
      
      * add notification to readme regarding launch of prototype!
      
      * fix compatibility check
      
      * reorganize mmmu configs
      
      * chat_template=None
      
      * add interleave chat_template
      
      * add condition
      
      * add max_images; interleave=true
      
      * nit
      
      * testmini_mcq
      
      * nit
      
      * pass image string; convert img
      
      * add vllm
      
      * add init
      
      * vlm add multi attr
      
      * fixup
      
      * pass max images to vllm model init
      
      * nit
      
      * encoding to device
      
      * fix HFMultimodalLM.chat_template ?
      
      * add mmmu readme
      
      * remove erroneous prints
      
      * use HFMultimodalLM.chat_template ; restore tasks/__init__.py
      
      * add docstring for replace_placeholders in utils
      
      * fix `replace_placeholders`; set image_string=None
      
      * fix typo
      
      * cleanup + fix merge conflicts
      
      * update MMMU readme
      
      * del mathvista
      
      * add some sample scores
      
      * Update README.md
      
      * add log msg for image_string value
      
      ---------
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      Co-authored-by: default avatarBaber Abbasi <baber@eleuther.ai>
      Co-authored-by: default avatarBaber <baber@hey.com>
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      fb963f0f
  2. 04 Sep, 2024 1 commit
  3. 30 Aug, 2024 2 commits
    • Baber Abbasi's avatar
      hotfix #2262 (#2264) · 928e8bb6
      Baber Abbasi authored
      * max_length - 1 (generation always >= 1)
      
      * vllm: fix rolling prefix_token
      
      * nit: add comment
      
      * fixup! max_length should be handled for logliklihoods
      
      * Revert "fixup! max_length should be handled for logliklihoods"
      
      This reverts commit 432d1a3b754c117c3a54ea2fe792ab3a1bd09ed3.
      928e8bb6
    • Baber Abbasi's avatar
      API: fix maxlen; vllm: prefix_token_id bug (#2262) · b31f92e8
      Baber Abbasi authored
      * max_length - 1 (generation always >= 1)
      
      * vllm: fix rolling prefix_token
      
      * nit: add comment
      
      * fixup! max_length should be handled for logliklihoods
      b31f92e8
  4. 28 Aug, 2024 1 commit
  5. 22 Aug, 2024 1 commit
  6. 20 Aug, 2024 1 commit
  7. 05 Aug, 2024 3 commits
  8. 01 Aug, 2024 1 commit
  9. 29 Jul, 2024 1 commit
    • Baber Abbasi's avatar
      bugfix and docs for API (#2139) · b70af4f5
      Baber Abbasi authored
      
      
      * encoding bugfix
      
      * encoding bugfix
      
      * overload logliklehood rather than loglikehood_tokens
      
      * add custom tokenizer
      
      * add docs
      
      * Update API_guide.md
      
      fix link; add note
      
      * Update API_guide.md
      
      typo
      
      * pre-commit
      
      * add link in readme
      
      * nit
      
      * nit
      
      * nit
      
      * Update API_guide.md
      
      nits
      
      * Update API_guide.md
      
      * Update API_guide.md
      
      * Update API_guide.md
      
      * Update API_guide.md
      
      * Update README.md
      
      * Update docs/API_guide.md
      
      * Update docs/API_guide.md
      
      * Update API_guide.md
      
      ---------
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      b70af4f5
  10. 22 Jul, 2024 1 commit
    • Baber Abbasi's avatar
      Refactor API models (#2008) · 42dc2448
      Baber Abbasi authored
      
      
      * refactor pad_token handling to fn
      
      * fix docs
      
      * add pad_token_handling to vllm
      
      * start on API superclass
      
      * don't detokenize the returned logits
      
      * streamline vllm tokenizer
      
      * add type hint
      
      * pre-commit
      
      * seems to be in working order
      
      * add model to init
      
      * refactor api models
      
      * nit
      
      * cleanup
      
      * add pbar
      
      * fix type hints
      
      * change optional dependencies
      
      * json encode chat template
      
      * add type hints
      
      * deal with different prompt input requiremnts
      
      * nits
      
      * fix
      
      * cache inside async
      
      * fix
      
      * fix
      
      * nits
      
      * nits
      
      * nits
      
      * nit
      
      * fixup
      
      * fixup
      
      * nit
      
      * add dummy retry
      
      * add dummy retry
      
      * handle imports; skip failing test
      
      * add type hint
      
      * add tests
      
      * add dependency to tests
      
      * add package names to exception
      
      * nit
      
      * docs; type hints
      
      * handle api key
      
      * nit
      
      * tokenizer bug
      
      * fix tokenizer
      
      * nit
      
      * nit
      
      * add better error messages
      
      * nit
      
      * remove decorator
      
      * CI: install api dep
      
      * revert evaluator.py
      
      * consolidate
      
      * consolidate
      
      * nits
      
      * nit
      
      * fix typealias
      
      * nit
      
      * nit
      
      * nit
      
      * Update lm_eval/models/api_models.py
      
      typo
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      
      * Update lm_eval/models/openai_completions.py
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      
      * Update lm_eval/models/anthropic_llms.py
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      
      * Update lm_eval/models/api_models.py
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      
      * fix typo
      
      * add news section
      
      * add info for API
      
      * pre-commit
      
      * typo
      
      * fix bug: unpack logliklehood requests
      
      * fix bug: shared gen_kwargs mutated
      
      * nit: handle copy properly
      
      * Update README.md
      
      * Update README.md
      
      * Update README.md
      
      * Update api_models.py
      
      * Update README.md
      
      ---------
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      42dc2448
  11. 15 Jul, 2024 1 commit
  12. 02 Jul, 2024 1 commit
  13. 28 Jun, 2024 1 commit
  14. 18 Jun, 2024 1 commit
  15. 13 Jun, 2024 2 commits
  16. 12 Jun, 2024 1 commit
  17. 11 Jun, 2024 1 commit
  18. 03 Jun, 2024 1 commit
  19. 30 May, 2024 1 commit
  20. 28 May, 2024 1 commit
  21. 24 May, 2024 1 commit
  22. 23 May, 2024 1 commit
  23. 19 May, 2024 1 commit
  24. 07 May, 2024 2 commits
  25. 05 May, 2024 2 commits
  26. 03 May, 2024 1 commit
    • KonradSzafer's avatar
      evaluation tracker implementation (#1766) · 59cf408a
      KonradSzafer authored
      * evaluation tracker implementation
      
      * OVModelForCausalLM test fix
      
      * typo fix
      
      * moved methods args
      
      * multiple args in one flag
      
      * loggers moved to dedicated dir
      
      * improved filename sanitization
      59cf408a
  27. 02 May, 2024 2 commits
  28. 18 Apr, 2024 1 commit
  29. 16 Apr, 2024 2 commits
  30. 05 Apr, 2024 1 commit
    • Seungwoo Ryu's avatar
      Anthropic Chat API (#1594) · 27924d77
      Seungwoo Ryu authored
      
      
      * claude3
      
      * supply for anthropic claude3
      
      * supply for anthropic claude3
      
      * anthropic config changes
      
      * add callback options on anthropic
      
      * line passed
      
      * claude3 tiny change
      
      * help anthropic installation
      
      * mention sysprompt / being careful with format in readme
      
      ---------
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      27924d77
  31. 01 Apr, 2024 1 commit
    • Michael Goin's avatar
      Fix CLI --batch_size arg for openai-completions/local-completions (#1656) · 9516087b
      Michael Goin authored
      The OpenAI interface supports batch size as an argument to the completions API, but does not seem to support specification of this on the CLI i.e. `lm_eval --model openai-completions --batch_size 16 ...` because of a simple lack of str->int conversion.
      
      This is confirmed by my usage and stacktrace from running `OPENAI_API_KEY=dummy lm_eval --model local-completions --tasks gsm8k --batch_size 16 --model_args model=nm-
      testing/zephyr-beta-7b-gptq-g128,tokenizer_backend=huggingface,base_url=http://localhost:8000/v1`:
      ```
      Traceback (most recent call last):
        File "/home/michael/venv/bin/lm_eval", line 8, in <module>
          sys.exit(cli_evaluate())
        File "/home/michael/code/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate
          results = evaluator.simple_evaluate(
        File "/home/michael/code/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper
          return fn(*args, **kwargs)
        File "/home/michael/code/lm-evaluation-harness/lm_eval/evaluator.py", line 251, in simple_evaluate
          results = evaluate(
        File "/home/michael/code/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper
          return fn(*args, **kwargs)
        File "/home/michael/code/lm-evaluation-harness/lm_eval/evaluator.py", line 390, in evaluate
          resps = getattr(lm, reqtype)(cloned_reqs)
        File "/home/michael/code/lm-evaluation-harness/lm_eval/models/openai_completions.py", line 263, in generate_until
          list(sameuntil_chunks(re_ord.get_reordered(), self.batch_size)),
        File "/home/michael/code/lm-evaluation-harness/lm_eval/models/openai_completions.py", line 251, in sameuntil_chunks
          if len(ret) >= size or x[1] != lastuntil:
      TypeError: '>=' not supported between instances of 'int' and 'str'
      ```
      9516087b
  32. 27 Mar, 2024 1 commit