1. 01 Jul, 2024 2 commits
  2. 29 Jun, 2024 1 commit
  3. 28 Jun, 2024 3 commits
    • Baber Abbasi's avatar
      Add chat template to `vllm` (#2034) · cc2d3463
      Baber Abbasi authored
      
      
      * add chat template
      
      * refactor token padding
      
      * nit
      
      * nit
      
      * check on failing test
      
      * check transformers version
      
      * remove transformers pin
      
      * add ids to test
      
      * nit
      
      * fixup
      
      * fix bos bug
      
      * nit
      
      * fixup! fix bos bug
      
      * increase tolerance for table test
      
      * don't detokenize vllm logprobs
      
      * Update lm_eval/models/utils.py
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      
      * pre-commit run --all-files
      
      ---------
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      cc2d3463
    • Baber Abbasi's avatar
      fix cache (#2037) · e922cceb
      Baber Abbasi authored
      e922cceb
    • Steven Basart's avatar
      Fixes scrolls task bug with few_shot examples (#2003) · 801322e0
      Steven Basart authored
      Bug:
      
      ```
      python -m scripts.write_out --task scrolls_quality --output_base_path ~/workspace/
      Traceback (most recent call last):
        File "<frozen runpy>", line 198, in _run_module_as_main
        File "<frozen runpy>", line 88, in _run_code
        File "/lm-evaluation-harness/scripts/write_out.py", line 92, in <module>
          main()
        File "/lm-evaluation-harness/scripts/write_out.py", line 51, in main
          task_dict = tasks.get_task_dict(task_names, task_manager)
                      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 423, in get_task_dict
          task_name_from_string_dict = task_manager.load_task_or_group(
                                       ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 271, in load_task_or_group
          collections.ChainMap(*map(self._load_individual_task_or_group, task_list))
        File "/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 162, in _load_individual_task_or_group
          return load_task(task_config, task=name_or_config, group=parent_name)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
        File "/lm-evaluation-harness/lm_eval/tasks/__init__.py", line 148, in load_task
          task_object = config["class"]()
                        ^^^^^^^^^^^^^^^^^
        File "/lm-evaluation-harness/lm_eval/tasks/scrolls/task.py", line 120, in __init__
          super().__init__()
        File "/lm-evaluation-harness/lm_eval/api/task.py", line 703, in __init__
          self._config = TaskConfig(**config)
                         ^^^^^^^^^^^^^^^^^^^^
      TypeError: lm_eval.api.task.TaskConfig() argument after ** must be a mapping, not NoneType
      ```
      801322e0
  4. 26 Jun, 2024 1 commit
  5. 25 Jun, 2024 5 commits
  6. 24 Jun, 2024 2 commits
  7. 20 Jun, 2024 1 commit
  8. 19 Jun, 2024 5 commits
  9. 18 Jun, 2024 2 commits
  10. 13 Jun, 2024 4 commits
  11. 12 Jun, 2024 2 commits
  12. 11 Jun, 2024 4 commits
  13. 10 Jun, 2024 1 commit
  14. 09 Jun, 2024 1 commit
  15. 07 Jun, 2024 4 commits
  16. 06 Jun, 2024 2 commits