1. 10 Feb, 2024 2 commits
  2. 07 Feb, 2024 1 commit
  3. 06 Feb, 2024 1 commit
  4. 05 Feb, 2024 1 commit
  5. 01 Feb, 2024 1 commit
  6. 31 Jan, 2024 3 commits
    • Baber Abbasi's avatar
      add bypass metric (#1156) · f8203de1
      Baber Abbasi authored
      * add bypass metric
      
      * fixed `bypass` metric.
      
      * add task attributes if predict_only
      
      * add `predict_only` checks
      
      * add docs
      
      * added `overide_metric`, `override_config` to `Task`
      
      * nits
      
      * nit
      
      * changed --predict_only to generations; nits
      
      * nits
      
      * nits
      
      * change gen_kwargs warning
      
      * add note about `--predict_only` in README.md
      
      * added `predict_only`
      
      * move table to bottom
      
      * nit
      
      * change null aggregation to bypass (conflict)
      
      * bugfix; default `temp=0.0`
      
      * typo
      f8203de1
    • Eugene Cheah's avatar
      Add support for RWKV models with World tokenizer (#1374) · 084b7050
      Eugene Cheah authored
      
      
      * Add support for RWKV models with World tokenizer
      
      The RWKV line of model with the World tokenizer, does not allow the padding token to be configured, and has its value preset as 0
      
      This however fails all the "if set" checks, and would cause the tokenizer to crash.
      
      A tokenizer class name check was added, in addition to a model type check, as there exists RWKV models which uses the neox tokenizers
      
      * Update huggingface.py
      
      Genericized so that this supports any RWKVWorld tokenizer, and added a fall-back for if the HF implementation name changes.
      
      * Comply with formatting guidelines
      
      * fix format
      
      ---------
      Co-authored-by: default avatarStella Biderman <stellabiderman@gmail.com>
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      084b7050
    • Hailey Schoelkopf's avatar
      Fix unintuitive `--gen_kwargs` behavior (#1329) · bd7d265a
      Hailey Schoelkopf authored
      * don't override do_sample if no value for it is passed
      
      * Update gen_kwargs override condition
      
      * Update huggingface.py
      
      * Update huggingface.py
      
      * run linters
      
      * silence an erroneous warning
      bd7d265a
  7. 28 Jan, 2024 1 commit
  8. 26 Jan, 2024 1 commit
    • NoushNabi's avatar
      Add causalLM OpenVino models (#1290) · 97a67d27
      NoushNabi authored
      
      
      * added intel optimum
      
      * added intel optimum in readme
      
      * modified intel optimum
      
      * modified intel optimum
      
      * modified intel optimum
      
      * modified install optimum
      
      * modified path of IR file
      
      * added openvino_device
      
      * added openvino_device2
      
      * changed optimum-causal to openvino-causal
      
      * Update README.md
      
      * Update README.md
      
      * remove `lm_eval.base` import
      
      * update openvino-causal -> openvino ; pass device through super().__init__()
      
      * Update README.md
      
      * Add optimum to tests dependencies
      
      * apply pre-commit
      
      * fix so tests pass
      
      ---------
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      97a67d27
  9. 24 Jan, 2024 1 commit
  10. 23 Jan, 2024 1 commit
  11. 22 Jan, 2024 2 commits
  12. 15 Jan, 2024 1 commit
  13. 11 Jan, 2024 1 commit
  14. 04 Jan, 2024 1 commit
  15. 02 Jan, 2024 2 commits
  16. 27 Dec, 2023 2 commits
  17. 23 Dec, 2023 1 commit
  18. 22 Dec, 2023 3 commits
    • Anjor Kanekar's avatar
    • Hailey Schoelkopf's avatar
      Upstream Mamba Support (`mamba_ssm`) (#1110) · 5503b274
      Hailey Schoelkopf authored
      * modularize HFLM code
      
      * pass through extra kwargs to AutoModel.from_pretrained call
      
      * remove explicit model_kwargs
      
      * rename gptq -> autogptq
      
      * fix tokenizer pad token errors
      
      * ensure model always respects device_map and autogptq's selected devices
      
      * add a _get_config helper fn
      
      * add mambaLMWrapper
      
      * add mamba extra
      
      * add mamba extra
      
      * fix conditional import
      
      * Fix botched merge commit
      
      * Remove beginning-of-file comment for consistency
      
      * Add docstring for mambaLM re: supported kwargs
      
      * Alphabetize extras
      
      * Update extras table
      
      * appease precommit
      
      * run precommit on mamba_lm
      5503b274
    • Zach Schillaci's avatar
      Generic decorator for handling rate limit errors (#1109) · 046ea6e2
      Zach Schillaci authored
      
      
      * Add retry error handler
      
      * fixup! Add retry error handler
      
      * Move to utils.py
      
      * Run isort on utils.py
      
      * Catch multiple exceptions
      
      * Update LMs with exception handler
      
      * Fixes to anthropic retry handler
      
      * fix callback kwarg
      
      * Update textsynth.py
      
      * fix python 3.8 incompatibility
      
      * fix indenterror I introduced
      
      * placate linter?
      
      * Update on_exception_callback kwarg name
      
      * fixup! Merge branch 'main' into add-retry-error-handler
      
      * fixup! fixup! Merge branch 'main' into add-retry-error-handler
      
      * Merge conflicts are fun
      
      * Run pre-commit
      
      ---------
      Co-authored-by: default avatarHailey Schoelkopf <65563625+haileyschoelkopf@users.noreply.github.com>
      046ea6e2
  19. 21 Dec, 2023 2 commits
  20. 20 Dec, 2023 2 commits
    • Vicki Boykis's avatar
      Implementing local OpenAI API-style chat completions on any given inference server (#1174) · fcfc0c60
      Vicki Boykis authored
      * LocalChatCompletionsLM add
      
      * clean up completions class
      
      * clean up completions class
      
      * update tokens
      
      * README
      
      * fix constructor
      
      * eos token
      
      * folding local-chat-completions into OpenAIChatCompletions
      
      * refactoring to include gen_kwargs as passable option
      
      * add todo on chat completion kwarg validation
      
      * Ruff and README fix
      
      * generalize to **kwargs
      
      * remove unnecessary kwargs
      
      * README and remove kwargs
      
      * README
      fcfc0c60
    • Baber Abbasi's avatar
      Switch Linting to `ruff` (#1166) · 65b8761d
      Baber Abbasi authored
      * add ruff and isort. remove black and flake8
      
      * remove unnecessary dependencies
      
      * remove dependency from table
      
      * change order
      
      * ran ruff
      
      * check 3.9
      
      * exclude evaluator
      
      * update CI workflow
      
      * use ruff config in pyproject.toml
      
      * test
      
      * add isort rules to ruff
      
      * sort imports
      
      * import `make_table`
      
      * try stages for no-commit-to-branch
      
      * turn on mypy for pre-commit
      
      * test
      
      * test
      
      * test
      
      * change no-commit-to-branch to default
      
      * nits
      
      * fixed dependency
      65b8761d
  21. 19 Dec, 2023 2 commits
  22. 16 Dec, 2023 2 commits
  23. 15 Dec, 2023 2 commits
  24. 14 Dec, 2023 3 commits
  25. 12 Dec, 2023 1 commit