1. 14 Dec, 2023 1 commit
    • Hailey Schoelkopf's avatar
      Refactor `hf` modeling code (#1096) · e0eda4d3
      Hailey Schoelkopf authored
      * modularize HFLM code
      
      * pass through extra kwargs to AutoModel.from_pretrained call
      
      * remove explicit model_kwargs
      
      * rename gptq -> autogptq
      
      * fix tokenizer pad token errors
      
      * ensure model always respects device_map and autogptq's selected devices
      
      * add a _get_config helper fn
      e0eda4d3
  2. 29 Nov, 2023 1 commit
  3. 26 Nov, 2023 1 commit
  4. 21 Nov, 2023 1 commit
  5. 20 Nov, 2023 1 commit
  6. 17 Nov, 2023 1 commit
  7. 10 Nov, 2023 1 commit
  8. 02 Nov, 2023 1 commit
  9. 01 Nov, 2023 1 commit
  10. 19 Oct, 2023 1 commit
  11. 17 Oct, 2023 1 commit
  12. 13 Oct, 2023 1 commit
  13. 11 Oct, 2023 3 commits
  14. 21 Sep, 2023 1 commit
  15. 13 Sep, 2023 1 commit
  16. 05 Sep, 2023 2 commits
  17. 04 Sep, 2023 2 commits
  18. 26 Aug, 2023 1 commit
  19. 25 Aug, 2023 2 commits
  20. 22 Aug, 2023 1 commit
  21. 11 Aug, 2023 1 commit
  22. 10 Aug, 2023 1 commit
  23. 07 Aug, 2023 1 commit
  24. 04 Aug, 2023 1 commit
  25. 03 Aug, 2023 1 commit
  26. 02 Aug, 2023 1 commit
  27. 27 Jul, 2023 2 commits
  28. 26 Jul, 2023 1 commit
  29. 24 Jul, 2023 1 commit
    • ZZR0's avatar
      Early stop bug of greedy_until (primary_until should be a list of str) · 984f8793
      ZZR0 authored
      I discovered that the accuracy of all models (e.g., llama7b, llama13b, starcoder) in the 'gsm8k-cot' task was 0%. After a thorough investigation, I realized that the generated text for each question was causing an early stop, preventing the 'regex_pattern' from finding any answers. This issue was caused by an incorrect assignment of the 'primary_until' variable in the 'greedy_until' function. Specifically, 'primary_until' should be a list of strings instead of a single string, as the 'stop_sequences' parameter in the 'stop_sequences_criteria' function requires a List[str]. Once I assigned 'primary_until' to '[until[0]]', the accuracy of llama7b in the 'gsm8k-cot' task increased to 1.67%.
      984f8793
  30. 22 Jul, 2023 1 commit
  31. 21 Jul, 2023 1 commit
  32. 17 Jul, 2023 1 commit
  33. 16 Jul, 2023 1 commit
  34. 15 Jul, 2023 1 commit