1. 19 May, 2024 1 commit
  2. 14 May, 2024 1 commit
  3. 13 May, 2024 2 commits
  4. 09 May, 2024 1 commit
    • Edd's avatar
      Copal task (#1803) · 1980a13c
      Edd authored
      * add copal
      
      * change name to copal id for clarity and the task name
      
      * remove `copal_id...` to yaml to make it work
      
      * checkmark on README
      
      * change group name to `copal_id`
      1980a13c
  5. 08 May, 2024 2 commits
  6. 07 May, 2024 5 commits
  7. 06 May, 2024 2 commits
  8. 05 May, 2024 4 commits
  9. 03 May, 2024 2 commits
  10. 02 May, 2024 2 commits
  11. 01 May, 2024 4 commits
  12. 26 Apr, 2024 2 commits
  13. 25 Apr, 2024 3 commits
  14. 18 Apr, 2024 2 commits
  15. 16 Apr, 2024 2 commits
  16. 08 Apr, 2024 1 commit
  17. 07 Apr, 2024 1 commit
  18. 05 Apr, 2024 2 commits
    • Seungwoo Ryu's avatar
      Anthropic Chat API (#1594) · 27924d77
      Seungwoo Ryu authored
      
      
      * claude3
      
      * supply for anthropic claude3
      
      * supply for anthropic claude3
      
      * anthropic config changes
      
      * add callback options on anthropic
      
      * line passed
      
      * claude3 tiny change
      
      * help anthropic installation
      
      * mention sysprompt / being careful with format in readme
      
      ---------
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      27924d77
    • ZoneTwelve's avatar
      TMMLU+ implementation (#1394) · 9ae96cdf
      ZoneTwelve authored
      
      
      * implementation of TMMLU+
      
      * implemented: TMMLU+
      
      ****TMMLU+ : large-scale Traditional chinese Massive Multitask language Understanding****
      
      - 4 categories
          - STEM
          - Social Science
          - Humanities
          - Other
      
      The TMMLU+ dataset, encompassing over 67 subjects and 20160 tasks, is six times larger and more balanced than its predecessor, TMMLU, and includes benchmark results from both closed-source and 20 open-weight Chinese large language models with 1.8B to 72B parameters. However, Traditional Chinese variants continue to underperform compared to major Simplified Chinese models.
      
      ```markdown
      Total number of tasks in the 'test' sets: 20160
      Total number of tasks in the 'validation' sets: 2247
      Total number of tasks in the 'train' sets: 335
      ```
      
      * Remove print from __init__.py
      
      There was my mistake in forgetting to remove the debug print from the code.
      
      * update: move TMMLU+ config generation program into default
      
      * fix: we should use training set as few shots example
      
      * update: README for TMMLU+
      
      * update: a small changes of TMMLU+ README file
      
      * pre-commit run thought
      
      * Add README for TMMLU+ dataset
      
      * run precommit
      
      * trigger precommit again
      
      * trigger precommit again
      
      * isort is fussy
      
      * isort is fussy
      
      * format, again
      
      * oops
      
      * oops
      
      ---------
      Co-authored-by: default avatarlintang <lintang@eleuther.ai>
      Co-authored-by: default avatarhaileyschoelkopf <hailey@eleuther.ai>
      9ae96cdf
  19. 04 Apr, 2024 1 commit