1. 04 Feb, 2021 2 commits
    • Leo Gao's avatar
      Massive refactor · 778e0f91
      Leo Gao authored
      - Extract evaluator (still needs work to clean up)
      - Add tests for evaluator
      - Fix all the things that break on the new tests
      - Misc cleanup
      778e0f91
    • Leo Gao's avatar
      Fix lambada · b57d059a
      Leo Gao authored
      b57d059a
  2. 03 Feb, 2021 2 commits
  3. 30 Jan, 2021 1 commit
  4. 29 Jan, 2021 1 commit
  5. 24 Jan, 2021 1 commit
  6. 21 Jan, 2021 1 commit
  7. 16 Jan, 2021 1 commit
  8. 12 Jan, 2021 1 commit
  9. 10 Jan, 2021 1 commit
  10. 09 Jan, 2021 2 commits
  11. 05 Jan, 2021 1 commit
  12. 28 Dec, 2020 2 commits
  13. 30 Nov, 2020 4 commits
    • Leo Gao's avatar
      Update docstring · 75db3899
      Leo Gao authored
      75db3899
    • Leo Gao's avatar
      Remove num_tokens · e3031e84
      Leo Gao authored
      e3031e84
    • Leo Gao's avatar
      Refactor to remove generate and fix some bad tokenization · 90e50b4c
      Leo Gao authored
      In particular, the following assumptions are FALSE in general:
      tokenize(context + continuation) = tokenize(context) + tokenize(continuation)
      len(tokenize(context + continuation)) = len(tokenize(context)) + len(tokenize(continuation))
      tokenize(context + continuation)[:len(tokenize(context))] = tokenize(context)
      
      So we need to tip-toe around the problem by being careful with how we do it.
      
      In particular, using Fast is not just for performance; while behavour of GPT2Tokenizer differs across Transformers 2 and 3, GPT2TokenizerFast doesn't.
      90e50b4c
    • Leo Gao's avatar
      Make fewshot_examples fast · 6de520af
      Leo Gao authored
      6de520af
  14. 06 Oct, 2020 1 commit
  15. 05 Oct, 2020 2 commits
  16. 17 Sep, 2020 1 commit
  17. 14 Sep, 2020 1 commit
  18. 07 Sep, 2020 9 commits
  19. 06 Sep, 2020 1 commit
  20. 28 Aug, 2020 1 commit