1. 08 Feb, 2023 2 commits
  2. 07 Feb, 2023 3 commits
  3. 06 Feb, 2023 1 commit
  4. 03 Feb, 2023 2 commits
  5. 02 Feb, 2023 3 commits
  6. 01 Feb, 2023 3 commits
  7. 31 Jan, 2023 8 commits
  8. 30 Jan, 2023 1 commit
  9. 26 Jan, 2023 2 commits
  10. 24 Jan, 2023 1 commit
  11. 23 Jan, 2023 3 commits
  12. 20 Jan, 2023 2 commits
  13. 17 Jan, 2023 2 commits
  14. 05 Jan, 2023 1 commit
  15. 03 Jan, 2023 2 commits
  16. 30 Dec, 2022 2 commits
    • Nick Hill's avatar
      fix(router): Include special tokens when tokenizing (#14) · 3efa5bbb
      Nick Hill authored
      There's currently a discrepancy in the tokenization between the router
      and python server code. The latter includes special tokens but former
      does not.
      
      This results in a token count mismatch for seq2seq models such as mt0
      where the tokenizer emits an EOS token at the end.
      
      This in turn results in some unexpected/incorrect output, in particular
      when batch concatenation is involved, because the python code uses the
      input length passed from the router for each row.
      
      As far as I can tell, it is better to include this token in the encoder
      `input_ids`, so I guess it's best to just adjust on the router side.
      3efa5bbb
    • Nick Hill's avatar
      fix(server): Check for device type correctly when determining initial padding (#16) · 686cc667
      Nick Hill authored
      AFAIK there is no torch device type called "gpu".
      686cc667
  17. 16 Dec, 2022 2 commits