1. 08 May, 2025 1 commit
  2. 07 Feb, 2025 1 commit
  3. 23 Jan, 2025 1 commit
  4. 21 Jan, 2025 1 commit
  5. 03 Dec, 2024 1 commit
  6. 02 Dec, 2024 1 commit
  7. 11 Nov, 2024 1 commit
  8. 25 Sep, 2024 1 commit
  9. 10 Sep, 2024 1 commit
  10. 01 Sep, 2024 1 commit
  11. 27 Aug, 2024 1 commit
  12. 27 Jul, 2024 1 commit
  13. 22 Jul, 2024 1 commit
  14. 03 May, 2024 1 commit
  15. 15 Apr, 2024 1 commit
  16. 26 Mar, 2024 1 commit
  17. 25 Mar, 2024 1 commit
  18. 12 Mar, 2024 1 commit
  19. 12 Feb, 2024 1 commit
  20. 09 Feb, 2024 1 commit
  21. 26 Jan, 2024 2 commits
  22. 08 Jan, 2024 1 commit
  23. 22 Dec, 2023 2 commits
  24. 19 Dec, 2023 1 commit
  25. 12 Dec, 2023 1 commit
  26. 11 Dec, 2023 1 commit
  27. 20 Nov, 2023 1 commit
  28. 09 Nov, 2023 1 commit
  29. 16 Oct, 2023 1 commit
  30. 14 Oct, 2023 1 commit
  31. 12 Oct, 2023 1 commit
  32. 02 Oct, 2023 2 commits
  33. 01 Oct, 2023 1 commit
  34. 28 Sep, 2023 1 commit
  35. 27 Sep, 2023 2 commits
  36. 30 Aug, 2023 1 commit
    • Quinn Slack's avatar
      treat stop as stop sequences, not exact tokens (#442) · f4432e1d
      Quinn Slack authored
      The `stop` option to the generate API is a list of sequences that should cause generation to stop. Although these are commonly called "stop tokens", they do not necessarily correspond to LLM tokens (per the LLM's tokenizer). For example, if the caller sends a generate request with `"stop":["\n"]`, then generation should stop on any token containing `\n` (and trim `\n` from the output), not just if the token exactly matches `\n`. If `stop` were interpreted strictly as LLM tokens, then it would require callers of the generate API to know the LLM's tokenizer and enumerate many tokens in the `stop` list.
      
      Fixes https://github.com/jmorganca/ollama/issues/295.
      f4432e1d