1. 18 Oct, 2024 1 commit
  2. 10 Oct, 2024 1 commit
    • Jesse Gross's avatar
      cli: Send all images in conversation history · 7fe39025
      Jesse Gross authored
      Currently the CLI only sends images from the most recent image-
      containing message. This prevents doing things like sending
      one message with an image and then a follow message with a
      second image and asking for comparision based on additional
      information not present in any text that was output.
      
      It's possible that some models have a problem with this but the
      CLI is not the right place to do this since any adjustments are
      model-specific and should affect all clients.
      
      Both llava:34b and minicpm-v do reasonable things with multiple
      images in the history.
      7fe39025
  3. 01 Oct, 2024 1 commit
  4. 11 Sep, 2024 2 commits
  5. 05 Sep, 2024 2 commits
  6. 01 Sep, 2024 1 commit
  7. 23 Aug, 2024 1 commit
  8. 21 Aug, 2024 1 commit
  9. 14 Aug, 2024 1 commit
  10. 12 Aug, 2024 1 commit
  11. 02 Aug, 2024 1 commit
  12. 27 Jul, 2024 1 commit
  13. 26 Jul, 2024 3 commits
  14. 23 Jul, 2024 1 commit
  15. 22 Jul, 2024 3 commits
    • Michael Yang's avatar
      bool · 55cd3ddc
      Michael Yang authored
      55cd3ddc
    • Michael Yang's avatar
      host · 4f1afd57
      Michael Yang authored
      4f1afd57
    • Daniel Hiltgen's avatar
      Remove no longer supported max vram var · cc269ba0
      Daniel Hiltgen authored
      The OLLAMA_MAX_VRAM env var was a temporary workaround for OOM
      scenarios.  With Concurrency this was no longer wired up, and the simplistic
      value doesn't map to multi-GPU setups.  Users can still set `num_gpu`
      to limit memory usage to avoid OOM if we get our predictions wrong.
      cc269ba0
  16. 14 Jul, 2024 1 commit
  17. 12 Jul, 2024 2 commits
  18. 28 Jun, 2024 2 commits
  19. 27 Jun, 2024 1 commit
  20. 25 Jun, 2024 1 commit
    • Blake Mizerany's avatar
      cmd: defer stating model info until necessary (#5248) · 2aa91a93
      Blake Mizerany authored
      This commit changes the 'ollama run' command to defer fetching model
      information until it really needs it. That is, when in interactive mode.
      
      It also removes one such case where the model information is fetch in
      duplicate, just before calling generateInteractive and then again, first
      thing, in generateInteractive.
      
      This positively impacts the performance of the command:
      
          ; time ./before run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./before run llama3 'hi'  0.02s user 0.01s system 2% cpu 1.168 total
          ; time ./before run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./before run llama3 'hi'  0.02s user 0.01s system 2% cpu 1.220 total
          ; time ./before run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./before run llama3 'hi'  0.02s user 0.01s system 2% cpu 1.217 total
          ; time ./after run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./after run llama3 'hi'  0.02s user 0.01s system 4% cpu 0.652 total
          ; time ./after run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./after run llama3 'hi'  0.01s user 0.01s system 5% cpu 0.498 total
          ; time ./after run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with or would you like to chat?
      
          ./after run llama3 'hi'  0.01s user 0.01s system 3% cpu 0.479 total
          ; time ./after run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./after run llama3 'hi'  0.02s user 0.01s system 5% cpu 0.507 total
          ; time ./after run llama3 'hi'
          Hi! It's nice to meet you. Is there something I can help you with, or would you like to chat?
      
          ./after run llama3 'hi'  0.02s user 0.01s system 5% cpu 0.507 total
      2aa91a93
  21. 19 Jun, 2024 1 commit
    • royjhan's avatar
      Extend api/show and ollama show to return more model info (#4881) · fedf7163
      royjhan authored
      
      
      * API Show Extended
      
      * Initial Draft of Information
      Co-Authored-By: default avatarPatrick Devine <pdevine@sonic.net>
      
      * Clean Up
      
      * Descriptive arg error messages and other fixes
      
      * Second Draft of Show with Projectors Included
      
      * Remove Chat Template
      
      * Touches
      
      * Prevent wrapping from files
      
      * Verbose functionality
      
      * Docs
      
      * Address Feedback
      
      * Lint
      
      * Resolve Conflicts
      
      * Function Name
      
      * Tests for api/show model info
      
      * Show Test File
      
      * Add Projector Test
      
      * Clean routes
      
      * Projector Check
      
      * Move Show Test
      
      * Touches
      
      * Doc update
      
      ---------
      Co-authored-by: default avatarPatrick Devine <pdevine@sonic.net>
      fedf7163
  22. 12 Jun, 2024 1 commit
  23. 04 Jun, 2024 4 commits
  24. 30 May, 2024 3 commits
  25. 24 May, 2024 1 commit
  26. 21 May, 2024 1 commit
  27. 20 May, 2024 1 commit