1. 15 May, 2025 1 commit
  2. 13 May, 2025 1 commit
  3. 09 May, 2025 2 commits
  4. 08 May, 2025 1 commit
    • Graham King's avatar
      feat: Qwen3, Gemma3 and Llama4 support (#1002) · ceaeba3e
      Graham King authored
      . New mistralrs and llamacpp version
      . mistralrs: Handle Gemma 3 and Llama 4 as vision models
      . Update the dynamo-run docs to use Qwen 3
      . Our pre-processor now supports Llama 4's newer multi-modal `config.json`
      . Upgrade minijinja to handle Qwen 3's prompt template
      
      For Llama 4 we'll need to limit the max seq len. vllm says:
      > To serve at least one request with the models's max seq len (10485760), (240.00 GiB KV cache is needed,...
      
      I was able to run Llama 4 with llamacpp and a quantized GGUF, with Dynamo doing the pre-processing.
      ceaeba3e
  5. 07 May, 2025 3 commits
  6. 06 May, 2025 3 commits
  7. 05 May, 2025 1 commit
  8. 29 Apr, 2025 2 commits
  9. 28 Apr, 2025 3 commits
  10. 26 Apr, 2025 2 commits
  11. 25 Apr, 2025 2 commits
  12. 24 Apr, 2025 1 commit
  13. 23 Apr, 2025 2 commits
  14. 22 Apr, 2025 1 commit
  15. 21 Apr, 2025 1 commit
  16. 18 Apr, 2025 4 commits
  17. 15 Apr, 2025 3 commits
  18. 11 Apr, 2025 3 commits
  19. 09 Apr, 2025 3 commits
  20. 08 Apr, 2025 1 commit