"examples/vscode:/vscode.git/clone" did not exist on "a720fa126b61e02b23afcb47c993c440c08aacd8"
  1. 13 May, 2025 1 commit
  2. 09 May, 2025 2 commits
  3. 08 May, 2025 1 commit
    • Graham King's avatar
      feat: Qwen3, Gemma3 and Llama4 support (#1002) · ceaeba3e
      Graham King authored
      . New mistralrs and llamacpp version
      . mistralrs: Handle Gemma 3 and Llama 4 as vision models
      . Update the dynamo-run docs to use Qwen 3
      . Our pre-processor now supports Llama 4's newer multi-modal `config.json`
      . Upgrade minijinja to handle Qwen 3's prompt template
      
      For Llama 4 we'll need to limit the max seq len. vllm says:
      > To serve at least one request with the models's max seq len (10485760), (240.00 GiB KV cache is needed,...
      
      I was able to run Llama 4 with llamacpp and a quantized GGUF, with Dynamo doing the pre-processing.
      ceaeba3e
  4. 07 May, 2025 3 commits
  5. 06 May, 2025 3 commits
  6. 05 May, 2025 1 commit
  7. 29 Apr, 2025 2 commits
  8. 28 Apr, 2025 3 commits
  9. 26 Apr, 2025 2 commits
  10. 25 Apr, 2025 2 commits
  11. 24 Apr, 2025 1 commit
  12. 23 Apr, 2025 2 commits
  13. 22 Apr, 2025 1 commit
  14. 21 Apr, 2025 1 commit
  15. 18 Apr, 2025 4 commits
  16. 15 Apr, 2025 3 commits
  17. 11 Apr, 2025 3 commits
  18. 09 Apr, 2025 3 commits
  19. 08 Apr, 2025 1 commit
  20. 07 Apr, 2025 1 commit