1. 08 May, 2025 1 commit
    • Graham King's avatar
      feat: Qwen3, Gemma3 and Llama4 support (#1002) · ceaeba3e
      Graham King authored
      . New mistralrs and llamacpp version
      . mistralrs: Handle Gemma 3 and Llama 4 as vision models
      . Update the dynamo-run docs to use Qwen 3
      . Our pre-processor now supports Llama 4's newer multi-modal `config.json`
      . Upgrade minijinja to handle Qwen 3's prompt template
      
      For Llama 4 we'll need to limit the max seq len. vllm says:
      > To serve at least one request with the models's max seq len (10485760), (240.00 GiB KV cache is needed,...
      
      I was able to run Llama 4 with llamacpp and a quantized GGUF, with Dynamo doing the pre-processing.
      ceaeba3e
  2. 03 Apr, 2025 1 commit
  3. 31 Mar, 2025 1 commit
  4. 19 Mar, 2025 1 commit
    • Graham King's avatar
      chore: Don't depend on openssl (#292) · 7c3fd5c9
      Graham King authored
      This makes the Rust parts all use ring / rustls library instead of local install of openssl. It's a step on the journey to being statically linked.
      
      Pieces:
      - `tokenizers` and `mistralrs` now support rustls (mistralrs by default, tokenizers with feature flag).
      - Move shared dependencies up into workspace
      - New `rand` crate has some renames for future rust
      - Ensure the dependency doesn't creep back in by enforcing it with cargo deny.
      7c3fd5c9
  5. 13 Mar, 2025 1 commit
  6. 12 Mar, 2025 1 commit
  7. 11 Mar, 2025 1 commit
  8. 08 Mar, 2025 1 commit
  9. 07 Mar, 2025 2 commits
  10. 06 Mar, 2025 1 commit
  11. 05 Mar, 2025 1 commit
  12. 28 Feb, 2025 1 commit
  13. 26 Feb, 2025 2 commits
  14. 25 Feb, 2025 1 commit
  15. 21 Feb, 2025 1 commit
  16. 13 Feb, 2025 1 commit
  17. 12 Feb, 2025 1 commit
  18. 10 Feb, 2025 1 commit
  19. 06 Feb, 2025 1 commit
  20. 05 Feb, 2025 2 commits