1. 06 Jan, 2025 1 commit
  2. 04 Jan, 2025 1 commit
  3. 03 Jan, 2025 2 commits
  4. 01 Jan, 2025 1 commit
  5. 29 Dec, 2024 4 commits
  6. 28 Dec, 2024 1 commit
  7. 27 Dec, 2024 2 commits
  8. 25 Dec, 2024 2 commits
  9. 23 Dec, 2024 3 commits
  10. 22 Dec, 2024 1 commit
  11. 20 Dec, 2024 2 commits
  12. 19 Dec, 2024 1 commit
  13. 18 Dec, 2024 1 commit
  14. 17 Dec, 2024 6 commits
    • Jesse Gross's avatar
      llama: Ensure KV cache is fully defragmented. · 08a832b4
      Jesse Gross authored
      Sometimes the KV cache requires defragmentation even without
      triggering the threshold heuristic. In this case, decoding
      will not being able to find a KV cache slot. This is particularly
      difficult for the caller to handle if it happens in between
      ubatches. To avoid this, we should immediately trigger a defrag.
      
      In addition, a heavily fragmented cache can require more than
      max_moves to defragment. Currently, we stop when we hit the limit
      but this can leave a cache that still does not have adequate space
      even after defragmentation is triggered. Instead, we should do
      multiple batches of processing until everything is complete.
      
      Fixes #7949
      08a832b4
    • Blake Mizerany's avatar
      llm: do not error on "null" format (#8139) · 2ddc32d5
      Blake Mizerany authored
      This fixes another regression in the previous commit that fixed other
      known bugs.
      2ddc32d5
    • Jascha Beste's avatar
    • Blake Mizerany's avatar
      llm: do not silently fail for supplied, but invalid formats (#8130) · 87f0a49f
      Blake Mizerany authored
      Changes in #8002 introduced fixes for bugs with mangling JSON Schemas.
      It also fixed a bug where the server would silently fail when clients
      requested invalid formats. It also, unfortunately, introduced a bug
      where the server would reject requests with an empty format, which
      should be allowed.
      
      The change in #8127 updated the code to allow the empty format, but also
      reintroduced the regression where the server would silently fail when
      the format was set, but invalid.
      
      This commit fixes both regressions. The server does not reject the empty
      format, but it does reject invalid formats. It also adds tests to help
      us catch regressions in the future.
      
      Also, the updated code provides a more detailed error message when a
      client sends a non-empty, but invalid format, echoing the invalid format
      in the response.
      
      This commits also takes the opportunity to remove superfluous linter
      checks.
      87f0a49f
    • Jeffrey Morgan's avatar
    • Daniel Hiltgen's avatar
      darwin: restore multiple runners for x86 (#8125) · 8f805dd7
      Daniel Hiltgen authored
      In 0.5.2 we simplified packaging to have avx only for macos x86.  It looks like
      there may still be some non-AVX systems out there, so this puts back the prior
      logic of building no-AVX for the primary binary, and now 2 runners for avx and avx2.
      These will be packaged in the App bundle only, so the stand-alone binary will now be
      without AVX support on macos.  On arm, we'll also see these runners reported
      as available in the log, but they're dormant and will never be used at runtime.
      8f805dd7
  15. 16 Dec, 2024 2 commits
  16. 15 Dec, 2024 1 commit
  17. 14 Dec, 2024 2 commits
  18. 13 Dec, 2024 2 commits
  19. 12 Dec, 2024 2 commits
  20. 11 Dec, 2024 3 commits
    • Blake Mizerany's avatar
      server: more support for mixed-case model names (#8017) · b1fd7fef
      Blake Mizerany authored
      Fixes #7944
      b1fd7fef
    • Daniel Hiltgen's avatar
      ci: fix linux version (#8054) · 36d111e7
      Daniel Hiltgen authored
      Pass through the version override so the makefiles use it
      36d111e7
    • Blake Mizerany's avatar
      llama: preserve field order in user-defined JSON schemas (#8002) · 9039c821
      Blake Mizerany authored
      Previously we decoded and re-encoded JSON schemas during validation,
      which served no purpose since json.RawMessage already validates JSON
      syntax. Worse, the re-encoding lost field ordering from the original
      schema, which affects inference quality during step-by-step reasoning.
      
      While fixing this ordering issue by using json.RawMessage directly,
      testing revealed that schema_to_grammar (from llama.cpp) also fails to
      preserve field order during grammar generation. This appears to be the
      root cause of inference degradation.
      
      This change prevents us from mangling the user's original schema order,
      but we still need to address the ordering issue in schema_to_grammar.
      That will be a separate change.
      
      Updates #7978
      9039c821