1. 08 Feb, 2024 1 commit
  2. 06 Feb, 2024 1 commit
  3. 02 Feb, 2024 1 commit
  4. 01 Feb, 2024 2 commits
  5. 31 Jan, 2024 1 commit
  6. 29 Jan, 2024 1 commit
  7. 25 Jan, 2024 3 commits
  8. 24 Jan, 2024 1 commit
  9. 23 Jan, 2024 2 commits
  10. 22 Jan, 2024 4 commits
  11. 21 Jan, 2024 3 commits
  12. 20 Jan, 2024 3 commits
  13. 19 Jan, 2024 4 commits
  14. 18 Jan, 2024 1 commit
  15. 17 Jan, 2024 1 commit
  16. 16 Jan, 2024 2 commits
    • Daniel Hiltgen's avatar
      Bump llama.cpp to b1842 and add new cuda lib dep · 795674dd
      Daniel Hiltgen authored
      Upstream llama.cpp has added a new dependency with the
      NVIDIA CUDA Driver Libraries (libcuda.so) which is part of the
      driver distribution, not the general cuda libraries, and is not
      available as an archive, so we can not statically link it.  This may
      introduce some additional compatibility challenges which we'll
      need to keep an eye on.
      795674dd
    • Bruce MacDonald's avatar
      do not cache prompt (#2018) · a897e833
      Bruce MacDonald authored
      - prompt cache causes inferance to hang after some time
      a897e833
  17. 14 Jan, 2024 3 commits
  18. 13 Jan, 2024 3 commits
  19. 12 Jan, 2024 2 commits
  20. 11 Jan, 2024 1 commit
    • Daniel Hiltgen's avatar
      Fix up the CPU fallback selection · 7427fa13
      Daniel Hiltgen authored
      The memory changes and multi-variant change had some merge
      glitches I missed.  This fixes them so we actually get the cpu llm lib
      and best variant for the given system.
      7427fa13