1. 22 Jan, 2024 3 commits
  2. 21 Jan, 2024 3 commits
  3. 20 Jan, 2024 1 commit
  4. 19 Jan, 2024 4 commits
  5. 18 Jan, 2024 1 commit
  6. 17 Jan, 2024 1 commit
  7. 16 Jan, 2024 2 commits
    • Daniel Hiltgen's avatar
      Bump llama.cpp to b1842 and add new cuda lib dep · 795674dd
      Daniel Hiltgen authored
      Upstream llama.cpp has added a new dependency with the
      NVIDIA CUDA Driver Libraries (libcuda.so) which is part of the
      driver distribution, not the general cuda libraries, and is not
      available as an archive, so we can not statically link it.  This may
      introduce some additional compatibility challenges which we'll
      need to keep an eye on.
      795674dd
    • Bruce MacDonald's avatar
      do not cache prompt (#2018) · a897e833
      Bruce MacDonald authored
      - prompt cache causes inferance to hang after some time
      a897e833
  8. 14 Jan, 2024 3 commits
  9. 13 Jan, 2024 3 commits
  10. 12 Jan, 2024 2 commits
  11. 11 Jan, 2024 9 commits
  12. 10 Jan, 2024 3 commits
  13. 09 Jan, 2024 5 commits