1. 07 Mar, 2024 2 commits
  2. 21 Jan, 2024 1 commit
  3. 20 Jan, 2024 1 commit
  4. 18 Jan, 2024 3 commits
  5. 11 Jan, 2024 1 commit
    • Daniel Hiltgen's avatar
      Build multiple CPU variants and pick the best · d88c527b
      Daniel Hiltgen authored
      This reduces the built-in linux version to not use any vector extensions
      which enables the resulting builds to run under Rosetta on MacOS in
      Docker.  Then at runtime it checks for the actual CPU vector
      extensions and loads the best CPU library available
      d88c527b
  6. 25 Dec, 2023 1 commit
  7. 22 Dec, 2023 1 commit
  8. 19 Dec, 2023 1 commit
  9. 01 Oct, 2023 1 commit
  10. 20 Sep, 2023 4 commits
  11. 14 Sep, 2023 1 commit
  12. 12 Sep, 2023 1 commit
  13. 30 Aug, 2023 1 commit
    • Bruce MacDonald's avatar
      subprocess llama.cpp server (#401) · 42998d79
      Bruce MacDonald authored
      * remove c code
      * pack llama.cpp
      * use request context for llama_cpp
      * let llama_cpp decide the number of threads to use
      * stop llama runner when app stops
      * remove sample count and duration metrics
      * use go generate to get libraries
      * tmp dir for running llm
      42998d79
  14. 25 Aug, 2023 1 commit
  15. 08 Aug, 2023 1 commit
  16. 24 Jul, 2023 1 commit
  17. 21 Jul, 2023 1 commit
  18. 18 Jul, 2023 1 commit
  19. 07 Jul, 2023 1 commit
  20. 28 Jun, 2023 4 commits
  21. 27 Jun, 2023 6 commits