1. 11 Jan, 2024 3 commits
    • Daniel Hiltgen's avatar
      Build multiple CPU variants and pick the best · d88c527b
      Daniel Hiltgen authored
      This reduces the built-in linux version to not use any vector extensions
      which enables the resulting builds to run under Rosetta on MacOS in
      Docker.  Then at runtime it checks for the actual CPU vector
      extensions and loads the best CPU library available
      d88c527b
    • Daniel Hiltgen's avatar
      Support multiple variants for a given llm lib type · 8da7bef0
      Daniel Hiltgen authored
      In some cases we may want multiple variants for a given GPU type or CPU.
      This adds logic to have an optional Variant which we can use to select
      an optimal library, but also allows us to try multiple variants in case
      some fail to load.
      
      This can be useful for scenarios such as ROCm v5 vs v6 incompatibility
      or potentially CPU features.
      8da7bef0
    • Jeffrey Morgan's avatar
      Increase minimum CUDA memory allocation overhead and fix minimum overhead for multi-gpu (#1896) · b24e8d17
      Jeffrey Morgan authored
      * increase minimum cuda overhead and fix minimum overhead for multi-gpu
      
      * fix multi gpu overhead
      
      * limit overhead to 10% of all gpus
      
      * better wording
      
      * allocate fixed amount before layers
      
      * fixed only includes graph alloc
      b24e8d17
  2. 09 Jan, 2024 3 commits
  3. 08 Jan, 2024 1 commit
  4. 04 Jan, 2024 4 commits
  5. 20 Dec, 2023 1 commit
    • Daniel Hiltgen's avatar
      Revamp the dynamic library shim · 7555ea44
      Daniel Hiltgen authored
      This switches the default llama.cpp to be CPU based, and builds the GPU variants
      as dynamically loaded libraries which we can select at runtime.
      
      This also bumps the ROCm library to version 6 given 5.7 builds don't work
      on the latest ROCm library that just shipped.
      7555ea44
  6. 19 Dec, 2023 4 commits
  7. 05 Dec, 2023 3 commits
  8. 04 Dec, 2023 1 commit
    • Bruce MacDonald's avatar
      chat api (#991) · 7a0899d6
      Bruce MacDonald authored
      - update chat docs
      - add messages chat endpoint
      - remove deprecated context and template generate parameters from docs
      - context and template are still supported for the time being and will continue to work as expected
      - add partial response to chat history
      7a0899d6
  9. 20 Nov, 2023 1 commit
  10. 10 Nov, 2023 1 commit
  11. 02 Nov, 2023 1 commit
  12. 19 Oct, 2023 2 commits
  13. 13 Oct, 2023 4 commits
  14. 11 Oct, 2023 1 commit
  15. 05 Oct, 2023 1 commit
  16. 25 Sep, 2023 1 commit
  17. 21 Sep, 2023 1 commit
  18. 12 Sep, 2023 1 commit
  19. 07 Sep, 2023 1 commit
  20. 30 Aug, 2023 1 commit
    • Bruce MacDonald's avatar
      subprocess llama.cpp server (#401) · 42998d79
      Bruce MacDonald authored
      * remove c code
      * pack llama.cpp
      * use request context for llama_cpp
      * let llama_cpp decide the number of threads to use
      * stop llama runner when app stops
      * remove sample count and duration metrics
      * use go generate to get libraries
      * tmp dir for running llm
      42998d79
  21. 26 Aug, 2023 2 commits
  22. 17 Aug, 2023 1 commit
  23. 14 Aug, 2023 1 commit