1. 23 Jan, 2024 1 commit
    • Daniel Hiltgen's avatar
      Report more information about GPUs in verbose mode · 987c16b2
      Daniel Hiltgen authored
      This adds additional calls to both CUDA and ROCm management libraries to
      discover additional attributes about the GPU(s) detected in the system, and
      wires up runtime verbosity selection.  When users hit problems with GPUs we can
      ask them to run with `OLLAMA_DEBUG=1 ollama serve` and share the results.
      987c16b2
  2. 11 Jan, 2024 1 commit
    • Daniel Hiltgen's avatar
      Support multiple variants for a given llm lib type · 8da7bef0
      Daniel Hiltgen authored
      In some cases we may want multiple variants for a given GPU type or CPU.
      This adds logic to have an optional Variant which we can use to select
      an optimal library, but also allows us to try multiple variants in case
      some fail to load.
      
      This can be useful for scenarios such as ROCm v5 vs v6 incompatibility
      or potentially CPU features.
      8da7bef0
  3. 10 Jan, 2024 1 commit
    • Daniel Hiltgen's avatar
      Harden GPU mgmt library lookup · 3c49c3ab
      Daniel Hiltgen authored
      When there are multiple management libraries installed on a system
      not every one will be compatible with the current driver.  This change
      improves our management library algorithm to build up a set of discovered
      libraries based on glob patterns, and then try all of them until we're able to
      load one without error.
      3c49c3ab
  4. 09 Jan, 2024 1 commit
  5. 03 Jan, 2024 1 commit
  6. 19 Dec, 2023 2 commits