Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
0d186f3b
Unverified
Commit
0d186f3b
authored
Sep 21, 2023
by
Michael Yang
Committed by
GitHub
Sep 21, 2023
Browse files
Merge pull request #569 from jmorganca/mxyng/update-submodules
silence warm up log
parents
ee1c994d
058d0cd0
Changes
6
Show whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
30 additions
and
0 deletions
+30
-0
.gitmodules
.gitmodules
+1
-0
llm/llama.cpp/generate_darwin_amd64.go
llm/llama.cpp/generate_darwin_amd64.go
+1
-0
llm/llama.cpp/generate_darwin_arm64.go
llm/llama.cpp/generate_darwin_arm64.go
+1
-0
llm/llama.cpp/generate_linux.go
llm/llama.cpp/generate_linux.go
+1
-0
llm/llama.cpp/generate_windows.go
llm/llama.cpp/generate_windows.go
+1
-0
llm/llama.cpp/patches/0001-remove-warm-up-logging.patch
llm/llama.cpp/patches/0001-remove-warm-up-logging.patch
+25
-0
No files found.
.gitmodules
View file @
0d186f3b
...
@@ -6,4 +6,5 @@
...
@@ -6,4 +6,5 @@
[submodule "llm/llama.cpp/gguf"]
[submodule "llm/llama.cpp/gguf"]
path = llm/llama.cpp/gguf
path = llm/llama.cpp/gguf
url = https://github.com/ggerganov/llama.cpp.git
url = https://github.com/ggerganov/llama.cpp.git
ignore = dirty
shallow = true
shallow = true
llm/llama.cpp/generate_darwin_amd64.go
View file @
0d186f3b
...
@@ -11,5 +11,6 @@ package llm
...
@@ -11,5 +11,6 @@ package llm
//go:generate cmake --build ggml/build/cpu --target server --config Release
//go:generate cmake --build ggml/build/cpu --target server --config Release
//go:generate git submodule update --force gguf
//go:generate git submodule update --force gguf
//go:generate git -C gguf apply ../patches/0001-remove-warm-up-logging.patch
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_ACCELERATE=on -DLLAMA_K_QUANTS=on -DCMAKE_SYSTEM_PROCESSOR=x86_64 -DCMAKE_OSX_ARCHITECTURES=x86_64 -DCMAKE_OSX_DEPLOYMENT_TARGET=11.0
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_ACCELERATE=on -DLLAMA_K_QUANTS=on -DCMAKE_SYSTEM_PROCESSOR=x86_64 -DCMAKE_OSX_ARCHITECTURES=x86_64 -DCMAKE_OSX_DEPLOYMENT_TARGET=11.0
//go:generate cmake --build gguf/build/cpu --target server --config Release
//go:generate cmake --build gguf/build/cpu --target server --config Release
llm/llama.cpp/generate_darwin_arm64.go
View file @
0d186f3b
...
@@ -11,5 +11,6 @@ package llm
...
@@ -11,5 +11,6 @@ package llm
//go:generate cmake --build ggml/build/metal --target server --config Release
//go:generate cmake --build ggml/build/metal --target server --config Release
//go:generate git submodule update --force gguf
//go:generate git submodule update --force gguf
//go:generate git -C gguf apply ../patches/0001-remove-warm-up-logging.patch
//go:generate cmake -S gguf -B gguf/build/metal -DLLAMA_METAL=on -DLLAMA_ACCELERATE=on -DLLAMA_K_QUANTS=on -DCMAKE_SYSTEM_PROCESSOR=arm64 -DCMAKE_OSX_ARCHITECTURES=arm64 -DCMAKE_OSX_DEPLOYMENT_TARGET=11.0
//go:generate cmake -S gguf -B gguf/build/metal -DLLAMA_METAL=on -DLLAMA_ACCELERATE=on -DLLAMA_K_QUANTS=on -DCMAKE_SYSTEM_PROCESSOR=arm64 -DCMAKE_OSX_ARCHITECTURES=arm64 -DCMAKE_OSX_DEPLOYMENT_TARGET=11.0
//go:generate cmake --build gguf/build/metal --target server --config Release
//go:generate cmake --build gguf/build/metal --target server --config Release
llm/llama.cpp/generate_linux.go
View file @
0d186f3b
...
@@ -12,6 +12,7 @@ package llm
...
@@ -12,6 +12,7 @@ package llm
//go:generate git submodule update --force gguf
//go:generate git submodule update --force gguf
//go:generate git -C gguf apply ../patches/0001-copy-cuda-runtime-libraries.patch
//go:generate git -C gguf apply ../patches/0001-copy-cuda-runtime-libraries.patch
//go:generate git -C gguf apply ../patches/0001-remove-warm-up-logging.patch
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_K_QUANTS=on
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_K_QUANTS=on
//go:generate cmake --build gguf/build/cpu --target server --config Release
//go:generate cmake --build gguf/build/cpu --target server --config Release
...
...
llm/llama.cpp/generate_windows.go
View file @
0d186f3b
...
@@ -9,5 +9,6 @@ package llm
...
@@ -9,5 +9,6 @@ package llm
//go:generate cmake --build ggml/build/cpu --target server --config Release
//go:generate cmake --build ggml/build/cpu --target server --config Release
//go:generate git submodule update --force gguf
//go:generate git submodule update --force gguf
//go:generate git -C gguf apply ../patches/0001-remove-warm-up-logging.patch
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_K_QUANTS=on
//go:generate cmake -S gguf -B gguf/build/cpu -DLLAMA_K_QUANTS=on
//go:generate cmake --build gguf/build/cpu --target server --config Release
//go:generate cmake --build gguf/build/cpu --target server --config Release
llm/llama.cpp/patches/0001-remove-warm-up-logging.patch
0 → 100644
View file @
0d186f3b
From 07993bdc35345b67b27aa649a7c099ad42d80c4c Mon Sep 17 00:00:00 2001
From: Michael Yang <mxyng@pm.me>
Date: Thu, 21 Sep 2023 14:43:21 -0700
Subject: [PATCH] remove warm up logging
---
common/common.cpp | 2 --
1 file changed, 2 deletions(-)
diff --git a/common/common.cpp b/common/common.cpp
index 2597ba0..b56549b 100644
--- a/common/common.cpp
+++ b/common/common.cpp
@@ -780,8 +780,6 @@
std::tuple<struct llama_model *, struct llama_context *> llama_init_from_gpt_par
}
{
- LOG("warming up the model with an empty run\n");
-
const std::vector<llama_token> tmp = { llama_token_bos(lctx), llama_token_eos(lctx), };
llama_eval(lctx, tmp.data(), std::min(tmp.size(), (size_t) params.n_batch), 0, params.n_threads);
llama_reset_timings(lctx);
--
2.42.0
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment