Unverified Commit 34f14279 authored by Jeffrey Morgan's avatar Jeffrey Morgan Committed by GitHub
Browse files

llm: always add bos token to prompt (#4941)



* fix embedding by adding fixes from llama.cpp upstream

* remove assert

---------
Co-authored-by: default avatarJesper Ek <deadbeef84@gmail.com>
parent 46a7f1e7
......@@ -835,7 +835,7 @@ struct llama_server_context
system_tokens.clear();
if (!system_prompt.empty()) {
system_tokens = ::llama_tokenize(ctx, system_prompt, add_bos_token);
system_tokens = ::llama_tokenize(ctx, system_prompt, true);
llama_batch_clear(batch);
......@@ -1656,7 +1656,7 @@ struct llama_server_context
slot.t_start_process_prompt = ggml_time_us();
slot.t_start_genereration = 0;
prompt_tokens = tokenize(slot.prompt, system_prompt.empty() && add_bos_token); // add BOS if there isn't system prompt
prompt_tokens = tokenize(slot.prompt, system_prompt.empty()); // add BOS if there isn't system prompt
slot.n_prompt_tokens = prompt_tokens.size();
......
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment