Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
be517e49
Commit
be517e49
authored
Apr 05, 2024
by
Michael Yang
Browse files
no rope parameters
parent
fc8e1086
Changes
4
Hide whitespace changes
Inline
Side-by-side
Showing
4 changed files
with
0 additions
and
14 deletions
+0
-14
api/types.go
api/types.go
+0
-4
convert/convert.go
convert/convert.go
+0
-1
convert/mistral.go
convert/mistral.go
+0
-1
llm/server.go
llm/server.go
+0
-8
No files found.
api/types.go
View file @
be517e49
...
...
@@ -121,8 +121,6 @@ type Runner struct {
VocabOnly
bool
`json:"vocab_only,omitempty"`
UseMMap
bool
`json:"use_mmap,omitempty"`
UseMLock
bool
`json:"use_mlock,omitempty"`
RopeFrequencyBase
float32
`json:"rope_frequency_base,omitempty"`
RopeFrequencyScale
float32
`json:"rope_frequency_scale,omitempty"`
NumThread
int
`json:"num_thread,omitempty"`
}
...
...
@@ -383,8 +381,6 @@ func DefaultOptions() Options {
Runner
:
Runner
{
// options set when the model is loaded
NumCtx
:
2048
,
RopeFrequencyBase
:
10000.0
,
RopeFrequencyScale
:
1.0
,
NumBatch
:
512
,
NumGPU
:
-
1
,
// -1 here indicates that NumGPU should be set dynamically
NumGQA
:
1
,
...
...
convert/convert.go
View file @
be517e49
...
...
@@ -32,7 +32,6 @@ type Params struct {
AttentionHeads
int
`json:"num_attention_heads"`
// n_head
KeyValHeads
int
`json:"num_key_value_heads"`
NormEPS
float64
`json:"rms_norm_eps"`
RopeFreqBase
float64
`json:"rope_theta"`
BoSTokenID
int
`json:"bos_token_id"`
EoSTokenID
int
`json:"eos_token_id"`
HeadDimension
int
`json:"head_dim"`
...
...
convert/mistral.go
View file @
be517e49
...
...
@@ -144,7 +144,6 @@ func (m *MistralModel) WriteGGUF() (string, error) {
"llama.attention.head_count"
:
uint32
(
m
.
Params
.
AttentionHeads
),
"llama.attention.head_count_kv"
:
uint32
(
m
.
Params
.
KeyValHeads
),
"llama.attention.layer_norm_rms_epsilon"
:
float32
(
m
.
Params
.
NormEPS
),
"llama.rope.freq_base"
:
float32
(
m
.
Params
.
RopeFreqBase
),
"general.file_type"
:
uint32
(
1
),
"tokenizer.ggml.model"
:
"llama"
,
...
...
llm/server.go
View file @
be517e49
...
...
@@ -172,14 +172,6 @@ func NewLlamaServer(model string, adapters, projectors []string, opts api.Option
params
=
append
(
params
,
"--main-gpu"
,
fmt
.
Sprintf
(
"%d"
,
opts
.
MainGPU
))
}
if
opts
.
RopeFrequencyBase
>
0
{
params
=
append
(
params
,
"--rope-freq-base"
,
fmt
.
Sprintf
(
"%f"
,
opts
.
RopeFrequencyBase
))
}
if
opts
.
RopeFrequencyScale
>
0
{
params
=
append
(
params
,
"--rope-freq-scale"
,
fmt
.
Sprintf
(
"%f"
,
opts
.
RopeFrequencyScale
))
}
if
len
(
adapters
)
>
0
{
// TODO: applying multiple adapters is not supported by the llama.cpp server yet
params
=
append
(
params
,
"--lora"
,
adapters
[
0
])
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment