Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
ad3a7d0e
"vscode:/vscode.git/clone" did not exist on "886aa32730618dbcac35edc8dbdbb69e826ef6bf"
Commit
ad3a7d0e
authored
Jul 27, 2023
by
Michael Yang
Browse files
add NumGQA
parent
18ffeeec
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
3 additions
and
0 deletions
+3
-0
api/types.go
api/types.go
+2
-0
llama/llama.go
llama/llama.go
+1
-0
No files found.
api/types.go
View file @
ad3a7d0e
...
...
@@ -153,6 +153,7 @@ type Options struct {
NumCtx
int
`json:"num_ctx,omitempty"`
NumKeep
int
`json:"num_keep,omitempty"`
NumBatch
int
`json:"num_batch,omitempty"`
NumGQA
int
`json:"num_gqa,omitempty"`
NumGPU
int
`json:"num_gpu,omitempty"`
MainGPU
int
`json:"main_gpu,omitempty"`
LowVRAM
bool
`json:"low_vram,omitempty"`
...
...
@@ -190,6 +191,7 @@ func DefaultOptions() Options {
NumCtx
:
2048
,
NumBatch
:
1024
,
NumGPU
:
1
,
NumGQA
:
1
,
LowVRAM
:
false
,
F16KV
:
true
,
UseMMap
:
true
,
...
...
llama/llama.go
View file @
ad3a7d0e
...
...
@@ -127,6 +127,7 @@ func New(model string, opts api.Options) (*LLM, error) {
params
.
seed
=
C
.
uint
(
llm
.
Seed
)
params
.
n_ctx
=
C
.
int
(
llm
.
NumCtx
)
params
.
n_batch
=
C
.
int
(
llm
.
NumBatch
)
params
.
n_gqa
=
C
.
int
(
llm
.
NumGQA
)
params
.
n_gpu_layers
=
C
.
int
(
llm
.
NumGPU
)
params
.
main_gpu
=
C
.
int
(
llm
.
MainGPU
)
params
.
low_vram
=
C
.
bool
(
llm
.
LowVRAM
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment