Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
ollama
Commits
6b32a2d5
"git@developer.sourcefind.cn:OpenDAS/ollama.git" did not exist on "4e986a823ca47eb16f563d15a6fe4cc393a00715"
Commit
6b32a2d5
authored
Mar 10, 2025
by
Michael Yang
Browse files
compat with upstream gguf
parent
c5cbe4fc
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
14 additions
and
14 deletions
+14
-14
convert/convert_gemma3.go
convert/convert_gemma3.go
+8
-8
model/models/gemma3/model.go
model/models/gemma3/model.go
+1
-1
model/models/gemma3/model_text.go
model/models/gemma3/model_text.go
+5
-5
No files found.
convert/convert_gemma3.go
View file @
6b32a2d5
...
@@ -76,19 +76,19 @@ func (p *gemma3Model) KV(t *Tokenizer) ggml.KV {
...
@@ -76,19 +76,19 @@ func (p *gemma3Model) KV(t *Tokenizer) ggml.KV {
switch
p
.
Architecture
{
switch
p
.
Architecture
{
case
"Gemma3ForCausalLM"
:
case
"Gemma3ForCausalLM"
:
kv
[
"gemma3.context_length"
]
=
p
.
MaxPositionEmbeddings
kv
[
"gemma3.context_length"
]
=
p
.
MaxPositionEmbeddings
kv
[
"gemma3.
text.
attention.layer_norm_rms_epsilon"
]
=
p
.
RMSNormEPS
kv
[
"gemma3.attention.layer_norm_rms_epsilon"
]
=
p
.
RMSNormEPS
kv
[
"gemma3.attention.key_length"
]
=
p
.
HeadDim
kv
[
"gemma3.attention.key_length"
]
=
p
.
HeadDim
kv
[
"gemma3.attention.value_length"
]
=
p
.
HeadDim
kv
[
"gemma3.attention.value_length"
]
=
p
.
HeadDim
kv
[
"gemma3.
text.
attention.sliding_window"
]
=
p
.
SlidingWindow
kv
[
"gemma3.attention.sliding_window"
]
=
p
.
SlidingWindow
kv
[
"gemma3.
text.
final_logit_softcapping"
]
=
p
.
FinalLogitSoftcap
kv
[
"gemma3.final_logit_softcapping"
]
=
p
.
FinalLogitSoftcap
kv
[
"gemma3.
text.
rope.local.freq_base"
]
=
p
.
RopeLocalTheta
kv
[
"gemma3.rope.local.freq_base"
]
=
p
.
RopeLocalTheta
kv
[
"gemma3.
text.
rope.global.freq_base"
]
=
p
.
RopeGlobalTheta
kv
[
"gemma3.rope.global.freq_base"
]
=
p
.
RopeGlobalTheta
kv
[
"gemma3.embedding_length"
]
=
p
.
HiddenSize
kv
[
"gemma3.embedding_length"
]
=
p
.
HiddenSize
kv
[
"gemma3.
text.
feed_forward_length"
]
=
p
.
IntermediateSize
kv
[
"gemma3.feed_forward_length"
]
=
p
.
IntermediateSize
default
:
default
:
kv
[
"gemma3.embedding_length"
]
=
p
.
TextModel
.
HiddenSize
kv
[
"gemma3.embedding_length"
]
=
p
.
TextModel
.
HiddenSize
kv
[
"gemma3.
text.
feed_forward_length"
]
=
p
.
TextModel
.
IntermediateSize
kv
[
"gemma3.feed_forward_length"
]
=
p
.
TextModel
.
IntermediateSize
kv
[
"gemma3.
text.
attention.sliding_window"
]
=
p
.
TextModel
.
SlidingWindow
kv
[
"gemma3.attention.sliding_window"
]
=
p
.
TextModel
.
SlidingWindow
kv
[
"gemma3.vision.block_count"
]
=
p
.
VisionModel
.
NumHiddenLayers
kv
[
"gemma3.vision.block_count"
]
=
p
.
VisionModel
.
NumHiddenLayers
kv
[
"gemma3.vision.embedding_length"
]
=
p
.
VisionModel
.
HiddenSize
kv
[
"gemma3.vision.embedding_length"
]
=
p
.
VisionModel
.
HiddenSize
kv
[
"gemma3.vision.feed_forward_length"
]
=
p
.
VisionModel
.
IntermediateSize
kv
[
"gemma3.vision.feed_forward_length"
]
=
p
.
VisionModel
.
IntermediateSize
...
...
model/models/gemma3/model.go
View file @
6b32a2d5
...
@@ -62,7 +62,7 @@ func New(c ml.Config) (model.Model, error) {
...
@@ -62,7 +62,7 @@ func New(c ml.Config) (model.Model, error) {
TextModel
:
newTextModel
(
c
),
TextModel
:
newTextModel
(
c
),
}
}
slidingWindowLen
:=
int32
(
c
.
Uint
(
"
text.
attention.sliding_window"
))
slidingWindowLen
:=
int32
(
c
.
Uint
(
"attention.sliding_window"
))
m
.
Cache
=
kvcache
.
NewWrapperCache
(
kvcache
.
NewSWACache
(
slidingWindowLen
,
m
.
Shift
),
kvcache
.
NewCausalCache
(
m
.
Shift
))
m
.
Cache
=
kvcache
.
NewWrapperCache
(
kvcache
.
NewSWACache
(
slidingWindowLen
,
m
.
Shift
),
kvcache
.
NewCausalCache
(
m
.
Shift
))
return
&
m
,
nil
return
&
m
,
nil
...
...
model/models/gemma3/model_text.go
View file @
6b32a2d5
...
@@ -62,11 +62,11 @@ func newTextModel(c ml.Config) *TextModel {
...
@@ -62,11 +62,11 @@ func newTextModel(c ml.Config) *TextModel {
numKVHeads
:
int
(
c
.
Uint
(
"attention.head_count_kv"
)),
numKVHeads
:
int
(
c
.
Uint
(
"attention.head_count_kv"
)),
attnKeyLen
:
int
(
c
.
Uint
(
"attention.key_length"
,
256
)),
attnKeyLen
:
int
(
c
.
Uint
(
"attention.key_length"
,
256
)),
attnValLen
:
int
(
c
.
Uint
(
"attention.value_length"
,
256
)),
attnValLen
:
int
(
c
.
Uint
(
"attention.value_length"
,
256
)),
eps
:
c
.
Float
(
"
text.
attention.layer_norm_rms_epsilon"
,
1e-06
),
eps
:
c
.
Float
(
"attention.layer_norm_rms_epsilon"
,
1e-06
),
ropeLocalBase
:
c
.
Float
(
"
text.
rope.local.freq_base"
,
10000.0
),
ropeLocalBase
:
c
.
Float
(
"rope.local.freq_base"
,
10000.0
),
ropeGlobalBase
:
c
.
Float
(
"
text.
rope.global.freq_base"
,
1000000.0
),
ropeGlobalBase
:
c
.
Float
(
"rope.global.freq_base"
,
1000000.0
),
ropeScale
:
c
.
Float
(
"
text.
rope.freq_scale"
,
1.0
),
ropeScale
:
c
.
Float
(
"rope.freq_scale"
,
1.0
),
finalLogitSoftcap
:
c
.
Float
(
"
text.
final_logit_softcapping"
,
30.0
),
finalLogitSoftcap
:
c
.
Float
(
"final_logit_softcapping"
,
30.0
),
},
},
}
}
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment