Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
change
sglang
Commits
2e341cd4
Unverified
Commit
2e341cd4
authored
Jul 18, 2024
by
zhyncs
Committed by
GitHub
Jul 17, 2024
Browse files
misc: add pre-commit config (#637)
parent
a8552cb1
Changes
43
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
28 additions
and
15 deletions
+28
-15
python/sglang/srt/models/qwen2_moe.py
python/sglang/srt/models/qwen2_moe.py
+6
-4
python/sglang/srt/server.py
python/sglang/srt/server.py
+17
-6
scripts/convert_yi_vl.py
scripts/convert_yi_vl.py
+5
-5
No files found.
python/sglang/srt/models/qwen2_moe.py
View file @
2e341cd4
...
@@ -400,10 +400,12 @@ class Qwen2MoeForCausalLM(nn.Module):
...
@@ -400,10 +400,12 @@ class Qwen2MoeForCausalLM(nn.Module):
expert_params_mapping
=
[
expert_params_mapping
=
[
# These are the weights for the experts
# These are the weights for the experts
# (param_name, weight_name, expert_id, shard_id)
# (param_name, weight_name, expert_id, shard_id)
(
(
(
"experts.w13_weight"
"experts.w13_weight"
if
weight_name
in
[
"gate_proj"
,
"up_proj"
]
if
weight_name
in
[
"gate_proj"
,
"up_proj"
]
else
"experts.w2_weight"
,
else
"experts.w2_weight"
),
f
"experts.
{
expert_id
}
.
{
weight_name
}
.weight"
,
f
"experts.
{
expert_id
}
.
{
weight_name
}
.weight"
,
expert_id
,
expert_id
,
shard_id
,
shard_id
,
...
@@ -418,7 +420,7 @@ class Qwen2MoeForCausalLM(nn.Module):
...
@@ -418,7 +420,7 @@ class Qwen2MoeForCausalLM(nn.Module):
for
name
,
loaded_weight
in
weights
:
for
name
,
loaded_weight
in
weights
:
if
"rotary_emb.inv_freq"
in
name
:
if
"rotary_emb.inv_freq"
in
name
:
continue
continue
for
(
param_name
,
weight_name
,
shard_id
)
in
stacked_params_mapping
:
for
param_name
,
weight_name
,
shard_id
in
stacked_params_mapping
:
# Skip non-stacked layers and experts (experts handled below).
# Skip non-stacked layers and experts (experts handled below).
if
weight_name
not
in
name
:
if
weight_name
not
in
name
:
continue
continue
...
...
python/sglang/srt/server.py
View file @
2e341cd4
...
@@ -32,8 +32,8 @@ from sglang.srt.hf_transformers_utils import get_tokenizer
...
@@ -32,8 +32,8 @@ from sglang.srt.hf_transformers_utils import get_tokenizer
from
sglang.srt.managers.controller.manager_multi
import
(
from
sglang.srt.managers.controller.manager_multi
import
(
start_controller_process
as
start_controller_process_multi
,
start_controller_process
as
start_controller_process_multi
,
)
)
from
sglang.srt.managers.controller.manager_single
import
launch_tp_servers
from
sglang.srt.managers.controller.manager_single
import
(
from
sglang.srt.managers.controller.manager_single
import
(
launch_tp_servers
,
start_controller_process
as
start_controller_process_single
,
start_controller_process
as
start_controller_process_single
,
)
)
from
sglang.srt.managers.detokenizer_manager
import
start_detokenizer_process
from
sglang.srt.managers.detokenizer_manager
import
start_detokenizer_process
...
@@ -198,11 +198,22 @@ def launch_server(server_args: ServerArgs, pipe_finish_writer, model_overide_arg
...
@@ -198,11 +198,22 @@ def launch_server(server_args: ServerArgs, pipe_finish_writer, model_overide_arg
if
server_args
.
node_rank
!=
0
:
if
server_args
.
node_rank
!=
0
:
tp_size_local
=
server_args
.
tp_size
//
server_args
.
nnodes
tp_size_local
=
server_args
.
tp_size
//
server_args
.
nnodes
gpu_ids
=
[
i
for
_
in
range
(
server_args
.
nnodes
)
for
i
in
range
(
tp_size_local
)]
gpu_ids
=
[
tp_rank_range
=
list
(
range
(
server_args
.
node_rank
*
tp_size_local
,
i
for
_
in
range
(
server_args
.
nnodes
)
for
i
in
range
(
tp_size_local
)
(
server_args
.
node_rank
+
1
)
*
tp_size_local
))
]
procs
=
launch_tp_servers
(
gpu_ids
,
tp_rank_range
,
server_args
,
tp_rank_range
=
list
(
port_args
.
model_port_args
[
0
],
model_overide_args
)
range
(
server_args
.
node_rank
*
tp_size_local
,
(
server_args
.
node_rank
+
1
)
*
tp_size_local
,
)
)
procs
=
launch_tp_servers
(
gpu_ids
,
tp_rank_range
,
server_args
,
port_args
.
model_port_args
[
0
],
model_overide_args
,
)
while
True
:
while
True
:
pass
pass
...
...
scripts/convert_yi_vl.py
View file @
2e341cd4
...
@@ -10,16 +10,15 @@ import os
...
@@ -10,16 +10,15 @@ import os
from
transformers
import
AutoConfig
,
AutoTokenizer
from
transformers
import
AutoConfig
,
AutoTokenizer
def
add_image_token
(
model_path
:
str
):
def
add_image_token
(
model_path
:
str
):
tokenizer
=
AutoTokenizer
.
from_pretrained
(
model_path
)
tokenizer
=
AutoTokenizer
.
from_pretrained
(
model_path
)
tokenizer
.
add_tokens
(
tokenizer
.
add_tokens
([
"<image_placeholder>"
],
special_tokens
=
True
)
[
"<image_placeholder>"
],
special_tokens
=
True
)
print
(
tokenizer
)
print
(
tokenizer
)
tokenizer
.
save_pretrained
(
model_path
)
tokenizer
.
save_pretrained
(
model_path
)
def
edit_model_config
(
model_path
):
def
edit_model_config
(
model_path
):
config
=
AutoConfig
.
from_pretrained
(
model_path
)
config
=
AutoConfig
.
from_pretrained
(
model_path
)
...
@@ -29,6 +28,7 @@ def edit_model_config(model_path):
...
@@ -29,6 +28,7 @@ def edit_model_config(model_path):
print
(
config
)
print
(
config
)
config
.
save_pretrained
(
model_path
)
config
.
save_pretrained
(
model_path
)
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
parser
=
argparse
.
ArgumentParser
()
parser
=
argparse
.
ArgumentParser
()
parser
.
add_argument
(
"--model-path"
,
type
=
str
)
parser
.
add_argument
(
"--model-path"
,
type
=
str
)
...
...
Prev
1
2
3
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment