Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
9119e2ac
Unverified
Commit
9119e2ac
authored
Feb 19, 2024
by
Fengzhe Zhou
Committed by
GitHub
Feb 19, 2024
Browse files
[Fix] rename qwen2-beta -> qwen1.5 (#894)
parent
b6e21ece
Changes
16
Hide whitespace changes
Inline
Side-by-side
Showing
16 changed files
with
88 additions
and
88 deletions
+88
-88
configs/models/qwen/hf_qwen1_5_0_5b.py
configs/models/qwen/hf_qwen1_5_0_5b.py
+3
-3
configs/models/qwen/hf_qwen1_5_0_5b_chat.py
configs/models/qwen/hf_qwen1_5_0_5b_chat.py
+2
-2
configs/models/qwen/hf_qwen1_5_14b.py
configs/models/qwen/hf_qwen1_5_14b.py
+3
-3
configs/models/qwen/hf_qwen1_5_14b_chat.py
configs/models/qwen/hf_qwen1_5_14b_chat.py
+2
-2
configs/models/qwen/hf_qwen1_5_1_8b.py
configs/models/qwen/hf_qwen1_5_1_8b.py
+3
-3
configs/models/qwen/hf_qwen1_5_1_8b_chat.py
configs/models/qwen/hf_qwen1_5_1_8b_chat.py
+2
-2
configs/models/qwen/hf_qwen1_5_4b.py
configs/models/qwen/hf_qwen1_5_4b.py
+3
-3
configs/models/qwen/hf_qwen1_5_4b_chat.py
configs/models/qwen/hf_qwen1_5_4b_chat.py
+2
-2
configs/models/qwen/hf_qwen1_5_72b.py
configs/models/qwen/hf_qwen1_5_72b.py
+3
-3
configs/models/qwen/hf_qwen1_5_72b_chat.py
configs/models/qwen/hf_qwen1_5_72b_chat.py
+2
-2
configs/models/qwen/hf_qwen1_5_7b.py
configs/models/qwen/hf_qwen1_5_7b.py
+25
-0
configs/models/qwen/hf_qwen1_5_7b_chat.py
configs/models/qwen/hf_qwen1_5_7b_chat.py
+34
-0
configs/models/qwen/hf_qwen2_beta_0_5b.py
configs/models/qwen/hf_qwen2_beta_0_5b.py
+0
-25
configs/models/qwen/hf_qwen2_beta_1_8b_chat.py
configs/models/qwen/hf_qwen2_beta_1_8b_chat.py
+0
-34
configs/models/qwen/vllm_qwen1_5_72b.py
configs/models/qwen/vllm_qwen1_5_72b.py
+2
-2
configs/models/qwen/vllm_qwen1_5_72b_chat.py
configs/models/qwen/vllm_qwen1_5_72b_chat.py
+2
-2
No files found.
configs/models/qwen/hf_qwen
2_beta_7
b.py
→
configs/models/qwen/hf_qwen
1_5_0_5
b.py
View file @
9119e2ac
...
...
@@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-7
b-hf'
,
path
=
"Qwen/Qwen
2-beta-7
B"
,
tokenizer_path
=
'Qwen/Qwen
2-beta-7
B'
,
abbr
=
'qwen
1.5-0.5
b-hf'
,
path
=
"Qwen/Qwen
1.5-0.5
B"
,
tokenizer_path
=
'Qwen/Qwen
1.5-0.5
B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_7
b_chat.py
→
configs/models/qwen/hf_qwen
1_5_0_5
b_chat.py
View file @
9119e2ac
...
...
@@ -11,8 +11,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-7
b-chat-hf'
,
path
=
"Qwen/Qwen
2-beta-7
B-Chat"
,
abbr
=
'qwen
1.5-0.5
b-chat-hf'
,
path
=
"Qwen/Qwen
1.5-0.5
B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_
4b.py
→
configs/models/qwen/hf_qwen
1_5_1
4b.py
View file @
9119e2ac
...
...
@@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-
4b-hf'
,
path
=
"Qwen/Qwen
2-beta-
4B"
,
tokenizer_path
=
'Qwen/Qwen
2-beta-
4B'
,
abbr
=
'qwen
1.5-1
4b-hf'
,
path
=
"Qwen/Qwen
1.5-1
4B"
,
tokenizer_path
=
'Qwen/Qwen
1.5-1
4B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_
4b_chat.py
→
configs/models/qwen/hf_qwen
1_5_1
4b_chat.py
View file @
9119e2ac
...
...
@@ -11,8 +11,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-
4b-chat-hf'
,
path
=
"Qwen/Qwen
2-beta-
4B-Chat"
,
abbr
=
'qwen
1.5-1
4b-chat-hf'
,
path
=
"Qwen/Qwen
1.5-1
4B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_14
b.py
→
configs/models/qwen/hf_qwen
1_5_1_8
b.py
View file @
9119e2ac
...
...
@@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-14
b-hf'
,
path
=
"Qwen/Qwen
2-beta-14
B"
,
tokenizer_path
=
'Qwen/Qwen
2-beta-14
B'
,
abbr
=
'qwen
1.5-1.8
b-hf'
,
path
=
"Qwen/Qwen
1.5-1.8
B"
,
tokenizer_path
=
'Qwen/Qwen
1.5-1.8
B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_14
b_chat.py
→
configs/models/qwen/hf_qwen
1_5_1_8
b_chat.py
View file @
9119e2ac
...
...
@@ -11,8 +11,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-14
b-chat-hf'
,
path
=
"Qwen/Qwen
2-beta-14
B-Chat"
,
abbr
=
'qwen
1.5-1.8
b-chat-hf'
,
path
=
"Qwen/Qwen
1.5-1.8
B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_1_8
b.py
→
configs/models/qwen/hf_qwen
1_5_4
b.py
View file @
9119e2ac
...
...
@@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-1.8
b-hf'
,
path
=
"Qwen/Qwen
2-beta-1_8
B"
,
tokenizer_path
=
'Qwen/Qwen
2-beta-1_8
B'
,
abbr
=
'qwen
1.5-4
b-hf'
,
path
=
"Qwen/Qwen
1.5-4
B"
,
tokenizer_path
=
'Qwen/Qwen
1.5-4
B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta_0_5
b_chat.py
→
configs/models/qwen/hf_qwen
1_5_4
b_chat.py
View file @
9119e2ac
...
...
@@ -11,8 +11,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta-0.5
b-chat-hf'
,
path
=
"Qwen/Qwen
2-beta-0_5
B-Chat"
,
abbr
=
'qwen
1.5-4
b-chat-hf'
,
path
=
"Qwen/Qwen
1.5-4
B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta
_72b.py
→
configs/models/qwen/hf_qwen
1_5
_72b.py
View file @
9119e2ac
...
...
@@ -3,9 +3,9 @@ from opencompass.models import HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta
-72b-hf'
,
path
=
"Qwen/Qwen
2-beta
-72B"
,
tokenizer_path
=
'Qwen/Qwen
2-beta
-72B'
,
abbr
=
'qwen
1.5
-72b-hf'
,
path
=
"Qwen/Qwen
1.5
-72B"
,
tokenizer_path
=
'Qwen/Qwen
1.5
-72B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen
2_beta
_72b_chat.py
→
configs/models/qwen/hf_qwen
1_5
_72b_chat.py
View file @
9119e2ac
...
...
@@ -11,8 +11,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen
2-beta
-72b-chat-hf'
,
path
=
"Qwen/Qwen
2-beta
-72B-Chat"
,
abbr
=
'qwen
1.5
-72b-chat-hf'
,
path
=
"Qwen/Qwen
1.5
-72B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/hf_qwen1_5_7b.py
0 → 100644
View file @
9119e2ac
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen1.5-7b-hf'
,
path
=
"Qwen/Qwen1.5-7B"
,
tokenizer_path
=
'Qwen/Qwen1.5-7B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
pad_token_id
=
151645
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen1_5_7b_chat.py
0 → 100644
View file @
9119e2ac
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
"BOT"
,
begin
=
"<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
151645
,
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen1.5-7b-chat-hf'
,
path
=
"Qwen/Qwen1.5-7B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
pad_token_id
=
151645
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/qwen/hf_qwen2_beta_0_5b.py
deleted
100644 → 0
View file @
b6e21ece
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen2-beta-0.5b-hf'
,
path
=
"Qwen/Qwen2-beta-0_5B"
,
tokenizer_path
=
'Qwen/Qwen2-beta-0_5B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
pad_token_id
=
151645
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen2_beta_1_8b_chat.py
deleted
100644 → 0
View file @
b6e21ece
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
"BOT"
,
begin
=
"<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
151645
,
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen2-beta-1.8b-chat-hf'
,
path
=
"Qwen/Qwen2-beta-1_8B-Chat"
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
pad_token_id
=
151645
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/qwen/vllm_qwen
2_beta
_72b.py
→
configs/models/qwen/vllm_qwen
1_5
_72b.py
View file @
9119e2ac
...
...
@@ -4,8 +4,8 @@ from opencompass.models import VLLM
models
=
[
dict
(
type
=
VLLM
,
abbr
=
'qwen
2-beta
-72b-vllm'
,
path
=
"Qwen/Qwen
2-beta
-72B"
,
abbr
=
'qwen
1.5
-72b-vllm'
,
path
=
"Qwen/Qwen
1.5
-72B"
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
max_out_len
=
100
,
max_seq_len
=
2048
,
...
...
configs/models/qwen/vllm_qwen
2_beta
_72b_chat.py
→
configs/models/qwen/vllm_qwen
1_5
_72b_chat.py
View file @
9119e2ac
...
...
@@ -12,8 +12,8 @@ _meta_template = dict(
models
=
[
dict
(
type
=
VLLM
,
abbr
=
'qwen
2-beta
-72b-chat-vllm'
,
path
=
"Qwen/Qwen
2-beta
-72B-Chat"
,
abbr
=
'qwen
1.5
-72b-chat-vllm'
,
path
=
"Qwen/Qwen
1.5
-72B-Chat"
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment