Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
aa2dd2b5
Unverified
Commit
aa2dd2b5
authored
May 14, 2024
by
Fengzhe Zhou
Committed by
GitHub
May 14, 2024
Browse files
[Format] Add config lints (#892)
parent
3dbba119
Changes
648
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
37 additions
and
37 deletions
+37
-37
configs/models/others/vllm_orionstar_14b_longchat.py
configs/models/others/vllm_orionstar_14b_longchat.py
+2
-2
configs/models/qwen/hf_qwen1_5_moe_a2_7b.py
configs/models/qwen/hf_qwen1_5_moe_a2_7b.py
+1
-1
configs/models/qwen/hf_qwen1_5_moe_a2_7b_chat.py
configs/models/qwen/hf_qwen1_5_moe_a2_7b_chat.py
+3
-3
configs/models/qwen/ms_qwen_7b_chat.py
configs/models/qwen/ms_qwen_7b_chat.py
+3
-3
configs/models/qwen/vllm_qwen1_5_14b_chat.py
configs/models/qwen/vllm_qwen1_5_14b_chat.py
+3
-3
configs/models/qwen/vllm_qwen1_5_72b.py
configs/models/qwen/vllm_qwen1_5_72b.py
+1
-1
configs/models/qwen/vllm_qwen1_5_72b_chat.py
configs/models/qwen/vllm_qwen1_5_72b_chat.py
+3
-3
configs/models/qwen/vllm_qwen_14b_chat.py
configs/models/qwen/vllm_qwen_14b_chat.py
+3
-3
configs/models/qwen/vllm_qwen_72b_chat.py
configs/models/qwen/vllm_qwen_72b_chat.py
+3
-3
configs/models/rwkv/rwkv5_3b.py
configs/models/rwkv/rwkv5_3b.py
+1
-1
configs/models/tigerbot/hf_tigerbot_13b_chat_v1.py
configs/models/tigerbot/hf_tigerbot_13b_chat_v1.py
+1
-1
configs/models/tigerbot/hf_tigerbot_13b_chat_v2.py
configs/models/tigerbot/hf_tigerbot_13b_chat_v2.py
+1
-1
configs/models/tigerbot/hf_tigerbot_70b_chat_v2.py
configs/models/tigerbot/hf_tigerbot_70b_chat_v2.py
+1
-1
configs/models/tigerbot/hf_tigerbot_70b_chat_v3.py
configs/models/tigerbot/hf_tigerbot_70b_chat_v3.py
+1
-1
configs/models/tigerbot/hf_tigerbot_7b_chat_v3.py
configs/models/tigerbot/hf_tigerbot_7b_chat_v3.py
+1
-1
configs/models/tigerbot/hf_tigerbot_7b_sft.py
configs/models/tigerbot/hf_tigerbot_7b_sft.py
+1
-1
configs/models/vicuna/vllm_vicuna_13b_v15_16k.py
configs/models/vicuna/vllm_vicuna_13b_v15_16k.py
+3
-3
configs/models/vicuna/vllm_vicuna_7b_v15_16k.py
configs/models/vicuna/vllm_vicuna_7b_v15_16k.py
+3
-3
configs/models/wizardcoder/hf_wizardcoder_15b.py
configs/models/wizardcoder/hf_wizardcoder_15b.py
+1
-1
configs/models/wizardcoder/hf_wizardcoder_1b.py
configs/models/wizardcoder/hf_wizardcoder_1b.py
+1
-1
No files found.
configs/models/others/vllm_orionstar_14b_longchat.py
View file @
aa2dd2b5
...
...
@@ -4,8 +4,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
begin
=
'<s>'
,
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'Human: '
,
end
=
'
\n
'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
Assistant:
"
,
end
=
'</s>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'Human: '
,
end
=
'
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
Assistant:
'
,
end
=
'</s>'
,
generate
=
True
),
],
)
...
...
configs/models/qwen/hf_qwen1_5_moe_a2_7b.py
View file @
aa2dd2b5
...
...
@@ -4,7 +4,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen1.5-moe-a2-7b-hf'
,
path
=
"
Qwen/Qwen1.5-MoE-A2.7B
"
,
path
=
'
Qwen/Qwen1.5-MoE-A2.7B
'
,
tokenizer_path
=
'Qwen/Qwen1.5-MoE-A2.7B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
...
...
configs/models/qwen/hf_qwen1_5_moe_a2_7b_chat.py
View file @
aa2dd2b5
...
...
@@ -2,8 +2,8 @@ from opencompass.models import HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
)
...
...
@@ -11,7 +11,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen1.5-moe-a2-7b-chat-hf'
,
path
=
"
Qwen/Qwen1.5-MoE-A2.7B-Chat
"
,
path
=
'
Qwen/Qwen1.5-MoE-A2.7B-Chat
'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
...
...
configs/models/qwen/ms_qwen_7b_chat.py
View file @
aa2dd2b5
...
...
@@ -3,8 +3,8 @@ from opencompass.models import ModelScopeCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
\n
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
\n
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>'
,
generate
=
True
),
],
)
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
ModelScopeCausalLM
,
abbr
=
'qwen-7b-chat-ms'
,
path
=
"
qwen/Qwen-7B-Chat
"
,
path
=
'
qwen/Qwen-7B-Chat
'
,
tokenizer_path
=
'qwen/Qwen-7B-Chat'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/qwen/vllm_qwen1_5_14b_chat.py
View file @
aa2dd2b5
...
...
@@ -3,8 +3,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
)
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'qwen1.5-14b-chat-vllm'
,
path
=
"
Qwen/Qwen1.5-14B-Chat
"
,
path
=
'
Qwen/Qwen1.5-14B-Chat
'
,
model_kwargs
=
dict
(
tensor_parallel_size
=
2
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
...
...
configs/models/qwen/vllm_qwen1_5_72b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'qwen1.5-72b-vllm'
,
path
=
"
Qwen/Qwen1.5-72B
"
,
path
=
'
Qwen/Qwen1.5-72B
'
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
max_out_len
=
100
,
max_seq_len
=
2048
,
...
...
configs/models/qwen/vllm_qwen1_5_72b_chat.py
View file @
aa2dd2b5
...
...
@@ -3,8 +3,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
)
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'qwen1.5-72b-chat-vllm'
,
path
=
"
Qwen/Qwen1.5-72B-Chat
"
,
path
=
'
Qwen/Qwen1.5-72B-Chat
'
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
...
...
configs/models/qwen/vllm_qwen_14b_chat.py
View file @
aa2dd2b5
...
...
@@ -3,8 +3,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
\n
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
\n
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>'
,
generate
=
True
),
],
)
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'qwen-14b-chat-vllm'
,
path
=
"
Qwen/Qwen-14B-Chat
"
,
path
=
'
Qwen/Qwen-14B-Chat
'
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
...
...
configs/models/qwen/vllm_qwen_72b_chat.py
View file @
aa2dd2b5
...
...
@@ -3,8 +3,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
"
BOT
"
,
begin
=
"
\n
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
\n
<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>'
,
generate
=
True
),
],
)
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'qwen-72b-chat-vllm'
,
path
=
"
Qwen/Qwen-72B-Chat
"
,
path
=
'
Qwen/Qwen-72B-Chat
'
,
model_kwargs
=
dict
(
tensor_parallel_size
=
4
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
...
...
configs/models/rwkv/rwkv5_3b.py
View file @
aa2dd2b5
...
...
@@ -4,7 +4,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'rwkv-5-3b'
,
path
=
"
RWKV/rwkv-5-world-3b
"
,
path
=
'
RWKV/rwkv-5-world-3b
'
,
tokenizer_path
=
'RWKV/rwkv-5-world-3b'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
...
...
configs/models/tigerbot/hf_tigerbot_13b_chat_v1.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-13b-chat-v1-hf'
,
path
=
"
TigerResearch/tigerbot-13b-chat-v1
"
,
path
=
'
TigerResearch/tigerbot-13b-chat-v1
'
,
tokenizer_path
=
'TigerResearch/tigerbot-13b-chat-v1'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/tigerbot/hf_tigerbot_13b_chat_v2.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-13b-chat-v2-hf'
,
path
=
"
TigerResearch/tigerbot-13b-chat
"
,
path
=
'
TigerResearch/tigerbot-13b-chat
'
,
tokenizer_path
=
'TigerResearch/tigerbot-13b-chat'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/tigerbot/hf_tigerbot_70b_chat_v2.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-70b-chat-v2-hf'
,
path
=
"
TigerResearch/tigerbot-70b-chat-v2
"
,
path
=
'
TigerResearch/tigerbot-70b-chat-v2
'
,
tokenizer_path
=
'TigerResearch/tigerbot-70b-chat-v2'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/tigerbot/hf_tigerbot_70b_chat_v3.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-70b-chat-v3-hf'
,
path
=
"
TigerResearch/tigerbot-70b-chat-v3
"
,
path
=
'
TigerResearch/tigerbot-70b-chat-v3
'
,
tokenizer_path
=
'TigerResearch/tigerbot-70b-chat-v3'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
configs/models/tigerbot/hf_tigerbot_7b_chat_v3.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-7b-chat-v3-hf'
,
path
=
"
TigerResearch/tigerbot-7b-chat
"
,
path
=
'
TigerResearch/tigerbot-7b-chat
'
,
tokenizer_path
=
'TigerResearch/tigerbot-7b-chat'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/tigerbot/hf_tigerbot_7b_sft.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'tigerbot-sft-7b-hf'
,
path
=
"
TigerResearch/tigerbot-7b-sft
"
,
path
=
'
TigerResearch/tigerbot-7b-sft
'
,
tokenizer_path
=
'TigerResearch/tigerbot-7b-sft'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/vicuna/vllm_vicuna_13b_v15_16k.py
View file @
aa2dd2b5
...
...
@@ -2,8 +2,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'USER: '
),
dict
(
role
=
"
BOT
"
,
begin
=
"
ASSISTANT:
"
,
end
=
'</s>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'USER: '
),
dict
(
role
=
'
BOT
'
,
begin
=
'
ASSISTANT:
'
,
end
=
'</s>'
,
generate
=
True
),
],
)
...
...
@@ -11,7 +11,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'vicuna-13b-v1.5-16k-vllm'
,
path
=
"
lmsys/vicuna-13b-v1.5-16k
"
,
path
=
'
lmsys/vicuna-13b-v1.5-16k
'
,
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
...
...
configs/models/vicuna/vllm_vicuna_7b_v15_16k.py
View file @
aa2dd2b5
...
...
@@ -2,8 +2,8 @@ from opencompass.models import VLLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'USER: '
),
dict
(
role
=
"
BOT
"
,
begin
=
"
ASSISTANT:
"
,
end
=
'</s>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'USER: '
),
dict
(
role
=
'
BOT
'
,
begin
=
'
ASSISTANT:
'
,
end
=
'</s>'
,
generate
=
True
),
],
)
...
...
@@ -11,7 +11,7 @@ models = [
dict
(
type
=
VLLM
,
abbr
=
'vicuna-7b-v1.5-16k-vllm'
,
path
=
"
lmsys/vicuna-7b-v1.5-16k
"
,
path
=
'
lmsys/vicuna-7b-v1.5-16k
'
,
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
...
...
configs/models/wizardcoder/hf_wizardcoder_15b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'WizardCoder-15B-V1.0'
,
path
=
"
WizardLM/WizardCoder-15B-V1.0
"
,
path
=
'
WizardLM/WizardCoder-15B-V1.0
'
,
tokenizer_path
=
'WizardLM/WizardCoder-15B-V1.0'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
configs/models/wizardcoder/hf_wizardcoder_1b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'WizardCoder-1B-V1.0'
,
path
=
"
WizardLM/WizardCoder-1B-V1.0
"
,
path
=
'
WizardLM/WizardCoder-1B-V1.0
'
,
tokenizer_path
=
'WizardLM/WizardCoder-1B-V1.0'
,
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
...
...
Prev
1
…
26
27
28
29
30
31
32
33
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment