Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
aa2dd2b5
Unverified
Commit
aa2dd2b5
authored
May 14, 2024
by
Fengzhe Zhou
Committed by
GitHub
May 14, 2024
Browse files
[Format] Add config lints (#892)
parent
3dbba119
Changes
648
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
36 additions
and
36 deletions
+36
-36
configs/models/hf_internlm/hf_internlm2_1_8b.py
configs/models/hf_internlm/hf_internlm2_1_8b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_20b.py
configs/models/hf_internlm/hf_internlm2_20b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_7b.py
configs/models/hf_internlm/hf_internlm2_7b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_base_20b.py
configs/models/hf_internlm/hf_internlm2_base_20b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_base_7b.py
configs/models/hf_internlm/hf_internlm2_base_7b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
...s/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
+2
-2
configs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
...gs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
+2
-2
configs/models/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
...els/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
+1
-1
configs/models/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
...dels/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
+1
-1
configs/models/hf_internlm/hf_internlm2_math_20b.py
configs/models/hf_internlm/hf_internlm2_math_20b.py
+1
-1
configs/models/hf_internlm/hf_internlm2_math_7b.py
configs/models/hf_internlm/hf_internlm2_math_7b.py
+1
-1
configs/models/hf_internlm/hf_internlm_20b.py
configs/models/hf_internlm/hf_internlm_20b.py
+1
-1
configs/models/hf_internlm/hf_internlm_7b.py
configs/models/hf_internlm/hf_internlm_7b.py
+1
-1
configs/models/hf_internlm/hf_internlm_chat_20b.py
configs/models/hf_internlm/hf_internlm_chat_20b.py
+2
-2
configs/models/hf_internlm/hf_internlm_chat_7b.py
configs/models/hf_internlm/hf_internlm_chat_7b.py
+2
-2
configs/models/hf_internlm/lmdeploy_internlm2_20b.py
configs/models/hf_internlm/lmdeploy_internlm2_20b.py
+3
-3
configs/models/hf_internlm/lmdeploy_internlm2_chat_20b.py
configs/models/hf_internlm/lmdeploy_internlm2_chat_20b.py
+5
-5
configs/models/hf_internlm/lmdeploy_internlm2_chat_7b.py
configs/models/hf_internlm/lmdeploy_internlm2_chat_7b.py
+5
-5
configs/models/hf_llama/lmdeploy_llama3_70b_instruct.py
configs/models/hf_llama/lmdeploy_llama3_70b_instruct.py
+2
-2
configs/models/hf_llama/lmdeploy_llama3_8b_instruct.py
configs/models/hf_llama/lmdeploy_llama3_8b_instruct.py
+2
-2
No files found.
configs/models/hf_internlm/hf_internlm2_1_8b.py
View file @
aa2dd2b5
...
...
@@ -4,7 +4,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-1.8b-hf'
,
path
=
"
internlm/internlm2-1_8b
"
,
path
=
'
internlm/internlm2-1_8b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
),
...
...
configs/models/hf_internlm/hf_internlm2_20b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-20b-hf'
,
path
=
"
internlm/internlm2-20b
"
,
path
=
'
internlm/internlm2-20b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
),
...
...
configs/models/hf_internlm/hf_internlm2_7b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-7b-hf'
,
path
=
"
internlm/internlm2-7b
"
,
path
=
'
internlm/internlm2-7b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
),
...
...
configs/models/hf_internlm/hf_internlm2_base_20b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-base-20b-hf'
,
path
=
"
internlm/internlm2-base-20b
"
,
path
=
'
internlm/internlm2-base-20b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
),
...
...
configs/models/hf_internlm/hf_internlm2_base_7b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-base-7b-hf'
,
path
=
"
internlm/internlm2-base-7b
"
,
path
=
'
internlm/internlm2-base-7b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
),
...
...
configs/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
View file @
aa2dd2b5
...
...
@@ -13,7 +13,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-20b-hf'
,
path
=
"
internlm/internlm2-chat-20b
"
,
path
=
'
internlm/internlm2-chat-20b
'
,
tokenizer_path
=
'internlm/internlm2-chat-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
@@ -31,7 +31,7 @@ models = [
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
generation_kwargs
=
{
"
eos_token_id
"
:
[
2
,
92542
]},
generation_kwargs
=
{
'
eos_token_id
'
:
[
2
,
92542
]},
batch_padding
=
True
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
View file @
aa2dd2b5
...
...
@@ -13,7 +13,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-7b-hf'
,
path
=
"
internlm/internlm2-chat-7b
"
,
path
=
'
internlm/internlm2-chat-7b
'
,
tokenizer_path
=
'internlm/internlm2-chat-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
@@ -31,7 +31,7 @@ models = [
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
generation_kwargs
=
{
"
eos_token_id
"
:
[
2
,
92542
]},
generation_kwargs
=
{
'
eos_token_id
'
:
[
2
,
92542
]},
batch_padding
=
True
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
View file @
aa2dd2b5
...
...
@@ -13,7 +13,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-20b-hf'
,
path
=
"
internlm/internlm2-math-20b
"
,
path
=
'
internlm/internlm2-math-20b
'
,
tokenizer_path
=
'internlm/internlm2-math-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
configs/models/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
View file @
aa2dd2b5
...
...
@@ -13,7 +13,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-7b-hf'
,
path
=
"
internlm/internlm2-math-7b
"
,
path
=
'
internlm/internlm2-math-7b
'
,
tokenizer_path
=
'internlm/internlm2-math-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
configs/models/hf_internlm/hf_internlm2_math_20b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-math-20b-hf'
,
path
=
"
internlm/internlm2-math-20b
"
,
path
=
'
internlm/internlm2-math-20b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
),
...
...
configs/models/hf_internlm/hf_internlm2_math_7b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm2-math-7b-hf'
,
path
=
"
internlm/internlm2-math-7b
"
,
path
=
'
internlm/internlm2-math-7b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
),
...
...
configs/models/hf_internlm/hf_internlm_20b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm-20b-hf'
,
path
=
"
internlm/internlm-20b
"
,
path
=
'
internlm/internlm-20b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
),
...
...
configs/models/hf_internlm/hf_internlm_7b.py
View file @
aa2dd2b5
...
...
@@ -5,7 +5,7 @@ models = [
dict
(
type
=
HuggingFaceBaseModel
,
abbr
=
'internlm-7b-hf'
,
path
=
"
internlm/internlm-7b
"
,
path
=
'
internlm/internlm-7b
'
,
max_out_len
=
1024
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
),
...
...
configs/models/hf_internlm/hf_internlm_chat_20b.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm-chat-20b-hf'
,
path
=
"
internlm/internlm-chat-20b
"
,
path
=
'
internlm/internlm-chat-20b
'
,
tokenizer_path
=
'internlm/internlm-chat-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
configs/models/hf_internlm/hf_internlm_chat_7b.py
View file @
aa2dd2b5
...
...
@@ -12,7 +12,7 @@ models = [
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm-chat-7b-hf'
,
path
=
"
internlm/internlm-chat-7b
"
,
path
=
'
internlm/internlm-chat-7b
'
,
tokenizer_path
=
'internlm/internlm-chat-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
...
...
configs/models/hf_internlm/lmdeploy_internlm2_20b.py
View file @
aa2dd2b5
...
...
@@ -4,12 +4,12 @@ from opencompass.models.turbomind import TurboMindModel
models
=
[
dict
(
type
=
TurboMindModel
,
abbr
=
"
internlm2-20b-turbomind
"
,
path
=
"
internlm/internlm2-20b
"
,
abbr
=
'
internlm2-20b-turbomind
'
,
path
=
'
internlm/internlm2-20b
'
,
engine_config
=
dict
(
session_len
=
32768
,
max_batch_size
=
32
,
model_name
=
"
internlm2-20b
"
,
model_name
=
'
internlm2-20b
'
,
tp
=
2
,
),
gen_config
=
dict
(
...
...
configs/models/hf_internlm/lmdeploy_internlm2_chat_20b.py
View file @
aa2dd2b5
...
...
@@ -3,21 +3,21 @@ from opencompass.models.turbomind import TurboMindModel
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
"
<|im_start|>user
\n
"
,
end
=
"
<|im_end|>
\n
"
),
dict
(
role
=
"
BOT
"
,
begin
=
"
<|im_start|>assistant
\n
"
,
end
=
"
<|im_end|>
\n
"
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'
<|im_start|>user
\n
'
,
end
=
'
<|im_end|>
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
<|im_start|>assistant
\n
'
,
end
=
'
<|im_end|>
\n
'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
TurboMindModel
,
abbr
=
"
internlm2-chat-20b-turbomind
"
,
path
=
"
internlm/internlm2-chat-20b
"
,
abbr
=
'
internlm2-chat-20b-turbomind
'
,
path
=
'
internlm/internlm2-chat-20b
'
,
meta_template
=
_meta_template
,
engine_config
=
dict
(
session_len
=
32768
,
max_batch_size
=
32
,
model_name
=
"
internlm2-chat-20b
"
,
model_name
=
'
internlm2-chat-20b
'
,
tp
=
2
,
stop_words
=
[
2
,
92542
],
),
...
...
configs/models/hf_internlm/lmdeploy_internlm2_chat_7b.py
View file @
aa2dd2b5
...
...
@@ -3,21 +3,21 @@ from opencompass.models.turbomind import TurboMindModel
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
"
<|im_start|>user
\n
"
,
end
=
"
<|im_end|>
\n
"
),
dict
(
role
=
"
BOT
"
,
begin
=
"
<|im_start|>assistant
\n
"
,
end
=
"
<|im_end|>
\n
"
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'
<|im_start|>user
\n
'
,
end
=
'
<|im_end|>
\n
'
),
dict
(
role
=
'
BOT
'
,
begin
=
'
<|im_start|>assistant
\n
'
,
end
=
'
<|im_end|>
\n
'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
TurboMindModel
,
abbr
=
"
internlm2-chat-7b-turbomind
"
,
path
=
"
internlm/internlm2-chat-7b
"
,
abbr
=
'
internlm2-chat-7b-turbomind
'
,
path
=
'
internlm/internlm2-chat-7b
'
,
meta_template
=
_meta_template
,
engine_config
=
dict
(
session_len
=
32768
,
max_batch_size
=
32
,
model_name
=
"
internlm2-chat-7b
"
,
model_name
=
'
internlm2-chat-7b
'
,
tp
=
1
,
stop_words
=
[
2
,
92542
],
),
...
...
configs/models/hf_llama/lmdeploy_llama3_70b_instruct.py
View file @
aa2dd2b5
...
...
@@ -2,8 +2,8 @@ from opencompass.models import TurboMindModel
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'<|begin_of_text|>user<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
),
dict
(
role
=
"
BOT
"
,
begin
=
'<|begin_of_text|>assistant<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'<|begin_of_text|>user<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
),
dict
(
role
=
'
BOT
'
,
begin
=
'<|begin_of_text|>assistant<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
,
generate
=
True
),
],
)
...
...
configs/models/hf_llama/lmdeploy_llama3_8b_instruct.py
View file @
aa2dd2b5
...
...
@@ -2,8 +2,8 @@ from opencompass.models import TurboMindModel
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"
HUMAN
"
,
begin
=
'<|begin_of_text|>user<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
),
dict
(
role
=
"
BOT
"
,
begin
=
'<|begin_of_text|>assistant<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
,
generate
=
True
),
dict
(
role
=
'
HUMAN
'
,
begin
=
'<|begin_of_text|>user<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
),
dict
(
role
=
'
BOT
'
,
begin
=
'<|begin_of_text|>assistant<|end_header_id|>
\n\n
'
,
end
=
'<|eot_id|>'
,
generate
=
True
),
],
)
...
...
Prev
1
…
23
24
25
26
27
28
29
30
31
…
33
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment