Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
d34ba111
Unverified
Commit
d34ba111
authored
Feb 05, 2024
by
Fengzhe Zhou
Committed by
GitHub
Feb 05, 2024
Browse files
[Sync] Merge branch 'dev' into zfz/update-keyset-demo (#876)
parent
32b5948f
Changes
97
Show whitespace changes
Inline
Side-by-side
Showing
20 changed files
with
472 additions
and
12 deletions
+472
-12
configs/models/hf_internlm/hf_internlm2_base_20b.py
configs/models/hf_internlm/hf_internlm2_base_20b.py
+26
-0
configs/models/hf_internlm/hf_internlm2_base_7b.py
configs/models/hf_internlm/hf_internlm2_base_7b.py
+26
-0
configs/models/hf_internlm/hf_internlm2_chat_1_8b_sft.py
configs/models/hf_internlm/hf_internlm2_chat_1_8b_sft.py
+35
-0
configs/models/hf_internlm/hf_internlm2_chat_20b.py
configs/models/hf_internlm/hf_internlm2_chat_20b.py
+0
-1
configs/models/hf_internlm/hf_internlm2_chat_20b_sft.py
configs/models/hf_internlm/hf_internlm2_chat_20b_sft.py
+35
-0
configs/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
...s/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
+36
-0
configs/models/hf_internlm/hf_internlm2_chat_7b.py
configs/models/hf_internlm/hf_internlm2_chat_7b.py
+0
-1
configs/models/hf_internlm/hf_internlm2_chat_7b_sft.py
configs/models/hf_internlm/hf_internlm2_chat_7b_sft.py
+35
-0
configs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
...gs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
+36
-0
configs/models/hf_internlm/hf_internlm2_chat_math_20b.py
configs/models/hf_internlm/hf_internlm2_chat_math_20b.py
+35
-0
configs/models/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
...els/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
+36
-0
configs/models/hf_internlm/hf_internlm2_chat_math_7b.py
configs/models/hf_internlm/hf_internlm2_chat_math_7b.py
+35
-0
configs/models/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
...dels/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
+36
-0
configs/models/nanbeige/hf_nanbeige_16b_chat.py
configs/models/nanbeige/hf_nanbeige_16b_chat.py
+4
-3
configs/models/nanbeige/hf_nanbeige_16b_chat_32k.py
configs/models/nanbeige/hf_nanbeige_16b_chat_32k.py
+5
-5
configs/models/others/hf_openchat_35_0106.py
configs/models/others/hf_openchat_35_0106.py
+33
-0
configs/models/others/hf_openchat_35_1210.py
configs/models/others/hf_openchat_35_1210.py
+33
-0
configs/models/others/hf_orionstar_14b_base.py
configs/models/others/hf_orionstar_14b_base.py
+24
-0
configs/models/others/hf_telechat_7b_chat.py
configs/models/others/hf_telechat_7b_chat.py
+1
-1
configs/models/others/hf_yayi2_30b_base.py
configs/models/others/hf_yayi2_30b_base.py
+1
-1
No files found.
configs/models/hf_internlm/hf_internlm2_base_20b.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-base-20b-hf'
,
path
=
"internlm/internlm2-base-20b"
,
tokenizer_path
=
'internlm/internlm2-base-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
min_out_len
=
1
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
)
]
configs/models/hf_internlm/hf_internlm2_base_7b.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-base-7b-hf'
,
path
=
"internlm/internlm2-base-7b"
,
tokenizer_path
=
'internlm/internlm2-base-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
min_out_len
=
1
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/hf_internlm/hf_internlm2_chat_1_8b_sft.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-1.8b-sft-hf'
,
path
=
"internlm/internlm2-chat-1_8b-sft"
,
tokenizer_path
=
'internlm/internlm2-chat-1_8b-sft'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_20b.py
View file @
d34ba111
...
...
@@ -4,7 +4,6 @@ from opencompass.models import HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'<|im_start|>system
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
...
...
configs/models/hf_internlm/hf_internlm2_chat_20b_sft.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-20b-sft-hf'
,
path
=
"internlm/internlm2-chat-20b-sft"
,
tokenizer_path
=
'internlm/internlm2-chat-20b-sft'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_20b_with_system.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'<|im_start|>system
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-20b-hf'
,
path
=
"internlm/internlm2-chat-20b"
,
tokenizer_path
=
'internlm/internlm2-chat-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_7b.py
View file @
d34ba111
...
...
@@ -4,7 +4,6 @@ from opencompass.models import HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'<|im_start|>system
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
...
...
configs/models/hf_internlm/hf_internlm2_chat_7b_sft.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-7b-sft-hf'
,
path
=
"internlm/internlm2-chat-7b-sft"
,
tokenizer_path
=
'internlm/internlm2-chat-7b-sft'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_7b_with_system.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'<|im_start|>user
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'<|im_start|>system
\n
'
,
end
=
'<|im_end|>
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'<|im_start|>assistant
\n
'
,
end
=
'<|im_end|>
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-7b-hf'
,
path
=
"internlm/internlm2-chat-7b"
,
tokenizer_path
=
'internlm/internlm2-chat-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|im_end|>'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_math_20b.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'[UNUSED_TOKEN_146]user
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'[UNUSED_TOKEN_146]assistant
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-20b-hf'
,
path
=
"internlm/internlm2-math-20b"
,
tokenizer_path
=
'internlm/internlm2-math-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'[UNUSED_TOKEN_145]'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_math_20b_with_system.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'[UNUSED_TOKEN_146]user
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'[UNUSED_TOKEN_146]system
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'[UNUSED_TOKEN_146]assistant
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-20b-hf'
,
path
=
"internlm/internlm2-math-20b"
,
tokenizer_path
=
'internlm/internlm2-math-20b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'[UNUSED_TOKEN_145]'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_math_7b.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'[UNUSED_TOKEN_146]user
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'[UNUSED_TOKEN_146]assistant
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-7b-hf'
,
path
=
"internlm/internlm2-math-7b"
,
tokenizer_path
=
'internlm/internlm2-math-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'[UNUSED_TOKEN_145]'
,
)
]
configs/models/hf_internlm/hf_internlm2_chat_math_7b_with_system.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
'HUMAN'
,
begin
=
'[UNUSED_TOKEN_146]user
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'SYSTEM'
,
begin
=
'[UNUSED_TOKEN_146]system
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
),
dict
(
role
=
'BOT'
,
begin
=
'[UNUSED_TOKEN_146]assistant
\n
'
,
end
=
'[UNUSED_TOKEN_145]
\n
'
,
generate
=
True
),
],
eos_token_id
=
92542
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'internlm2-chat-math-7b-hf'
,
path
=
"internlm/internlm2-math-7b"
,
tokenizer_path
=
'internlm/internlm2-math-7b'
,
model_kwargs
=
dict
(
trust_remote_code
=
True
,
device_map
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
use_fast
=
False
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'[UNUSED_TOKEN_145]'
,
)
]
configs/models/nanbeige/hf_nanbeige_16b_chat.py
View file @
d34ba111
...
...
@@ -19,16 +19,17 @@ models = [
torch_dtype
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'
righ
t'
,
padding_side
=
'
lef
t'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
batch_padding
=
False
,
max_out_len
=
10
24
,
max_out_len
=
10
0
,
max_seq_len
=
4096
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'</s>'
,
)
]
configs/models/nanbeige/hf_nanbeige_16b_chat_32k.py
View file @
d34ba111
...
...
@@ -19,16 +19,16 @@ models = [
torch_dtype
=
'auto'
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'
righ
t'
,
padding_side
=
'
lef
t'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
batch_padding
=
False
,
max_out_len
=
1024
,
max_seq_len
=
8192
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
end_str
=
'</s>'
,
)
]
configs/models/others/hf_openchat_35_0106.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'GPT4 Correct User: '
,
end
=
'<|end_of_turn|>'
),
dict
(
role
=
"BOT"
,
begin
=
"GPT4 Correct Assistant: "
,
end
=
'<|end_of_turn|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
abbr
=
'openchat-3.5-0106-hf'
,
type
=
HuggingFaceCausalLM
,
path
=
'openchat/openchat-3.5-0106'
,
tokenizer_path
=
'openchat/openchat-3.5-0106'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|end_of_turn|>'
,
)
]
configs/models/others/hf_openchat_35_1210.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'GPT4 Correct User: '
,
end
=
'<|end_of_turn|>'
),
dict
(
role
=
"BOT"
,
begin
=
"GPT4 Correct Assistant: "
,
end
=
'<|end_of_turn|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
abbr
=
'openchat-3.5-1210-hf'
,
type
=
HuggingFaceCausalLM
,
path
=
'openchat/openchat-3.5-1210'
,
tokenizer_path
=
'openchat/openchat-3.5-1210'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
end_str
=
'<|end_of_turn|>'
,
)
]
configs/models/others/hf_orionstar_14b_base.py
0 → 100644
View file @
d34ba111
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
abbr
=
'orionstar-14b-base-hf'
,
type
=
HuggingFaceCausalLM
,
path
=
'OrionStarAI/Orion-14B-Base'
,
tokenizer_path
=
'OrionStarAI/Orion-14B-Base'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
),
max_out_len
=
100
,
min_out_len
=
1
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
2
,
num_procs
=
1
),
)
]
configs/models/others/hf_telechat_7b_chat.py
View file @
d34ba111
...
...
@@ -11,7 +11,7 @@ _meta_template = dict(
models
=
[
dict
(
abbr
=
'telechat-7b-hf'
,
abbr
=
'telechat-7b-hf
--rerun
'
,
type
=
HuggingFaceCausalLM
,
path
=
'Tele-AI/telechat-7B'
,
tokenizer_path
=
'Tele-AI/telechat-7B'
,
...
...
configs/models/others/hf_yayi2_30b_base.py
View file @
d34ba111
...
...
@@ -17,7 +17,7 @@ models = [
trust_remote_code
=
True
,
),
max_out_len
=
100
,
min_out_len
=
3
,
min_out_len
=
1
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
...
...
Prev
1
2
3
4
5
Next
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment