Skip to content
GitLab
Menu
Projects
Groups
Snippets
Loading...
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
Menu
Open sidebar
OpenDAS
opencompass
Commits
3a354bd1
"docs/git@developer.sourcefind.cn:OpenDAS/opencompass.git" did not exist on "3fe5ee096c9a8b08633874438a386bea649fa088"
Unverified
Commit
3a354bd1
authored
Dec 07, 2023
by
Fengzhe Zhou
Committed by
GitHub
Dec 07, 2023
Browse files
add qwen and deepseek configs (#672)
parent
1c95790f
Changes
8
Hide whitespace changes
Inline
Side-by-side
Showing
8 changed files
with
226 additions
and
0 deletions
+226
-0
configs/models/deepseek/hf_deepseek_67b_base.py
configs/models/deepseek/hf_deepseek_67b_base.py
+24
-0
configs/models/deepseek/hf_deepseek_67b_chat.py
configs/models/deepseek/hf_deepseek_67b_chat.py
+32
-0
configs/models/deepseek/hf_deepseek_7b_base.py
configs/models/deepseek/hf_deepseek_7b_base.py
+24
-0
configs/models/deepseek/hf_deepseek_7b_chat.py
configs/models/deepseek/hf_deepseek_7b_chat.py
+32
-0
configs/models/qwen/hf_qwen_1_8b.py
configs/models/qwen/hf_qwen_1_8b.py
+25
-0
configs/models/qwen/hf_qwen_1_8b_chat.py
configs/models/qwen/hf_qwen_1_8b_chat.py
+32
-0
configs/models/qwen/hf_qwen_72b.py
configs/models/qwen/hf_qwen_72b.py
+25
-0
configs/models/qwen/hf_qwen_72b_chat.py
configs/models/qwen/hf_qwen_72b_chat.py
+32
-0
No files found.
configs/models/deepseek/hf_deepseek_67b_base.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'deepseek-67b-base-hf'
,
path
=
"deepseek-ai/deepseek-llm-67b-base"
,
tokenizer_path
=
'deepseek-ai/deepseek-llm-67b-base'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
)
]
configs/models/deepseek/hf_deepseek_67b_chat.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'User: '
,
end
=
'
\n\n
'
),
dict
(
role
=
"BOT"
,
begin
=
"Assistant: "
,
end
=
'<|end▁of▁sentence|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'deepseek-67b-chat-hf'
,
path
=
"deepseek-ai/deepseek-llm-67b-chat"
,
tokenizer_path
=
'deepseek-ai/deepseek-llm-67b-chat'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
)
]
configs/models/deepseek/hf_deepseek_7b_base.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'deepseek-7b-base-hf'
,
path
=
"deepseek-ai/deepseek-llm-7b-base"
,
tokenizer_path
=
'deepseek-ai/deepseek-llm-7b-base'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/deepseek/hf_deepseek_7b_chat.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'User: '
,
end
=
'
\n\n
'
),
dict
(
role
=
"BOT"
,
begin
=
"Assistant: "
,
end
=
'<|end▁of▁sentence|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'deepseek-7b-chat-hf'
,
path
=
"deepseek-ai/deepseek-llm-7b-chat"
,
tokenizer_path
=
'deepseek-ai/deepseek-llm-7b-chat'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
meta_template
=
_meta_template
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen_1_8b.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen-1.8b-hf'
,
path
=
"Qwen/Qwen-1_8B"
,
tokenizer_path
=
'Qwen/Qwen-1_8B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
pad_token_id
=
151643
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen_1_8b_chat.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
"BOT"
,
begin
=
"
\n
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen-1.8b-chat-hf'
,
path
=
"Qwen/Qwen-1_8B-Chat"
,
tokenizer_path
=
'Qwen/Qwen-1_8B-Chat'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,),
pad_token_id
=
151643
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
1
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen_72b.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen-72b-hf'
,
path
=
"Qwen/Qwen-72B"
,
tokenizer_path
=
'Qwen/Qwen-72B'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
,
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,
),
pad_token_id
=
151643
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
)
]
configs/models/qwen/hf_qwen_72b_chat.py
0 → 100644
View file @
3a354bd1
from
opencompass.models
import
HuggingFaceCausalLM
_meta_template
=
dict
(
round
=
[
dict
(
role
=
"HUMAN"
,
begin
=
'
\n
<|im_start|>user
\n
'
,
end
=
'<|im_end|>'
),
dict
(
role
=
"BOT"
,
begin
=
"
\n
<|im_start|>assistant
\n
"
,
end
=
'<|im_end|>'
,
generate
=
True
),
],
)
models
=
[
dict
(
type
=
HuggingFaceCausalLM
,
abbr
=
'qwen-72b-chat-hf'
,
path
=
"Qwen/Qwen-72B-Chat"
,
tokenizer_path
=
'Qwen/Qwen-72B-Chat'
,
model_kwargs
=
dict
(
device_map
=
'auto'
,
trust_remote_code
=
True
),
tokenizer_kwargs
=
dict
(
padding_side
=
'left'
,
truncation_side
=
'left'
,
trust_remote_code
=
True
,
use_fast
=
False
,),
pad_token_id
=
151643
,
max_out_len
=
100
,
max_seq_len
=
2048
,
batch_size
=
8
,
meta_template
=
_meta_template
,
run_cfg
=
dict
(
num_gpus
=
4
,
num_procs
=
1
),
)
]
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment