Unverified Commit 32f40a8f authored by Fengzhe Zhou's avatar Fengzhe Zhou Committed by GitHub
Browse files

[Sync] Sync with internal codes 2023.01.08 (#777)

parent 8194199d
...@@ -16,6 +16,6 @@ models = [ ...@@ -16,6 +16,6 @@ models = [
batch_size=8, batch_size=8,
model_kwargs=dict(device_map='auto'), model_kwargs=dict(device_map='auto'),
batch_padding=False, # if false, inference with for-loop without batch padding batch_padding=False, # if false, inference with for-loop without batch padding
run_cfg=dict(num_gpus=8, num_procs=1), run_cfg=dict(num_gpus=4, num_procs=1),
) )
] ]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin=' [INST] ', end=' [/INST] '),
dict(role="BOT", begin='', end='', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='llama-2-70b-chat-hf',
path="meta-llama/Llama-2-70b-chat-hf",
tokenizer_path='meta-llama/Llama-2-70b-chat-hf',
model_kwargs=dict(
device_map='auto'
),
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
use_fast=False,
),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
run_cfg=dict(num_gpus=4, num_procs=1),
end_str='[INST]',
)
]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin=' [INST] ', end=' [/INST] '),
dict(role="BOT", begin='', end='', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='llama-2-7b-chat-hf',
path="meta-llama/Llama-2-7b-chat-hf",
tokenizer_path='meta-llama/Llama-2-7b-chat-hf',
model_kwargs=dict(
device_map='auto'
),
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
use_fast=False,
),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
run_cfg=dict(num_gpus=1, num_procs=1),
end_str='[INST]',
)
]
...@@ -16,6 +16,6 @@ models = [ ...@@ -16,6 +16,6 @@ models = [
batch_size=8, batch_size=8,
model_kwargs=dict(device_map='auto'), model_kwargs=dict(device_map='auto'),
batch_padding=False, # if false, inference with for-loop without batch padding batch_padding=False, # if false, inference with for-loop without batch padding
run_cfg=dict(num_gpus=8, num_procs=1), run_cfg=dict(num_gpus=4, num_procs=1),
) )
] ]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='\n<|im_start|>user\n', end='<|im_end|>'),
dict(role="BOT", begin="\n<|im_start|>assistant\n", end='<|im_end|>', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='lemur-70b-chat-v1',
path="OpenLemur/lemur-70b-chat-v1",
tokenizer_path='OpenLemur/lemur-70b-chat-v1',
# tokenizer_kwargs=dict(
# padding_side='left',
# truncation_side='left',
# trust_remote_code=True,
# use_fast=False,),
# pad_token_id=151643,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
meta_template=_meta_template,
model_kwargs=dict(device_map='auto', trust_remote_code=True),
run_cfg=dict(num_gpus=4, num_procs=1),
)
]
...@@ -30,5 +30,6 @@ models = [ ...@@ -30,5 +30,6 @@ models = [
max_seq_len=2048, max_seq_len=2048,
batch_size=8, batch_size=8,
run_cfg=dict(num_gpus=1, num_procs=1), run_cfg=dict(num_gpus=1, num_procs=1),
end_str='</s>',
) )
] ]
from opencompass.models import VLLM
_meta_template = dict(
begin="<s>",
round=[
dict(role="HUMAN", begin='[INST]', end='[/INST]'),
dict(role="BOT", begin="", end='</s>', generate=True),
],
eos_token_id=2
)
models = [
dict(
type=VLLM,
abbr='mistral-7b-instruct-v0.2-vllm',
path='mistralai/Mistral-7B-Instruct-v0.2',
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='</s>',
run_cfg=dict(num_gpus=1, num_procs=1),
)
]
...@@ -29,6 +29,7 @@ models = [ ...@@ -29,6 +29,7 @@ models = [
max_out_len=100, max_out_len=100,
max_seq_len=2048, max_seq_len=2048,
batch_size=8, batch_size=8,
run_cfg=dict(num_gpus=1, num_procs=1), run_cfg=dict(num_gpus=2, num_procs=1),
end_str='</s>',
) )
] ]
...@@ -19,6 +19,6 @@ models = [ ...@@ -19,6 +19,6 @@ models = [
max_out_len=100, max_out_len=100,
max_seq_len=2048, max_seq_len=2048,
batch_size=8, batch_size=8,
run_cfg=dict(num_gpus=1, num_procs=1), run_cfg=dict(num_gpus=2, num_procs=1),
) )
] ]
from opencompass.models import VLLM
_meta_template = dict(
begin="<s>",
round=[
dict(role="HUMAN", begin='[INST]', end='[/INST]'),
dict(role="BOT", begin="", end='</s>', generate=True),
],
eos_token_id=2
)
models = [
dict(
type=VLLM,
abbr='mixtral-8x7b-instruct-v0.1-vllm',
path='mistralai/Mixtral-8x7B-Instruct-v0.1',
model_kwargs=dict(tensor_parallel_size=2),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='</s>',
run_cfg=dict(num_gpus=2, num_procs=1),
)
]
...@@ -30,5 +30,6 @@ models = [ ...@@ -30,5 +30,6 @@ models = [
max_seq_len=2048, max_seq_len=2048,
batch_size=8, batch_size=8,
run_cfg=dict(num_gpus=4, num_procs=1), run_cfg=dict(num_gpus=4, num_procs=1),
end_str='<|endoftext|>',
) )
] ]
from opencompass.models import HuggingFaceCausalLM from opencompass.models import HuggingFaceCausalLM
_meta_template = dict( _meta_template = dict(
round=[ round=[
dict(role="HUMAN", begin='\n<|im_start|>user\n', end='<|im_end|>'), dict(role="HUMAN", begin='\n<|im_start|>user\n', end='<|im_end|>'),
...@@ -28,5 +29,6 @@ models = [ ...@@ -28,5 +29,6 @@ models = [
batch_size=8, batch_size=8,
meta_template=_meta_template, meta_template=_meta_template,
run_cfg=dict(num_gpus=4, num_procs=1), run_cfg=dict(num_gpus=4, num_procs=1),
end_str='<|im_end|>',
) )
] ]
from opencompass.models import VLLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='\n<|im_start|>user\n', end='<|im_end|>'),
dict(role="BOT", begin="\n<|im_start|>assistant\n", end='<|im_end|>', generate=True),
],
)
models = [
dict(
type=VLLM,
abbr='qwen-72b-chat-vllm',
path="Qwen/Qwen-72B-Chat",
model_kwargs=dict(tensor_parallel_size=4),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='<|im_end|>',
run_cfg=dict(num_gpus=4, num_procs=1),
)
]
from opencompass.models import HuggingFaceCausalLM from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='USER: '),
dict(role="BOT", begin=" ASSISTANT:", end='</s>', generate=True),
],
)
models = [ models = [
dict( dict(
...@@ -12,12 +18,13 @@ models = [ ...@@ -12,12 +18,13 @@ models = [
truncation_side='left', truncation_side='left',
use_fast=False, use_fast=False,
), ),
meta_template=_meta_template,
max_out_len=100, max_out_len=100,
max_seq_len=8192, max_seq_len=8192,
batch_size=8, batch_size=8,
model_kwargs=dict(device_map='auto'), model_kwargs=dict(device_map='auto'),
batch_padding=False, # if false, inference with for-loop without batch padding batch_padding=False, # if false, inference with for-loop without batch padding
use_fastchat_template=True, run_cfg=dict(num_gpus=2, num_procs=1),
run_cfg=dict(num_gpus=2, num_procs=1) end_str='</s>',
) )
] ]
from opencompass.models import HuggingFaceCausalLM from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='USER: '),
dict(role="BOT", begin=" ASSISTANT:", end='</s>', generate=True),
],
)
models = [ models = [
dict( dict(
...@@ -12,12 +18,13 @@ models = [ ...@@ -12,12 +18,13 @@ models = [
truncation_side='left', truncation_side='left',
use_fast=False, use_fast=False,
), ),
meta_template=_meta_template,
max_out_len=100, max_out_len=100,
max_seq_len=8192, max_seq_len=8192,
batch_size=8, batch_size=8,
model_kwargs=dict(device_map='auto'), model_kwargs=dict(device_map='auto'),
batch_padding=False, # if false, inference with for-loop without batch padding batch_padding=False, # if false, inference with for-loop without batch padding
use_fastchat_template=True, run_cfg=dict(num_gpus=1, num_procs=1),
run_cfg=dict(num_gpus=1, num_procs=1) end_str='</s>',
) )
] ]
from opencompass.models import VLLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='USER: '),
dict(role="BOT", begin=" ASSISTANT:", end='</s>', generate=True),
],
)
models = [
dict(
type=VLLM,
abbr='vicuna-13b-v1.5-16k-vllm',
path="lmsys/vicuna-13b-v1.5-16k",
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='</s>',
run_cfg=dict(num_gpus=2, num_procs=1),
)
]
from opencompass.models import VLLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='USER: '),
dict(role="BOT", begin=" ASSISTANT:", end='</s>', generate=True),
],
)
models = [
dict(
type=VLLM,
abbr='vicuna-7b-v1.5-16k-vllm',
path="lmsys/vicuna-7b-v1.5-16k",
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=32,
generation_kwargs=dict(temperature=0),
end_str='</s>',
run_cfg=dict(num_gpus=1, num_procs=1),
)
]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role="HUMAN", begin='USER: ', end=' '),
dict(role="BOT", begin="ASSISTANT: ", end='</s>', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='wizardlm-13b-v1.2-hf',
path='WizardLM/WizardLM-13B-V1.2',
tokenizer_path='WizardLM/WizardLM-13B-V1.2',
model_kwargs=dict(
device_map='auto',
trust_remote_code=True,
),
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
trust_remote_code=True,
),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
run_cfg=dict(num_gpus=2, num_procs=1),
end_str='</s>',
)
]
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment