Unverified Commit 2c915218 authored by Leymore's avatar Leymore Committed by GitHub
Browse files

[Feaure] Add new models: baichuan2, tigerbot, vicuna v1.5 (#373)

* add bag of new models: baichuan2, tigerbot, vicuna v1.5

* update

* re-organize models

* update readme

* update
parent b48d0840
...@@ -8,7 +8,7 @@ with read_base(): ...@@ -8,7 +8,7 @@ with read_base():
from .datasets.collections.chat_medium import datasets from .datasets.collections.chat_medium import datasets
# and output the results in a choosen format # and output the results in a choosen format
from .summarizers.medium import summarizer from .summarizers.medium import summarizer
from .models.claude import models from .models.claude.claude import models
infer = dict( infer = dict(
partitioner=dict(type=NaivePartitioner), partitioner=dict(type=NaivePartitioner),
......
...@@ -2,6 +2,6 @@ from mmengine.config import read_base ...@@ -2,6 +2,6 @@ from mmengine.config import read_base
with read_base(): with read_base():
from .datasets.humanevalx.humanevalx_gen import humanevalx_datasets from .datasets.humanevalx.humanevalx_gen import humanevalx_datasets
from .models.hf_codegeex2_6b import models from .models.codegeex2.hf_codegeex2_6b import models
datasets = humanevalx_datasets datasets = humanevalx_datasets
...@@ -3,8 +3,8 @@ from mmengine.config import read_base ...@@ -3,8 +3,8 @@ from mmengine.config import read_base
with read_base(): with read_base():
from .datasets.siqa.siqa_gen import siqa_datasets from .datasets.siqa.siqa_gen import siqa_datasets
from .datasets.winograd.winograd_ppl import winograd_datasets from .datasets.winograd.winograd_ppl import winograd_datasets
from .models.hf_opt_125m import opt125m from .models.opt.hf_opt_125m import opt125m
from .models.hf_opt_350m import opt350m from .models.opt.hf_opt_350m import opt350m
datasets = [*siqa_datasets, *winograd_datasets] datasets = [*siqa_datasets, *winograd_datasets]
models = [opt125m, opt350m] models = [opt125m, opt350m]
...@@ -2,7 +2,7 @@ from mmengine.config import read_base ...@@ -2,7 +2,7 @@ from mmengine.config import read_base
with read_base(): with read_base():
from .datasets.collections.base_medium_llama import piqa_datasets, siqa_datasets from .datasets.collections.base_medium_llama import piqa_datasets, siqa_datasets
from .models.hf_llama_7b import models from .models.hf_llama.hf_llama_7b import models
datasets = [*piqa_datasets, *siqa_datasets] datasets = [*piqa_datasets, *siqa_datasets]
...@@ -4,6 +4,6 @@ with read_base(): ...@@ -4,6 +4,6 @@ with read_base():
# choose a list of datasets # choose a list of datasets
from .datasets.collections.base_medium import datasets from .datasets.collections.base_medium import datasets
# choose a model of interest # choose a model of interest
from .models.internlm_7b import models from .models.internlm.internlm_7b import models
# and output the results in a choosen format # and output the results in a choosen format
from .summarizers.medium import summarizer from .summarizers.medium import summarizer
...@@ -4,6 +4,6 @@ with read_base(): ...@@ -4,6 +4,6 @@ with read_base():
# choose a list of datasets # choose a list of datasets
from .datasets.collections.base_medium import datasets from .datasets.collections.base_medium import datasets
# choose a model of interest # choose a model of interest
from .models.hf_internlm_7b import models from .models.hf_internlm.hf_internlm_7b import models
# and output the results in a choosen format # and output the results in a choosen format
from .summarizers.medium import summarizer from .summarizers.medium import summarizer
...@@ -2,7 +2,7 @@ from mmengine.config import read_base ...@@ -2,7 +2,7 @@ from mmengine.config import read_base
with read_base(): with read_base():
from .datasets.collections.base_medium_llama import piqa_datasets, siqa_datasets from .datasets.collections.base_medium_llama import piqa_datasets, siqa_datasets
from .models.llama2_7b import models from .models.llama.llama2_7b import models
datasets = [*piqa_datasets, *siqa_datasets] datasets = [*piqa_datasets, *siqa_datasets]
from mmengine.config import read_base from mmengine.config import read_base
with read_base(): with read_base():
from .models.hf_qwen_7b import models from .models.qwen.hf_qwen_7b import models
from .datasets.collections.leaderboard.qwen import datasets from .datasets.collections.leaderboard.qwen import datasets
from .summarizers.leaderboard import summarizer from .summarizers.leaderboard import summarizer
......
from mmengine.config import read_base from mmengine.config import read_base
with read_base(): with read_base():
from .models.hf_qwen_7b_chat import models from .models.qwen.hf_qwen_7b_chat import models
from .datasets.collections.leaderboard.qwen_chat import datasets from .datasets.collections.leaderboard.qwen_chat import datasets
from .summarizers.leaderboard import summarizer from .summarizers.leaderboard import summarizer
......
from opencompass.models import HuggingFaceCausalLM
models = [
dict(
type=HuggingFaceCausalLM,
abbr='baichuan2-13b-base-hf',
path="baichuan-inc/Baichuan2-13B-Base",
tokenizer_path='baichuan-inc/Baichuan2-13B-Base',
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
trust_remote_code=True,
use_fast=False,
),
max_out_len=100,
max_seq_len=2048,
batch_size=8,
model_kwargs=dict(device_map='auto', trust_remote_code=True),
run_cfg=dict(num_gpus=2, num_procs=1),
)
]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role='HUMAN', begin='<reserved_106>'),
dict(role='BOT', begin='<reserved_107>', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='baichuan2-13b-chat-hf',
path="baichuan-inc/Baichuan2-13B-Chat",
tokenizer_path='baichuan-inc/Baichuan2-13B-Chat',
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
trust_remote_code=True,
use_fast=False,
),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
model_kwargs=dict(device_map='auto', trust_remote_code=True),
run_cfg=dict(num_gpus=2, num_procs=1),
)
]
from opencompass.models import HuggingFaceCausalLM
models = [
dict(
type=HuggingFaceCausalLM,
abbr='baichuan2-7b-base-hf',
path="baichuan-inc/Baichuan2-7B-Base",
tokenizer_path='baichuan-inc/Baichuan2-7B-Base',
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
trust_remote_code=True,
use_fast=False,
),
max_out_len=100,
max_seq_len=2048,
batch_size=8,
model_kwargs=dict(device_map='auto', trust_remote_code=True),
run_cfg=dict(num_gpus=1, num_procs=1),
)
]
from opencompass.models import HuggingFaceCausalLM
_meta_template = dict(
round=[
dict(role='HUMAN', begin='<reserved_106>'),
dict(role='BOT', begin='<reserved_107>', generate=True),
],
)
models = [
dict(
type=HuggingFaceCausalLM,
abbr='baichuan2-7b-chat-hf',
path="baichuan-inc/Baichuan2-7B-Chat",
tokenizer_path='baichuan-inc/Baichuan2-7B-Chat',
tokenizer_kwargs=dict(
padding_side='left',
truncation_side='left',
trust_remote_code=True,
use_fast=False,
),
meta_template=_meta_template,
max_out_len=100,
max_seq_len=2048,
batch_size=8,
model_kwargs=dict(device_map='auto', trust_remote_code=True),
run_cfg=dict(num_gpus=1, num_procs=1),
)
]
Markdown is supported
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment