vllm_qwen1_5_7b_chat.py 328 Bytes
Newer Older
1
2
3
4
5
6
7
8
9
10
11
12
13
from opencompass.models import VLLMwithChatTemplate

models = [
    dict(
        type=VLLMwithChatTemplate,
        abbr='qwen1.5-7b-chat-vllm',
        path='Qwen/Qwen1.5-7B-Chat',
        model_kwargs=dict(tensor_parallel_size=1),
        max_out_len=1024,
        batch_size=32768,
        run_cfg=dict(num_gpus=1),
    )
]