lmdeploy_internlm2_chat_7b.py 979 Bytes
Newer Older
1
2
3
4
5
from opencompass.models.turbomind import TurboMindModel


_meta_template = dict(
    round=[
6
7
        dict(role="HUMAN", begin="<|im_start|>user\n", end="<|im_end|>\n"),
        dict(role="BOT", begin="<|im_start|>assistant\n", end="<|im_end|>\n", generate=True),
8
9
10
11
12
13
    ],
)

models = [
    dict(
        type=TurboMindModel,
14
        abbr="internlm2-chat-7b-turbomind",
15
16
        path="internlm/internlm2-chat-7b",
        meta_template=_meta_template,
17
18
19
20
21
22
23
24
25
26
27
28
29
30
        engine_config=dict(
            session_len=210000,
            max_batch_size=8,
            rope_scaling_factor=2.0,
            model_name="internlm2-chat-7b",
            tp=1,
            stop_words=[2, 92542],
        ),
        gen_config=dict(
            top_k=1,
            top_p=0.8,
            temperature=1.0,
            max_new_tokens=2000,
        ),
31
32
        max_out_len=2000,
        max_seq_len=210000,
33
        batch_size=1,
34
35
36
37
        concurrency=8,
        run_cfg=dict(num_gpus=1, num_procs=1),
    )
]