request_demo.py 2.62 KB
Newer Older
Rayyyyy's avatar
Rayyyyy committed
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
"""
This script is an example of using the OpenAI API to create various interactions with a Yuan2.0 model.
"""

from openai import OpenAI

# import os
# os.environ["no_proxy"] = "localhost,127.0.0.1,::1"

base_url = "http://127.0.0.1:8001/v1/"
client = OpenAI(
    api_key="EMPTY",
    base_url=base_url
)

def function_chat():
    messages = [{"role": "user", "content": "What's the weather like in San Francisco, Tokyo, and Paris?"}]
    tools = [
        {
            "type": "function",
            "function": {
                "name": "get_current_weather",
                "description": "Get the current weather in a given location",
                "parameters": {
                    "type": "object",
                    "properties": {
                        "location": {
                            "type": "string",
                            "description": "The city and state, e.g. San Francisco, CA",
                        },
                        "unit": {"type": "string", "enum": ["celsius", "fahrenheit"]},
                    },
                    "required": ["location"],
                },
            },
        }
    ]

    response = client.chat.completions.create(
        model="yuan2",
        messages=messages,
        tools=tools,
        tool_choice="auto",
    )
    if response:
        content = response.choices[0].message.content
        print(content)
    else:
        print("Error:", response.status_code)


def simple_chat(use_stream=True):
    messages = [
        {
            "role": "system",
            "content": "你是浪潮信息研发的大语言模型。",
        },
        {
            "role": "user",
            "content": "你好,请给我写一首诗,主题是春节。"
        }
    ]
    response = client.chat.completions.create(
        model="Yuan2-2B-Janus-hf".lower(),
        messages=messages,
        stream=use_stream,
        max_tokens=1024,
        temperature=1.0,
        seed=1234,
        top_p=0.9)
    if response:
        if use_stream:
            for chunk in response:
                print(chunk.choices[0].delta.content)
        else:
            content = response.choices[0].message.content
            print(content)
    else:
        print("Error:", response.status_code)


def embedding():
    response = client.embeddings.create(
        model="bge-large-zh-1.5",
        input=["你好,写一个春节晚会致辞,100字左右。"],
    )
    embeddings = response.data[0].embedding
    print(f"embeddings length: {len(embeddings)}")


if __name__ == "__main__":
    simple_chat(use_stream=False)
    simple_chat(use_stream=True)
    embedding()
    # function_chat()