glm4/basic_demo/openai_api_request.py

94 lines
2.9 KiB
Python
Raw Normal View History

2024-06-05 10:22:16 +08:00
"""
This script creates a OpenAI Request demo for the glm-4-9b model, just Use OpenAI API to interact with the model.
"""
from openai import OpenAI
base_url = "http://127.0.0.1:8000/v1/"
client = OpenAI(api_key="EMPTY", base_url=base_url)
def function_chat():
messages = [{"role": "user", "content": "What's the weather like in San Francisco, Tokyo, and Paris?"}]
tools = [
{
"type": "function",
"function": {
"name": "get_current_weather",
2024-06-07 22:14:00 +08:00
"description": "Get the current weather",
2024-06-05 10:22:16 +08:00
"parameters": {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "The city and state, e.g. San Francisco, CA",
},
2024-06-07 22:14:00 +08:00
"format": {
"type": "string",
"enum": ["celsius", "fahrenheit"],
"description": "The temperature unit to use. Infer this from the users location.",
},
2024-06-05 10:22:16 +08:00
},
2024-06-07 22:14:00 +08:00
"required": ["location", "format"],
2024-06-05 10:22:16 +08:00
},
2024-06-07 22:14:00 +08:00
}
},
2024-06-05 10:22:16 +08:00
]
2024-06-07 22:14:00 +08:00
# # All Tools 能力: 绘图
2024-06-05 10:22:16 +08:00
# messages = [{"role": "user", "content": "帮我画一张天空的画画吧"}]
# tools = [{"type": "cogview"}]
#
# All Tools 能力: 联网查询
# messages = [{"role": "user", "content": "今天黄金的价格"}]
# tools = [{"type": "simple_browser"}]
response = client.chat.completions.create(
model="glm-4",
messages=messages,
tools=tools,
2024-06-07 22:14:00 +08:00
stream=False, # must use False
2024-06-05 10:22:16 +08:00
tool_choice="auto", # use "auto" to let the model choose the tool automatically
# tool_choice={"type": "function", "function": {"name": "my_function"}},
)
if response:
2024-06-07 22:14:00 +08:00
print(response.choices[0].message)
2024-06-05 10:22:16 +08:00
else:
print("Error:", response.status_code)
2024-06-07 22:14:00 +08:00
2024-06-05 10:22:16 +08:00
def simple_chat(use_stream=False):
messages = [
{
"role": "system",
2024-06-07 22:14:00 +08:00
"content": "请在你输出的时候都带上“喵喵喵”三个字,放在开头。",
2024-06-05 10:22:16 +08:00
},
{
"role": "user",
2024-06-07 22:14:00 +08:00
"content": "你好,你是谁"
2024-06-05 10:22:16 +08:00
}
]
response = client.chat.completions.create(
model="glm-4",
messages=messages,
stream=use_stream,
2024-06-07 22:14:00 +08:00
max_tokens=256,
temperature=0.1,
2024-06-05 10:22:16 +08:00
presence_penalty=1.1,
top_p=0.8)
if response:
if use_stream:
for chunk in response:
print(chunk.choices[0].delta.content)
else:
content = response.choices[0].message.content
print(content)
else:
print("Error:", response.status_code)
if __name__ == "__main__":
2024-06-07 22:14:00 +08:00
# simple_chat(use_stream=False)
2024-06-05 10:22:16 +08:00
function_chat()