b802ccd289 | ||
---|---|---|
.. | ||
README.md | ||
README_en.md | ||
cli_demo_tool.py | ||
openai_api_demo.bak.py | ||
openai_api_demo.py | ||
tool_register.py |
README_en.md
Tool Invocation
This document will introduce how to use the ChatGLM3-6B for tool invocation. Currently, only the ChatGLM3-6B model supports tool invocation, while the ChatGLM3-6B-Base and ChatGLM3-6B-32K models do not support it.
Building System Prompt
Here are two examples of tool invocation. First, prepare the description information of the data to be built.
tools = [
{
"name": "track",
"description": "Track the real-time price of a specified stock",
"parameters": {
"type": "object",
"properties": {
"symbol": {
"description": "The stock code that needs to be tracked"
}
},
"required": ['symbol']
}
},
{
"name": "text-to-speech",
"description": "Convert text to speech",
"parameters": {
"type": "object",
"properties": {
"text": {
"description": "The text that needs to be converted into speech"
},
"voice": {
"description": "The type of voice to use (male, female, etc.)"
},
"speed": {
"description": "The speed of the speech (fast, medium, slow, etc.)"
}
},
"required": ['text']
}
}
]
system_info = {"role": "system", "content": "Answer the following questions as best as you can. You have access to the following tools:", "tools": tools}
Please ensure that the definition format of the tool is consistent with the example to obtain optimal performance.
Asking Questions
Note: Currently, the tool invocation of ChatGLM3-6B only supports the chat
method and does not support the stream_chat
method.
history = [system_info]
query = "Help me inquire the price of stock 10111"
response, history = model.chat(tokenizer, query, history=history)
print(response)
The expected output here is
{"name": "track", "parameters": {"symbol": "10111"}}
This indicates that the model needs to call the tool track
, and the parameter symbol
needs to be passed in.
Invoke Tool, Generate Response
Here, you need to implement the logic of calling the tool yourself. Assuming that the return result has been obtained, return the result to the model in json format and get a response.
result = json.dumps({"price": 12412}, ensure_ascii=False)
response, history = model.chat(tokenizer, result, history=history, role="observation")
print(response)
Here role="observation"
indicates that the input is the return value of the tool invocation rather than user input, and it cannot be omitted.
The expected output is
Based on your query, after the API call, the price of stock 10111 is 12412.
This indicates that this tool invocation has ended, and the model generates a response based on the return result. For more complex questions, the model may need to make multiple tool invocations. At this time, you can judge whether the returned response
is str
or dict
to determine whether the return is a generated response or a tool invocation request.