thinking.yaml

 1---
 2version: 2
 3interactions:
 4- id: 0
 5  request:
 6    proto: HTTP/1.1
 7    proto_major: 1
 8    proto_minor: 1
 9    content_length: 458
10    host: ""
11    body: '{"messages":[{"content":"You are a helpful assistant","role":"system"},{"content":"What''s the weather in Florence, Italy?","role":"user"}],"model":"glm-4.5","reasoning_effort":"high","tool_choice":"auto","tools":[{"function":{"name":"weather","strict":false,"description":"Get weather information for a location","parameters":{"properties":{"location":{"description":"the city","type":"string"}},"required":["location"],"type":"object"}},"type":"function"}]}'
12    headers:
13      Accept:
14      - application/json
15      Content-Type:
16      - application/json
17      User-Agent:
18      - OpenAI/Go 2.3.0
19    url: https://api.z.ai/api/coding/paas/v4/chat/completions
20    method: POST
21  response:
22    proto: HTTP/2.0
23    proto_major: 2
24    proto_minor: 0
25    content_length: -1
26    uncompressed: true
27    body: '{"choices":[{"finish_reason":"tool_calls","index":0,"message":{"content":"\n\nI''ll check the weather in Florence, Italy for you.\n","reasoning_content":"The user is asking for weather information for Florence, Italy. I have access to a weather function that requires a \"location\" parameter. The user has provided \"Florence, Italy\" as the location, so I have all the required information to make the function call.","role":"assistant","tool_calls":[{"function":{"arguments":"{\"location\": \"Florence, Italy\"}","name":"weather"},"id":"call_-8303617501649735809","index":0,"type":"function"}]}}],"created":1758884808,"id":"20250926190643068def4b4247493f","model":"glm-4.5","request_id":"20250926190643068def4b4247493f","usage":{"completion_tokens":86,"prompt_tokens":179,"prompt_tokens_details":{"cached_tokens":43},"total_tokens":265}}'
28    headers:
29      Content-Type:
30      - application/json; charset=UTF-8
31    status: 200 OK
32    code: 200
33    duration: 4.969463833s
34- id: 1
35  request:
36    proto: HTTP/1.1
37    proto_major: 1
38    proto_minor: 1
39    content_length: 772
40    host: ""
41    body: '{"messages":[{"content":"You are a helpful assistant","role":"system"},{"content":"What''s the weather in Florence, Italy?","role":"user"},{"content":"\n\nI''ll check the weather in Florence, Italy for you.\n","tool_calls":[{"id":"call_-8303617501649735809","function":{"arguments":"{\"location\": \"Florence, Italy\"}","name":"weather"},"type":"function"}],"role":"assistant"},{"content":"40 C","tool_call_id":"call_-8303617501649735809","role":"tool"}],"model":"glm-4.5","reasoning_effort":"high","tool_choice":"auto","tools":[{"function":{"name":"weather","strict":false,"description":"Get weather information for a location","parameters":{"properties":{"location":{"description":"the city","type":"string"}},"required":["location"],"type":"object"}},"type":"function"}]}'
42    headers:
43      Accept:
44      - application/json
45      Content-Type:
46      - application/json
47      User-Agent:
48      - OpenAI/Go 2.3.0
49    url: https://api.z.ai/api/coding/paas/v4/chat/completions
50    method: POST
51  response:
52    proto: HTTP/2.0
53    proto_major: 2
54    proto_minor: 0
55    content_length: -1
56    uncompressed: true
57    body: '{"choices":[{"finish_reason":"stop","index":0,"message":{"content":"\n\nThe weather in Florence, Italy is currently 40°C (104°F). That''s quite hot! Florence typically has warm summers, but this temperature is particularly high even for the region. Make sure to stay hydrated and seek shade if you''re there, as this level of heat can be quite intense.","reasoning_content":"The weather function returned \"40 C\" for Florence, Italy. This seems like a very high temperature (40°C/104°F), which would be extremely hot for Florence. Let me provide this information to the user while noting that this seems unusually warm.","role":"assistant"}}],"created":1758884811,"id":"2025092619064878c373b988234a18","model":"glm-4.5","request_id":"2025092619064878c373b988234a18","usage":{"completion_tokens":116,"prompt_tokens":219,"prompt_tokens_details":{"cached_tokens":43},"total_tokens":335}}'
58    headers:
59      Content-Type:
60      - application/json; charset=UTF-8
61    status: 200 OK
62    code: 200
63    duration: 3.68313875s