1---
2version: 2
3interactions:
4- id: 0
5 request:
6 proto: HTTP/1.1
7 proto_major: 1
8 proto_minor: 1
9 content_length: 550
10 host: generativelanguage.googleapis.com
11 body: |
12 {"contents":[{"parts":[{"text":"What's the weather in Florence, Italy?"}],"role":"user"}],"generationConfig":{"thinkingConfig":{"includeThoughts":true,"thinkingBudget":128}},"systemInstruction":{"parts":[{"text":"You are a helpful assistant"}],"role":"user"},"toolConfig":{"functionCallingConfig":{"mode":"AUTO"}},"tools":[{"functionDeclarations":[{"description":"Get weather information for a location","name":"weather","parameters":{"properties":{"location":{"description":"the city","type":"STRING"}},"required":["location"],"type":"OBJECT"}}]}]}
13 headers:
14 Content-Type:
15 - application/json
16 User-Agent:
17 - google-genai-sdk/1.29.0 gl-go/go1.25.1
18 url: https://generativelanguage.googleapis.com/v1beta/models/gemini-2.5-pro:generateContent
19 method: POST
20 response:
21 proto: HTTP/2.0
22 proto_major: 2
23 proto_minor: 0
24 content_length: -1
25 uncompressed: true
26 body: |
27 {
28 "candidates": [
29 {
30 "content": {
31 "parts": [
32 {
33 "text": "**Getting the Weather in Florence**\n\nOkay, so I'm thinking about the weather, specifically in Florence, Italy. That immediately clicks - I know I've got a function, `weather.get_weather`, that's designed for exactly this kind of query. It takes a `location` parameter, which is perfect. Therefore, I'll set that `location` parameter to \"Florence, Italy\", and that should be all I need to get the forecast.\n",
34 "thought": true
35 },
36 {
37 "functionCall": {
38 "name": "weather",
39 "args": {
40 "location": "Florence, Italy"
41 }
42 },
43 "thoughtSignature": "Cp4CAdHtim8JgCsNhv84SCzZ8bo/kRaAn5bY0cEQ0mfo7nl3GCS7sRlN2T7G4P/Epnwc7/9wbdFzcvcdaqNgWk1flXACXm4QluykEizpys7U+TyP8I2vaUUDIl4Tii0yBZNwjaJxmYp0yb5E9LjpsuEFolG0Ao2vqoJar/0XgY5bzViogURnnUHqVzP2euy5PJEdAd+rOqz1hnK4lCumsNfF5vrkx8aeyGL99C1Eu1/ZETTU7TSnVht+z+4qlysMOlhx4ko0anJ5MWzjewwrZBFkMPgzzw0zVmsHZg9d4t0Hnhmk6Ft7Ge6JOhvI54NgL8MkiviZNvIYT6iHNRqSd71uYeCdCdiw9UJCwASXvbRIv/ktz6OZKrlGld61RP1BXw=="
44 }
45 ],
46 "role": "model"
47 },
48 "finishReason": "STOP",
49 "index": 0,
50 "finishMessage": "Model generated function call(s)."
51 }
52 ],
53 "usageMetadata": {
54 "promptTokenCount": 54,
55 "candidatesTokenCount": 15,
56 "totalTokenCount": 135,
57 "promptTokensDetails": [
58 {
59 "modality": "TEXT",
60 "tokenCount": 54
61 }
62 ],
63 "thoughtsTokenCount": 66
64 },
65 "modelVersion": "gemini-2.5-pro",
66 "responseId": "29n5aPKhFJrivdIPhYnAuQE"
67 }
68 headers:
69 Content-Type:
70 - application/json; charset=UTF-8
71 status: 200 OK
72 code: 200
73 duration: 3.824523542s
74- id: 1
75 request:
76 proto: HTTP/1.1
77 proto_major: 1
78 proto_minor: 1
79 content_length: 764
80 host: generativelanguage.googleapis.com
81 body: |
82 {"contents":[{"parts":[{"text":"What's the weather in Florence, Italy?"}],"role":"user"},{"parts":[{"functionCall":{"args":{"location":"Florence, Italy"},"id":"1","name":"weather"}}],"role":"model"},{"parts":[{"functionResponse":{"id":"1","name":"weather","response":{"result":"40 C"}}}],"role":"user"}],"generationConfig":{"thinkingConfig":{"includeThoughts":true,"thinkingBudget":128}},"systemInstruction":{"parts":[{"text":"You are a helpful assistant"}],"role":"user"},"toolConfig":{"functionCallingConfig":{"mode":"AUTO"}},"tools":[{"functionDeclarations":[{"description":"Get weather information for a location","name":"weather","parameters":{"properties":{"location":{"description":"the city","type":"STRING"}},"required":["location"],"type":"OBJECT"}}]}]}
83 headers:
84 Content-Type:
85 - application/json
86 User-Agent:
87 - google-genai-sdk/1.29.0 gl-go/go1.25.1
88 url: https://generativelanguage.googleapis.com/v1beta/models/gemini-2.5-pro:generateContent
89 method: POST
90 response:
91 proto: HTTP/2.0
92 proto_major: 2
93 proto_minor: 0
94 content_length: -1
95 uncompressed: true
96 body: |
97 {
98 "candidates": [
99 {
100 "content": {
101 "parts": [
102 {
103 "text": "The weather in Florence, Italy is 40 C. \n"
104 }
105 ],
106 "role": "model"
107 },
108 "finishReason": "STOP",
109 "index": 0
110 }
111 ],
112 "usageMetadata": {
113 "promptTokenCount": 84,
114 "candidatesTokenCount": 12,
115 "totalTokenCount": 96,
116 "promptTokensDetails": [
117 {
118 "modality": "TEXT",
119 "tokenCount": 84
120 }
121 ]
122 },
123 "modelVersion": "gemini-2.5-pro",
124 "responseId": "3Nn5aMLPIMahxN8P5c69gQc"
125 }
126 headers:
127 Content-Type:
128 - application/json; charset=UTF-8
129 status: 200 OK
130 code: 200
131 duration: 1.228113333s