thinking.yaml

  1---
  2version: 2
  3interactions:
  4- id: 0
  5  request:
  6    proto: HTTP/1.1
  7    proto_major: 1
  8    proto_minor: 1
  9    content_length: 552
 10    host: generativelanguage.googleapis.com
 11    body: |
 12      {"contents":[{"parts":[{"text":"What's the weather in Florence, Italy?"}],"role":"user"}],"generationConfig":{"thinkingConfig":{"includeThoughts":true,"thinkingLevel":"HIGH"}},"systemInstruction":{"parts":[{"text":"You are a helpful assistant"}],"role":"user"},"toolConfig":{"functionCallingConfig":{"mode":"AUTO"}},"tools":[{"functionDeclarations":[{"description":"Get weather information for a location","name":"weather","parameters":{"properties":{"location":{"description":"the city","type":"STRING"}},"required":["location"],"type":"OBJECT"}}]}]}
 13    headers:
 14      Content-Type:
 15      - application/json
 16      User-Agent:
 17      - google-genai-sdk/1.47.0 gl-go/go1.26.0
 18    url: https://generativelanguage.googleapis.com/v1beta/models/gemini-3-pro-preview:generateContent
 19    method: POST
 20  response:
 21    proto: HTTP/2.0
 22    proto_major: 2
 23    proto_minor: 0
 24    content_length: -1
 25    uncompressed: true
 26    body: |
 27      {
 28        "candidates": [
 29          {
 30            "content": {
 31              "parts": [
 32                {
 33                  "text": "**My Approach to Gathering Weather Information**\n\nOkay, so the user wants the weather for Florence, Italy. My primary task here is to find the most efficient way to get that information. I immediately recognize the need for a weather data retrieval tool; that's the obvious starting point. I quickly scan my available tools and, from what I see, the `weather` tool seems like the most relevant option. That's a good sign.\n\nNow, I need to ensure I'm using this tool correctly. It takes a `location` parameter, which is exactly what I need to specify the city. Excellent, that aligns perfectly with the user's request for Florence. Therefore, the logical action is to call the `weather` tool and pass \"Florence, Italy\" as the value for the `location` parameter. The structure will be `weather(location=\"Florence, Italy\")`. That should provide the user with the weather details they're after.\n\n\n",
 34                  "thought": true
 35                },
 36                {
 37                  "functionCall": {
 38                    "name": "weather",
 39                    "args": {
 40                      "location": "Florence, Italy"
 41                    }
 42                  },
 43                  "thoughtSignature": "EoMDCoADAb4+9vvcql04LjFTQa1nTMMudhrlpmFSqgQWtecZs9Gzm8GC6izmq42etV5ICgg5qr/h5KO05LHjo7OT/Y/O6zTq5vweOOvvAZHK3IDBgU7FBiCW5bW+nI6TjhgDHSNPn/Uk14iuv5RGXrH1QgxQ2ApBiLQlEmkjzmzmBUgI0gRmjRX+sDMF0IJu66tmZfkaqvR9Tp2xxEZjJitWL7P6jZvql6JE54UmL8dI02Dk/xnNS76rZiipzg8013aXxfIEgTEHMnsRvzK7QCGaukQ+/DU8GnomhZRYZPEW32OxQEjUyqm10c8o62WPtThyfIUmUBHwc8oojkH/2IT5fRy2nJwvUar3O82/of2miqe/9z7TNTKNK83wl9Gy5R3Gbc/XjI7A+LrwlYTynsVRqXFg3YURlzRnfFcGt4KEUcd2wLKqpeCNSngzS0VgiEqK5j9RTPedm3S2/oTgtl8jt2gff3NEDkJe8xUlXLS+l28rJ22Loc9BooDebKZYBBHvL9qu"
 44                }
 45              ],
 46              "role": "model"
 47            },
 48            "finishReason": "STOP",
 49            "index": 0,
 50            "finishMessage": "Model generated function call(s)."
 51          }
 52        ],
 53        "usageMetadata": {
 54          "promptTokenCount": 65,
 55          "candidatesTokenCount": 16,
 56          "totalTokenCount": 161,
 57          "promptTokensDetails": [
 58            {
 59              "modality": "TEXT",
 60              "tokenCount": 65
 61            }
 62          ],
 63          "thoughtsTokenCount": 80
 64        },
 65        "modelVersion": "gemini-3-pro-preview",
 66        "responseId": "cf-dacuQCbPoz7IPiZ25mAE"
 67      }
 68    headers:
 69      Content-Type:
 70      - application/json; charset=UTF-8
 71    status: 200 OK
 72    code: 200
 73    duration: 4.104216583s
 74- id: 1
 75  request:
 76    proto: HTTP/1.1
 77    proto_major: 1
 78    proto_minor: 1
 79    content_length: 1308
 80    host: generativelanguage.googleapis.com
 81    body: |
 82      {"contents":[{"parts":[{"text":"What's the weather in Florence, Italy?"}],"role":"user"},{"parts":[{"functionCall":{"args":{"location":"Florence, Italy"},"id":"1","name":"weather"},"thoughtSignature":"EoMDCoADAb4+9vvcql04LjFTQa1nTMMudhrlpmFSqgQWtecZs9Gzm8GC6izmq42etV5ICgg5qr/h5KO05LHjo7OT/Y/O6zTq5vweOOvvAZHK3IDBgU7FBiCW5bW+nI6TjhgDHSNPn/Uk14iuv5RGXrH1QgxQ2ApBiLQlEmkjzmzmBUgI0gRmjRX+sDMF0IJu66tmZfkaqvR9Tp2xxEZjJitWL7P6jZvql6JE54UmL8dI02Dk/xnNS76rZiipzg8013aXxfIEgTEHMnsRvzK7QCGaukQ+/DU8GnomhZRYZPEW32OxQEjUyqm10c8o62WPtThyfIUmUBHwc8oojkH/2IT5fRy2nJwvUar3O82/of2miqe/9z7TNTKNK83wl9Gy5R3Gbc/XjI7A+LrwlYTynsVRqXFg3YURlzRnfFcGt4KEUcd2wLKqpeCNSngzS0VgiEqK5j9RTPedm3S2/oTgtl8jt2gff3NEDkJe8xUlXLS+l28rJ22Loc9BooDebKZYBBHvL9qu"}],"role":"model"},{"parts":[{"functionResponse":{"id":"1","name":"weather","response":{"result":"40 C"}}}],"role":"user"}],"generationConfig":{"thinkingConfig":{"includeThoughts":true,"thinkingLevel":"HIGH"}},"systemInstruction":{"parts":[{"text":"You are a helpful assistant"}],"role":"user"},"toolConfig":{"functionCallingConfig":{"mode":"AUTO"}},"tools":[{"functionDeclarations":[{"description":"Get weather information for a location","name":"weather","parameters":{"properties":{"location":{"description":"the city","type":"STRING"}},"required":["location"],"type":"OBJECT"}}]}]}
 83    headers:
 84      Content-Type:
 85      - application/json
 86      User-Agent:
 87      - google-genai-sdk/1.47.0 gl-go/go1.26.0
 88    url: https://generativelanguage.googleapis.com/v1beta/models/gemini-3-pro-preview:generateContent
 89    method: POST
 90  response:
 91    proto: HTTP/2.0
 92    proto_major: 2
 93    proto_minor: 0
 94    content_length: -1
 95    uncompressed: true
 96    body: |
 97      {
 98        "candidates": [
 99          {
100            "content": {
101              "parts": [
102                {
103                  "text": "The weather in Florence, Italy is currently 40 degrees Celsius."
104                }
105              ],
106              "role": "model"
107            },
108            "finishReason": "STOP",
109            "index": 0
110          }
111        ],
112        "usageMetadata": {
113          "promptTokenCount": 175,
114          "candidatesTokenCount": 14,
115          "totalTokenCount": 189,
116          "promptTokensDetails": [
117            {
118              "modality": "TEXT",
119              "tokenCount": 175
120            }
121          ]
122        },
123        "modelVersion": "gemini-3-pro-preview",
124        "responseId": "c_-daaKZE47oz7IP_YXG8AY"
125      }
126    headers:
127      Content-Type:
128      - application/json; charset=UTF-8
129    status: 200 OK
130    code: 200
131    duration: 2.255511958s