simple_streaming.yaml

 1---
 2version: 2
 3interactions:
 4- id: 0
 5  request:
 6    proto: HTTP/1.1
 7    proto_major: 1
 8    proto_minor: 1
 9    content_length: 202
10    host: generativelanguage.googleapis.com
11    body: |
12      {"contents":[{"parts":[{"text":"Say hi in Portuguese"}],"role":"user"}],"generationConfig":{"maxOutputTokens":4000},"systemInstruction":{"parts":[{"text":"You are a helpful assistant"}],"role":"user"}}
13    form:
14      alt:
15      - sse
16    headers:
17      Content-Type:
18      - application/json
19      User-Agent:
20      - google-genai-sdk/1.29.0 gl-go/go1.25.1
21    url: https://generativelanguage.googleapis.com/v1beta/models/gemini-2.5-pro:streamGenerateContent?alt=sse
22    method: POST
23  response:
24    proto: HTTP/2.0
25    proto_major: 2
26    proto_minor: 0
27    content_length: -1
28    body: "data: {\"candidates\": [{\"content\": {\"parts\": [{\"text\": \"Of course! Here are the most common ways:\\n\\n*   **Olá** (Hello) - This works in any situation,\"}],\"role\": \"model\"},\"index\": 0}],\"usageMetadata\": {\"promptTokenCount\": 11,\"candidatesTokenCount\": 26,\"totalTokenCount\": 840,\"promptTokensDetails\": [{\"modality\": \"TEXT\",\"tokenCount\": 11}],\"thoughtsTokenCount\": 803},\"modelVersion\": \"gemini-2.5-pro\",\"responseId\": \"adH5aP3PAq3d7M8PqtjQ2AI\"}\r\n\r\ndata: {\"candidates\": [{\"content\": {\"parts\": [{\"text\": \" formal or informal.\\n*   **Oi** (Hi) - This is more informal and very common, especially in\"}],\"role\": \"model\"},\"index\": 0}],\"usageMetadata\": {\"promptTokenCount\": 11,\"candidatesTokenCount\": 50,\"totalTokenCount\": 864,\"promptTokensDetails\": [{\"modality\": \"TEXT\",\"tokenCount\": 11}],\"thoughtsTokenCount\": 803},\"modelVersion\": \"gemini-2.5-pro\",\"responseId\": \"adH5aP3PAq3d7M8PqtjQ2AI\"}\r\n\r\ndata: {\"candidates\": [{\"content\": {\"parts\": [{\"text\": \" Brazil.\"}],\"role\": \"model\"},\"finishReason\": \"STOP\",\"index\": 0}],\"usageMetadata\": {\"promptTokenCount\": 11,\"candidatesTokenCount\": 52,\"totalTokenCount\": 866,\"promptTokensDetails\": [{\"modality\": \"TEXT\",\"tokenCount\": 11}],\"thoughtsTokenCount\": 803},\"modelVersion\": \"gemini-2.5-pro\",\"responseId\": \"adH5aP3PAq3d7M8PqtjQ2AI\"}\r\n\r\n"
29    headers:
30      Content-Type:
31      - text/event-stream
32    status: 200 OK
33    code: 200
34    duration: 8.964147375s