1---
2version: 2
3interactions:
4- id: 0
5 request:
6 proto: HTTP/1.1
7 proto_major: 1
8 proto_minor: 1
9 content_length: 205
10 host: ""
11 body: '{"max_output_tokens":4000,"store":false,"input":[{"content":"You are a helpful assistant","role":"system"},{"content":[{"text":"Say hi in Portuguese","type":"input_text"}],"role":"user"}],"model":"gpt-4o"}'
12 headers:
13 Accept:
14 - application/json
15 Content-Type:
16 - application/json
17 User-Agent:
18 - Charm-Fantasy/0.12.2 (https://charm.land/fantasy)
19 url: https://api.openai.com/v1/responses
20 method: POST
21 response:
22 proto: HTTP/2.0
23 proto_major: 2
24 proto_minor: 0
25 content_length: -1
26 uncompressed: true
27 body: |-
28 {
29 "id": "resp_0e7dfdc132fa96a90169b3014c5470819eadf32151a0dbfd15",
30 "object": "response",
31 "created_at": 1773338956,
32 "status": "completed",
33 "background": false,
34 "billing": {
35 "payer": "developer"
36 },
37 "completed_at": 1773338957,
38 "error": null,
39 "frequency_penalty": 0.0,
40 "incomplete_details": null,
41 "instructions": null,
42 "max_output_tokens": 4000,
43 "max_tool_calls": null,
44 "model": "gpt-4o-2024-08-06",
45 "output": [
46 {
47 "id": "msg_0e7dfdc132fa96a90169b3014db10c819ea5b69df79c7f2810",
48 "type": "message",
49 "status": "completed",
50 "content": [
51 {
52 "type": "output_text",
53 "annotations": [],
54 "logprobs": [],
55 "text": "Oi!"
56 }
57 ],
58 "role": "assistant"
59 }
60 ],
61 "parallel_tool_calls": true,
62 "presence_penalty": 0.0,
63 "previous_response_id": null,
64 "prompt_cache_key": null,
65 "prompt_cache_retention": null,
66 "reasoning": {
67 "effort": null,
68 "summary": null
69 },
70 "safety_identifier": null,
71 "service_tier": "default",
72 "store": false,
73 "temperature": 1.0,
74 "text": {
75 "format": {
76 "type": "text"
77 },
78 "verbosity": "medium"
79 },
80 "tool_choice": "auto",
81 "tools": [],
82 "top_logprobs": 0,
83 "top_p": 1.0,
84 "truncation": "disabled",
85 "usage": {
86 "input_tokens": 20,
87 "input_tokens_details": {
88 "cached_tokens": 0
89 },
90 "output_tokens": 3,
91 "output_tokens_details": {
92 "reasoning_tokens": 0
93 },
94 "total_tokens": 23
95 },
96 "user": null,
97 "metadata": {}
98 }
99 headers:
100 Content-Type:
101 - application/json
102 status: 200 OK
103 code: 200
104 duration: 2.574510083s