1{
2 "name": "Synthetic",
3 "id": "synthetic",
4 "type": "openai-compat",
5 "api_key": "$SYNTHETIC_API_KEY",
6 "api_endpoint": "https://api.synthetic.new/openai/v1",
7 "default_large_model_id": "hf:zai-org/GLM-4.6",
8 "default_small_model_id": "hf:openai/gpt-oss-120b",
9 "models": [
10 {
11 "id": "hf:deepseek-ai/DeepSeek-R1",
12 "name": "DeepSeek R1",
13 "cost_per_1m_in": 0.55,
14 "cost_per_1m_out": 2.19,
15 "cost_per_1m_in_cached": 0.07,
16 "cost_per_1m_out_cached": 0.14,
17 "context_window": 128000,
18 "default_max_tokens": 65536,
19 "can_reason": true,
20 "reasoning_levels": [
21 "low",
22 "medium",
23 "high"
24 ],
25 "default_reasoning_effort": "medium",
26 "supports_attachments": false
27 },
28 {
29 "id": "hf:deepseek-ai/DeepSeek-R1-0528",
30 "name": "DeepSeek R1 0528",
31 "cost_per_1m_in": 3.0,
32 "cost_per_1m_out": 8.0,
33 "cost_per_1m_in_cached": 0.07,
34 "cost_per_1m_out_cached": 0.14,
35 "context_window": 128000,
36 "default_max_tokens": 65536,
37 "can_reason": true,
38 "reasoning_levels": [
39 "low",
40 "medium",
41 "high"
42 ],
43 "default_reasoning_effort": "medium",
44 "supports_attachments": false
45 },
46 {
47 "id": "hf:deepseek-ai/DeepSeek-V3",
48 "name": "DeepSeek V3",
49 "cost_per_1m_in": 1.25,
50 "cost_per_1m_out": 1.25,
51 "cost_per_1m_in_cached": 0.07,
52 "cost_per_1m_out_cached": 0.14,
53 "context_window": 128000,
54 "default_max_tokens": 8192,
55 "can_reason": false,
56 "supports_attachments": false
57 },
58 {
59 "id": "hf:deepseek-ai/DeepSeek-V3-0324",
60 "name": "DeepSeek V3 0324",
61 "cost_per_1m_in": 1.2,
62 "cost_per_1m_out": 1.2,
63 "cost_per_1m_in_cached": 0.07,
64 "cost_per_1m_out_cached": 0.14,
65 "context_window": 128000,
66 "default_max_tokens": 8192,
67 "can_reason": false,
68 "supports_attachments": false
69 },
70 {
71 "id": "hf:deepseek-ai/DeepSeek-V3.1",
72 "name": "DeepSeek V3.1",
73 "cost_per_1m_in": 0.56,
74 "cost_per_1m_out": 1.68,
75 "cost_per_1m_in_cached": 0.07,
76 "cost_per_1m_out_cached": 0.14,
77 "context_window": 131072,
78 "default_max_tokens": 8192,
79 "can_reason": true,
80 "reasoning_levels": [
81 "low",
82 "medium",
83 "high"
84 ],
85 "default_reasoning_effort": "medium",
86 "supports_attachments": false
87 },
88 {
89 "id": "hf:deepseek-ai/DeepSeek-V3.1-Terminus",
90 "name": "DeepSeek V3.1 Terminus",
91 "cost_per_1m_in": 1.2,
92 "cost_per_1m_out": 1.2,
93 "cost_per_1m_in_cached": 0.07,
94 "cost_per_1m_out_cached": 0.14,
95 "context_window": 128000,
96 "default_max_tokens": 8192,
97 "can_reason": true,
98 "reasoning_levels": [
99 "low",
100 "medium",
101 "high"
102 ],
103 "default_reasoning_effort": "medium",
104 "supports_attachments": false
105 },
106 {
107 "id": "hf:meta-llama/Llama-3.1-405B-Instruct",
108 "name": "Llama 3.1 405B Instruct",
109 "cost_per_1m_in": 3.0,
110 "cost_per_1m_out": 3.0,
111 "cost_per_1m_in_cached": 0.27,
112 "cost_per_1m_out_cached": 0.55,
113 "context_window": 131072,
114 "default_max_tokens": 4096,
115 "can_reason": true,
116 "reasoning_levels": [
117 "low",
118 "medium",
119 "high"
120 ],
121 "default_reasoning_effort": "medium",
122 "supports_attachments": false
123 },
124 {
125 "id": "hf:meta-llama/Llama-3.1-70B-Instruct",
126 "name": "Llama 3.1 70B Instruct",
127 "cost_per_1m_in": 0.9,
128 "cost_per_1m_out": 0.9,
129 "cost_per_1m_in_cached": 0.59,
130 "cost_per_1m_out_cached": 1.1,
131 "context_window": 131072,
132 "default_max_tokens": 4096,
133 "can_reason": false,
134 "supports_attachments": false
135 },
136 {
137 "id": "hf:meta-llama/Llama-3.1-8B-Instruct",
138 "name": "Llama 3.1 8B Instruct",
139 "cost_per_1m_in": 0.2,
140 "cost_per_1m_out": 0.2,
141 "cost_per_1m_in_cached": 0.07,
142 "cost_per_1m_out_cached": 0.2,
143 "context_window": 128000,
144 "default_max_tokens": 4096,
145 "can_reason": false,
146 "supports_attachments": false
147 },
148 {
149 "id": "hf:meta-llama/Llama-3.3-70B-Instruct",
150 "name": "Llama 3.3 70B Instruct",
151 "cost_per_1m_in": 0.9,
152 "cost_per_1m_out": 0.9,
153 "cost_per_1m_in_cached": 0.59,
154 "cost_per_1m_out_cached": 1.1,
155 "context_window": 128000,
156 "default_max_tokens": 4096,
157 "can_reason": false,
158 "supports_attachments": false
159 },
160 {
161 "id": "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8",
162 "name": "Llama 4 Maverick 17B 128E Instruct FP8",
163 "cost_per_1m_in": 0.22,
164 "cost_per_1m_out": 0.88,
165 "cost_per_1m_in_cached": 0.14,
166 "cost_per_1m_out_cached": 0.55,
167 "context_window": 536576,
168 "default_max_tokens": 4096,
169 "can_reason": false,
170 "supports_attachments": true
171 },
172 {
173 "id": "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct",
174 "name": "Llama 4 Scout 17B 16E Instruct",
175 "cost_per_1m_in": 0.15,
176 "cost_per_1m_out": 0.6,
177 "cost_per_1m_in_cached": 0.14,
178 "cost_per_1m_out_cached": 0.55,
179 "context_window": 328000,
180 "default_max_tokens": 8192,
181 "can_reason": false,
182 "supports_attachments": true
183 },
184 {
185 "id": "hf:MiniMaxAI/MiniMax-M2",
186 "name": "MiniMax M2",
187 "cost_per_1m_in": 0.55,
188 "cost_per_1m_out": 2.19,
189 "cost_per_1m_in_cached": 0.27,
190 "cost_per_1m_out_cached": 0.55,
191 "context_window": 196608,
192 "default_max_tokens": 65536,
193 "can_reason": true,
194 "reasoning_levels": [
195 "low",
196 "medium",
197 "high"
198 ],
199 "default_reasoning_effort": "medium",
200 "supports_attachments": false
201 },
202 {
203 "id": "hf:moonshotai/Kimi-K2-Instruct",
204 "name": "Kimi K2 Instruct",
205 "cost_per_1m_in": 0.6,
206 "cost_per_1m_out": 2.5,
207 "cost_per_1m_in_cached": 0.27,
208 "cost_per_1m_out_cached": 0.55,
209 "context_window": 128000,
210 "default_max_tokens": 131072,
211 "can_reason": false,
212 "supports_attachments": false
213 },
214 {
215 "id": "hf:moonshotai/Kimi-K2-Thinking",
216 "name": "Kimi K2 Thinking",
217 "cost_per_1m_in": 0.55,
218 "cost_per_1m_out": 2.19,
219 "cost_per_1m_in_cached": 0.55,
220 "cost_per_1m_out_cached": 2.19,
221 "context_window": 196608,
222 "default_max_tokens": 65536,
223 "can_reason": true,
224 "reasoning_levels": [
225 "low",
226 "medium",
227 "high"
228 ],
229 "default_reasoning_effort": "medium",
230 "supports_attachments": false
231 },
232 {
233 "id": "hf:moonshotai/Kimi-K2-Instruct-0905",
234 "name": "Kimi K2 Instruct 0905",
235 "cost_per_1m_in": 1.2,
236 "cost_per_1m_out": 1.2,
237 "cost_per_1m_in_cached": 0.55,
238 "cost_per_1m_out_cached": 1.1,
239 "context_window": 262144,
240 "default_max_tokens": 262144,
241 "can_reason": false,
242 "supports_attachments": false
243 },
244 {
245 "id": "hf:openai/gpt-oss-120b",
246 "name": "GPT-OSS 120B",
247 "cost_per_1m_in": 0.1,
248 "cost_per_1m_out": 0.1,
249 "cost_per_1m_in_cached": 0.55,
250 "cost_per_1m_out_cached": 1.1,
251 "context_window": 128000,
252 "default_max_tokens": 65536,
253 "can_reason": true,
254 "reasoning_levels": [
255 "low",
256 "medium",
257 "high"
258 ],
259 "default_reasoning_effort": "medium",
260 "supports_attachments": false
261 },
262 {
263 "id": "hf:Qwen/Qwen2.5-Coder-32B-Instruct",
264 "name": "Qwen2.5 Coder 32B Instruct",
265 "cost_per_1m_in": 0.14,
266 "cost_per_1m_out": 0.55,
267 "cost_per_1m_in_cached": 0.14,
268 "cost_per_1m_out_cached": 0.55,
269 "context_window": 32768,
270 "default_max_tokens": 32768,
271 "can_reason": true,
272 "reasoning_levels": [
273 "low",
274 "medium",
275 "high"
276 ],
277 "default_reasoning_effort": "medium",
278 "supports_attachments": false
279 },
280 {
281 "id": "hf:Qwen/Qwen3-235B-A22B-Instruct-2507",
282 "name": "Qwen3 235B A22B Instruct 2507",
283 "cost_per_1m_in": 0.22,
284 "cost_per_1m_out": 0.88,
285 "cost_per_1m_in_cached": 0.55,
286 "cost_per_1m_out_cached": 1.1,
287 "context_window": 262144,
288 "default_max_tokens": 6912,
289 "can_reason": true,
290 "reasoning_levels": [
291 "low",
292 "medium",
293 "high"
294 ],
295 "default_reasoning_effort": "medium",
296 "supports_attachments": false
297 },
298 {
299 "id": "hf:Qwen/Qwen3-235B-A22B-Thinking-2507",
300 "name": "Qwen3 235B A22B Thinking 2507",
301 "cost_per_1m_in": 0.65,
302 "cost_per_1m_out": 3.0,
303 "cost_per_1m_in_cached": 0.55,
304 "cost_per_1m_out_cached": 1.1,
305 "context_window": 256000,
306 "default_max_tokens": 81920,
307 "can_reason": true,
308 "reasoning_levels": [
309 "low",
310 "medium",
311 "high"
312 ],
313 "default_reasoning_effort": "medium",
314 "supports_attachments": false
315 },
316 {
317 "id": "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct",
318 "name": "Qwen3 Coder 480B A35B Instruct",
319 "cost_per_1m_in": 0.45,
320 "cost_per_1m_out": 1.8,
321 "cost_per_1m_in_cached": 0.82,
322 "cost_per_1m_out_cached": 1.65,
323 "context_window": 256000,
324 "default_max_tokens": 262144,
325 "can_reason": true,
326 "reasoning_levels": [
327 "low",
328 "medium",
329 "high"
330 ],
331 "default_reasoning_effort": "medium",
332 "supports_attachments": false
333 },
334 {
335 "id": "hf:zai-org/GLM-4.5",
336 "name": "GLM-4.5",
337 "cost_per_1m_in": 0.55,
338 "cost_per_1m_out": 2.19,
339 "cost_per_1m_in_cached": 0.14,
340 "cost_per_1m_out_cached": 0.55,
341 "context_window": 128000,
342 "default_max_tokens": 98304,
343 "can_reason": true,
344 "reasoning_levels": [
345 "low",
346 "medium",
347 "high"
348 ],
349 "default_reasoning_effort": "medium",
350 "supports_attachments": false
351 },
352 {
353 "id": "hf:zai-org/GLM-4.6",
354 "name": "GLM-4.6",
355 "cost_per_1m_in": 0.55,
356 "cost_per_1m_out": 0.55,
357 "cost_per_1m_in_cached": 0.27,
358 "cost_per_1m_out_cached": 0.55,
359 "context_window": 202752,
360 "default_max_tokens": 65536,
361 "can_reason": true,
362 "reasoning_levels": [
363 "low",
364 "medium",
365 "high"
366 ],
367 "default_reasoning_effort": "medium",
368 "supports_attachments": false
369 }
370 ]
371}