synthetic.json

  1{
  2  "name": "Synthetic",
  3  "id": "synthetic",
  4  "type": "openai-compat",
  5  "api_key": "$SYNTHETIC_API_KEY",
  6  "api_endpoint": "https://api.synthetic.new/openai/v1",
  7  "default_large_model_id": "hf:zai-org/GLM-4.6",
  8  "default_small_model_id": "hf:openai/gpt-oss-120b",
  9  "models": [
 10    {
 11      "id": "hf:deepseek-ai/DeepSeek-R1",
 12      "name": "DeepSeek R1",
 13      "cost_per_1m_in": 0.55,
 14      "cost_per_1m_out": 2.19,
 15      "cost_per_1m_in_cached": 0.07,
 16      "cost_per_1m_out_cached": 0.14,
 17      "context_window": 128000,
 18      "default_max_tokens": 65536,
 19      "can_reason": true,
 20      "reasoning_levels": [
 21        "low",
 22        "medium",
 23        "high"
 24      ],
 25      "default_reasoning_effort": "medium",
 26      "supports_attachments": false
 27    },
 28    {
 29      "id": "hf:deepseek-ai/DeepSeek-R1-0528",
 30      "name": "DeepSeek R1 0528",
 31      "cost_per_1m_in": 3.0,
 32      "cost_per_1m_out": 8.0,
 33      "cost_per_1m_in_cached": 0.07,
 34      "cost_per_1m_out_cached": 0.14,
 35      "context_window": 128000,
 36      "default_max_tokens": 65536,
 37      "can_reason": true,
 38      "reasoning_levels": [
 39        "low",
 40        "medium",
 41        "high"
 42      ],
 43      "default_reasoning_effort": "medium",
 44      "supports_attachments": false
 45    },
 46    {
 47      "id": "hf:deepseek-ai/DeepSeek-V3",
 48      "name": "DeepSeek V3",
 49      "cost_per_1m_in": 1.25,
 50      "cost_per_1m_out": 1.25,
 51      "cost_per_1m_in_cached": 0.07,
 52      "cost_per_1m_out_cached": 0.14,
 53      "context_window": 128000,
 54      "default_max_tokens": 8192,
 55      "can_reason": false,
 56      "supports_attachments": false
 57    },
 58    {
 59      "id": "hf:deepseek-ai/DeepSeek-V3-0324",
 60      "name": "DeepSeek V3 0324",
 61      "cost_per_1m_in": 1.2,
 62      "cost_per_1m_out": 1.2,
 63      "cost_per_1m_in_cached": 0.07,
 64      "cost_per_1m_out_cached": 0.14,
 65      "context_window": 128000,
 66      "default_max_tokens": 8192,
 67      "can_reason": false,
 68      "supports_attachments": false
 69    },
 70    {
 71      "id": "hf:deepseek-ai/DeepSeek-V3.1",
 72      "name": "DeepSeek V3.1",
 73      "cost_per_1m_in": 0.56,
 74      "cost_per_1m_out": 1.68,
 75      "cost_per_1m_in_cached": 0.07,
 76      "cost_per_1m_out_cached": 0.14,
 77      "context_window": 131072,
 78      "default_max_tokens": 8192,
 79      "can_reason": true,
 80      "reasoning_levels": [
 81        "low",
 82        "medium",
 83        "high"
 84      ],
 85      "default_reasoning_effort": "medium",
 86      "supports_attachments": false
 87    },
 88    {
 89      "id": "hf:deepseek-ai/DeepSeek-V3.1-Terminus",
 90      "name": "DeepSeek V3.1 Terminus",
 91      "cost_per_1m_in": 1.2,
 92      "cost_per_1m_out": 1.2,
 93      "cost_per_1m_in_cached": 0.07,
 94      "cost_per_1m_out_cached": 0.14,
 95      "context_window": 128000,
 96      "default_max_tokens": 8192,
 97      "can_reason": true,
 98      "reasoning_levels": [
 99        "low",
100        "medium",
101        "high"
102      ],
103      "default_reasoning_effort": "medium",
104      "supports_attachments": false
105    },
106    {
107      "id": "hf:meta-llama/Llama-3.1-405B-Instruct",
108      "name": "Llama 3.1 405B Instruct",
109      "cost_per_1m_in": 3.0,
110      "cost_per_1m_out": 3.0,
111      "cost_per_1m_in_cached": 0.27,
112      "cost_per_1m_out_cached": 0.55,
113      "context_window": 131072,
114      "default_max_tokens": 4096,
115      "can_reason": true,
116      "reasoning_levels": [
117        "low",
118        "medium",
119        "high"
120      ],
121      "default_reasoning_effort": "medium",
122      "supports_attachments": false
123    },
124    {
125      "id": "hf:meta-llama/Llama-3.1-70B-Instruct",
126      "name": "Llama 3.1 70B Instruct",
127      "cost_per_1m_in": 0.9,
128      "cost_per_1m_out": 0.9,
129      "cost_per_1m_in_cached": 0.59,
130      "cost_per_1m_out_cached": 1.1,
131      "context_window": 131072,
132      "default_max_tokens": 4096,
133      "can_reason": false,
134      "supports_attachments": false
135    },
136    {
137      "id": "hf:meta-llama/Llama-3.1-8B-Instruct",
138      "name": "Llama 3.1 8B Instruct",
139      "cost_per_1m_in": 0.2,
140      "cost_per_1m_out": 0.2,
141      "cost_per_1m_in_cached": 0.07,
142      "cost_per_1m_out_cached": 0.2,
143      "context_window": 128000,
144      "default_max_tokens": 4096,
145      "can_reason": false,
146      "supports_attachments": false
147    },
148    {
149      "id": "hf:meta-llama/Llama-3.3-70B-Instruct",
150      "name": "Llama 3.3 70B Instruct",
151      "cost_per_1m_in": 0.9,
152      "cost_per_1m_out": 0.9,
153      "cost_per_1m_in_cached": 0.59,
154      "cost_per_1m_out_cached": 1.1,
155      "context_window": 128000,
156      "default_max_tokens": 4096,
157      "can_reason": false,
158      "supports_attachments": false
159    },
160    {
161      "id": "hf:meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8",
162      "name": "Llama 4 Maverick 17B 128E Instruct FP8",
163      "cost_per_1m_in": 0.22,
164      "cost_per_1m_out": 0.88,
165      "cost_per_1m_in_cached": 0.14,
166      "cost_per_1m_out_cached": 0.55,
167      "context_window": 536576,
168      "default_max_tokens": 4096,
169      "can_reason": false,
170      "supports_attachments": true
171    },
172    {
173      "id": "hf:meta-llama/Llama-4-Scout-17B-16E-Instruct",
174      "name": "Llama 4 Scout 17B 16E Instruct",
175      "cost_per_1m_in": 0.15,
176      "cost_per_1m_out": 0.6,
177      "cost_per_1m_in_cached": 0.14,
178      "cost_per_1m_out_cached": 0.55,
179      "context_window": 328000,
180      "default_max_tokens": 8192,
181      "can_reason": false,
182      "supports_attachments": true
183    },
184    {
185      "id": "hf:MiniMaxAI/MiniMax-M2",
186      "name": "MiniMax M2",
187      "cost_per_1m_in": 0.55,
188      "cost_per_1m_out": 2.19,
189      "cost_per_1m_in_cached": 0.27,
190      "cost_per_1m_out_cached": 0.55,
191      "context_window": 196608,
192      "default_max_tokens": 65536,
193      "can_reason": true,
194      "reasoning_levels": [
195        "low",
196        "medium",
197        "high"
198      ],
199      "default_reasoning_effort": "medium",
200      "supports_attachments": false
201    },
202    {
203      "id": "hf:moonshotai/Kimi-K2-Instruct",
204      "name": "Kimi K2 Instruct",
205      "cost_per_1m_in": 0.6,
206      "cost_per_1m_out": 2.5,
207      "cost_per_1m_in_cached": 0.27,
208      "cost_per_1m_out_cached": 0.55,
209      "context_window": 128000,
210      "default_max_tokens": 131072,
211      "can_reason": false,
212      "supports_attachments": false
213    },
214    {
215      "id": "hf:moonshotai/Kimi-K2-Thinking",
216      "name": "Kimi K2 Thinking",
217      "cost_per_1m_in": 0.55,
218      "cost_per_1m_out": 2.19,
219      "cost_per_1m_in_cached": 0.55,
220      "cost_per_1m_out_cached": 2.19,
221      "context_window": 196608,
222      "default_max_tokens": 65536,
223      "can_reason": true,
224      "reasoning_levels": [
225        "low",
226        "medium",
227        "high"
228      ],
229      "default_reasoning_effort": "medium",
230      "supports_attachments": false
231    },
232    {
233      "id": "hf:moonshotai/Kimi-K2-Instruct-0905",
234      "name": "Kimi K2 Instruct 0905",
235      "cost_per_1m_in": 1.2,
236      "cost_per_1m_out": 1.2,
237      "cost_per_1m_in_cached": 0.55,
238      "cost_per_1m_out_cached": 1.1,
239      "context_window": 262144,
240      "default_max_tokens": 262144,
241      "can_reason": false,
242      "supports_attachments": false
243    },
244    {
245      "id": "hf:openai/gpt-oss-120b",
246      "name": "GPT-OSS 120B",
247      "cost_per_1m_in": 0.1,
248      "cost_per_1m_out": 0.1,
249      "cost_per_1m_in_cached": 0.55,
250      "cost_per_1m_out_cached": 1.1,
251      "context_window": 128000,
252      "default_max_tokens": 65536,
253      "can_reason": true,
254      "reasoning_levels": [
255        "low",
256        "medium",
257        "high"
258      ],
259      "default_reasoning_effort": "medium",
260      "supports_attachments": false
261    },
262    {
263      "id": "hf:Qwen/Qwen2.5-Coder-32B-Instruct",
264      "name": "Qwen2.5 Coder 32B Instruct",
265      "cost_per_1m_in": 0.14,
266      "cost_per_1m_out": 0.55,
267      "cost_per_1m_in_cached": 0.14,
268      "cost_per_1m_out_cached": 0.55,
269      "context_window": 32768,
270      "default_max_tokens": 32768,
271      "can_reason": true,
272      "reasoning_levels": [
273        "low",
274        "medium",
275        "high"
276      ],
277      "default_reasoning_effort": "medium",
278      "supports_attachments": false
279    },
280    {
281      "id": "hf:Qwen/Qwen3-235B-A22B-Instruct-2507",
282      "name": "Qwen3 235B A22B Instruct 2507",
283      "cost_per_1m_in": 0.22,
284      "cost_per_1m_out": 0.88,
285      "cost_per_1m_in_cached": 0.55,
286      "cost_per_1m_out_cached": 1.1,
287      "context_window": 262144,
288      "default_max_tokens": 6912,
289      "can_reason": true,
290      "reasoning_levels": [
291        "low",
292        "medium",
293        "high"
294      ],
295      "default_reasoning_effort": "medium",
296      "supports_attachments": false
297    },
298    {
299      "id": "hf:Qwen/Qwen3-235B-A22B-Thinking-2507",
300      "name": "Qwen3 235B A22B Thinking 2507",
301      "cost_per_1m_in": 0.65,
302      "cost_per_1m_out": 3.0,
303      "cost_per_1m_in_cached": 0.55,
304      "cost_per_1m_out_cached": 1.1,
305      "context_window": 256000,
306      "default_max_tokens": 81920,
307      "can_reason": true,
308      "reasoning_levels": [
309        "low",
310        "medium",
311        "high"
312      ],
313      "default_reasoning_effort": "medium",
314      "supports_attachments": false
315    },
316    {
317      "id": "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct",
318      "name": "Qwen3 Coder 480B A35B Instruct",
319      "cost_per_1m_in": 0.45,
320      "cost_per_1m_out": 1.8,
321      "cost_per_1m_in_cached": 0.82,
322      "cost_per_1m_out_cached": 1.65,
323      "context_window": 256000,
324      "default_max_tokens": 262144,
325      "can_reason": true,
326      "reasoning_levels": [
327        "low",
328        "medium",
329        "high"
330      ],
331      "default_reasoning_effort": "medium",
332      "supports_attachments": false
333    },
334    {
335      "id": "hf:zai-org/GLM-4.5",
336      "name": "GLM-4.5",
337      "cost_per_1m_in": 0.55,
338      "cost_per_1m_out": 2.19,
339      "cost_per_1m_in_cached": 0.14,
340      "cost_per_1m_out_cached": 0.55,
341      "context_window": 128000,
342      "default_max_tokens": 98304,
343      "can_reason": true,
344      "reasoning_levels": [
345        "low",
346        "medium",
347        "high"
348      ],
349      "default_reasoning_effort": "medium",
350      "supports_attachments": false
351    },
352    {
353      "id": "hf:zai-org/GLM-4.6",
354      "name": "GLM-4.6",
355      "cost_per_1m_in": 0.55,
356      "cost_per_1m_out": 0.55,
357      "cost_per_1m_in_cached": 0.27,
358      "cost_per_1m_out_cached": 0.55,
359      "context_window": 202752,
360      "default_max_tokens": 65536,
361      "can_reason": true,
362      "reasoning_levels": [
363        "low",
364        "medium",
365        "high"
366      ],
367      "default_reasoning_effort": "medium",
368      "supports_attachments": false
369   }
370  ]
371}