1{
2 "name": "Venice AI",
3 "id": "venice",
4 "api_key": "$VENICE_API_KEY",
5 "api_endpoint": "https://api.venice.ai/api/v1",
6 "type": "openai-compat",
7 "default_large_model_id": "claude-opus-4-6",
8 "default_small_model_id": "minimax-m25",
9 "models": [
10 {
11 "id": "claude-opus-4-5",
12 "name": "Claude Opus 4.5",
13 "cost_per_1m_in": 6,
14 "cost_per_1m_out": 30,
15 "cost_per_1m_in_cached": 0,
16 "cost_per_1m_out_cached": 0,
17 "context_window": 198000,
18 "default_max_tokens": 32768,
19 "can_reason": true,
20 "reasoning_levels": [
21 "low",
22 "medium",
23 "high"
24 ],
25 "default_reasoning_effort": "medium",
26 "supports_attachments": true,
27 "options": {}
28 },
29 {
30 "id": "claude-opus-4-6",
31 "name": "Claude Opus 4.6",
32 "cost_per_1m_in": 6,
33 "cost_per_1m_out": 30,
34 "cost_per_1m_in_cached": 0,
35 "cost_per_1m_out_cached": 0,
36 "context_window": 1000000,
37 "default_max_tokens": 32768,
38 "can_reason": true,
39 "reasoning_levels": [
40 "low",
41 "medium",
42 "high"
43 ],
44 "default_reasoning_effort": "medium",
45 "supports_attachments": true,
46 "options": {}
47 },
48 {
49 "id": "claude-sonnet-4-5",
50 "name": "Claude Sonnet 4.5",
51 "cost_per_1m_in": 3.75,
52 "cost_per_1m_out": 18.75,
53 "cost_per_1m_in_cached": 0,
54 "cost_per_1m_out_cached": 0,
55 "context_window": 198000,
56 "default_max_tokens": 32768,
57 "can_reason": true,
58 "reasoning_levels": [
59 "low",
60 "medium",
61 "high"
62 ],
63 "default_reasoning_effort": "medium",
64 "supports_attachments": true,
65 "options": {}
66 },
67 {
68 "id": "claude-sonnet-4-6",
69 "name": "Claude Sonnet 4.6",
70 "cost_per_1m_in": 3.6,
71 "cost_per_1m_out": 18,
72 "cost_per_1m_in_cached": 0,
73 "cost_per_1m_out_cached": 0,
74 "context_window": 1000000,
75 "default_max_tokens": 32768,
76 "can_reason": true,
77 "reasoning_levels": [
78 "low",
79 "medium",
80 "high"
81 ],
82 "default_reasoning_effort": "medium",
83 "supports_attachments": true,
84 "options": {}
85 },
86 {
87 "id": "deepseek-v3.2",
88 "name": "DeepSeek V3.2",
89 "cost_per_1m_in": 0.33,
90 "cost_per_1m_out": 0.48,
91 "cost_per_1m_in_cached": 0,
92 "cost_per_1m_out_cached": 0,
93 "context_window": 160000,
94 "default_max_tokens": 32768,
95 "can_reason": true,
96 "reasoning_levels": [
97 "low",
98 "medium",
99 "high"
100 ],
101 "default_reasoning_effort": "medium",
102 "supports_attachments": false,
103 "options": {}
104 },
105 {
106 "id": "zai-org-glm-4.6",
107 "name": "GLM 4.6",
108 "cost_per_1m_in": 0.85,
109 "cost_per_1m_out": 2.75,
110 "cost_per_1m_in_cached": 0,
111 "cost_per_1m_out_cached": 0,
112 "context_window": 198000,
113 "default_max_tokens": 32768,
114 "can_reason": false,
115 "supports_attachments": false,
116 "options": {}
117 },
118 {
119 "id": "zai-org-glm-4.7",
120 "name": "GLM 4.7",
121 "cost_per_1m_in": 0.55,
122 "cost_per_1m_out": 2.65,
123 "cost_per_1m_in_cached": 0,
124 "cost_per_1m_out_cached": 0,
125 "context_window": 198000,
126 "default_max_tokens": 32768,
127 "can_reason": true,
128 "reasoning_levels": [
129 "low",
130 "medium",
131 "high"
132 ],
133 "default_reasoning_effort": "medium",
134 "supports_attachments": false,
135 "options": {}
136 },
137 {
138 "id": "zai-org-glm-4.7-flash",
139 "name": "GLM 4.7 Flash",
140 "cost_per_1m_in": 0.125,
141 "cost_per_1m_out": 0.5,
142 "cost_per_1m_in_cached": 0,
143 "cost_per_1m_out_cached": 0,
144 "context_window": 128000,
145 "default_max_tokens": 32000,
146 "can_reason": true,
147 "reasoning_levels": [
148 "low",
149 "medium",
150 "high"
151 ],
152 "default_reasoning_effort": "medium",
153 "supports_attachments": false,
154 "options": {}
155 },
156 {
157 "id": "olafangensan-glm-4.7-flash-heretic",
158 "name": "GLM 4.7 Flash Heretic",
159 "cost_per_1m_in": 0.14,
160 "cost_per_1m_out": 0.8,
161 "cost_per_1m_in_cached": 0,
162 "cost_per_1m_out_cached": 0,
163 "context_window": 200000,
164 "default_max_tokens": 32768,
165 "can_reason": true,
166 "reasoning_levels": [
167 "low",
168 "medium",
169 "high"
170 ],
171 "default_reasoning_effort": "medium",
172 "supports_attachments": false,
173 "options": {}
174 },
175 {
176 "id": "zai-org-glm-5",
177 "name": "GLM 5",
178 "cost_per_1m_in": 1,
179 "cost_per_1m_out": 3.2,
180 "cost_per_1m_in_cached": 0,
181 "cost_per_1m_out_cached": 0,
182 "context_window": 198000,
183 "default_max_tokens": 32768,
184 "can_reason": true,
185 "reasoning_levels": [
186 "low",
187 "medium",
188 "high"
189 ],
190 "default_reasoning_effort": "medium",
191 "supports_attachments": false,
192 "options": {}
193 },
194 {
195 "id": "openai-gpt-4o-2024-11-20",
196 "name": "GPT-4o",
197 "cost_per_1m_in": 3.125,
198 "cost_per_1m_out": 12.5,
199 "cost_per_1m_in_cached": 0,
200 "cost_per_1m_out_cached": 0,
201 "context_window": 128000,
202 "default_max_tokens": 32000,
203 "can_reason": false,
204 "supports_attachments": true,
205 "options": {}
206 },
207 {
208 "id": "openai-gpt-4o-mini-2024-07-18",
209 "name": "GPT-4o Mini",
210 "cost_per_1m_in": 0.1875,
211 "cost_per_1m_out": 0.75,
212 "cost_per_1m_in_cached": 0,
213 "cost_per_1m_out_cached": 0,
214 "context_window": 128000,
215 "default_max_tokens": 32000,
216 "can_reason": false,
217 "supports_attachments": true,
218 "options": {}
219 },
220 {
221 "id": "openai-gpt-52",
222 "name": "GPT-5.2",
223 "cost_per_1m_in": 2.19,
224 "cost_per_1m_out": 17.5,
225 "cost_per_1m_in_cached": 0,
226 "cost_per_1m_out_cached": 0,
227 "context_window": 256000,
228 "default_max_tokens": 32768,
229 "can_reason": true,
230 "reasoning_levels": [
231 "low",
232 "medium",
233 "high"
234 ],
235 "default_reasoning_effort": "medium",
236 "supports_attachments": false,
237 "options": {}
238 },
239 {
240 "id": "openai-gpt-52-codex",
241 "name": "GPT-5.2 Codex",
242 "cost_per_1m_in": 2.19,
243 "cost_per_1m_out": 17.5,
244 "cost_per_1m_in_cached": 0,
245 "cost_per_1m_out_cached": 0,
246 "context_window": 256000,
247 "default_max_tokens": 32768,
248 "can_reason": true,
249 "reasoning_levels": [
250 "low",
251 "medium",
252 "high"
253 ],
254 "default_reasoning_effort": "medium",
255 "supports_attachments": true,
256 "options": {}
257 },
258 {
259 "id": "openai-gpt-53-codex",
260 "name": "GPT-5.3 Codex",
261 "cost_per_1m_in": 2.19,
262 "cost_per_1m_out": 17.5,
263 "cost_per_1m_in_cached": 0,
264 "cost_per_1m_out_cached": 0,
265 "context_window": 400000,
266 "default_max_tokens": 32768,
267 "can_reason": true,
268 "reasoning_levels": [
269 "low",
270 "medium",
271 "high"
272 ],
273 "default_reasoning_effort": "medium",
274 "supports_attachments": true,
275 "options": {}
276 },
277 {
278 "id": "openai-gpt-54",
279 "name": "GPT-5.4",
280 "cost_per_1m_in": 3.13,
281 "cost_per_1m_out": 18.8,
282 "cost_per_1m_in_cached": 0,
283 "cost_per_1m_out_cached": 0,
284 "context_window": 1000000,
285 "default_max_tokens": 32768,
286 "can_reason": true,
287 "reasoning_levels": [
288 "low",
289 "medium",
290 "high"
291 ],
292 "default_reasoning_effort": "medium",
293 "supports_attachments": true,
294 "options": {}
295 },
296 {
297 "id": "openai-gpt-54-pro",
298 "name": "GPT-5.4 Pro",
299 "cost_per_1m_in": 37.5,
300 "cost_per_1m_out": 225,
301 "cost_per_1m_in_cached": 0,
302 "cost_per_1m_out_cached": 0,
303 "context_window": 1000000,
304 "default_max_tokens": 32768,
305 "can_reason": true,
306 "reasoning_levels": [
307 "low",
308 "medium",
309 "high"
310 ],
311 "default_reasoning_effort": "medium",
312 "supports_attachments": true,
313 "options": {}
314 },
315 {
316 "id": "gemini-3-flash-preview",
317 "name": "Gemini 3 Flash Preview",
318 "cost_per_1m_in": 0.7,
319 "cost_per_1m_out": 3.75,
320 "cost_per_1m_in_cached": 0,
321 "cost_per_1m_out_cached": 0,
322 "context_window": 256000,
323 "default_max_tokens": 32768,
324 "can_reason": true,
325 "reasoning_levels": [
326 "low",
327 "medium",
328 "high"
329 ],
330 "default_reasoning_effort": "medium",
331 "supports_attachments": true,
332 "options": {}
333 },
334 {
335 "id": "gemini-3-pro-preview",
336 "name": "Gemini 3 Pro Preview",
337 "cost_per_1m_in": 2.5,
338 "cost_per_1m_out": 15,
339 "cost_per_1m_in_cached": 0,
340 "cost_per_1m_out_cached": 0,
341 "context_window": 198000,
342 "default_max_tokens": 32768,
343 "can_reason": true,
344 "reasoning_levels": [
345 "low",
346 "medium",
347 "high"
348 ],
349 "default_reasoning_effort": "medium",
350 "supports_attachments": true,
351 "options": {}
352 },
353 {
354 "id": "gemini-3-1-pro-preview",
355 "name": "Gemini 3.1 Pro Preview",
356 "cost_per_1m_in": 2.5,
357 "cost_per_1m_out": 15,
358 "cost_per_1m_in_cached": 0,
359 "cost_per_1m_out_cached": 0,
360 "context_window": 1000000,
361 "default_max_tokens": 32768,
362 "can_reason": true,
363 "reasoning_levels": [
364 "low",
365 "medium",
366 "high"
367 ],
368 "default_reasoning_effort": "medium",
369 "supports_attachments": true,
370 "options": {}
371 },
372 {
373 "id": "google-gemma-3-27b-it",
374 "name": "Google Gemma 3 27B Instruct",
375 "cost_per_1m_in": 0.12,
376 "cost_per_1m_out": 0.2,
377 "cost_per_1m_in_cached": 0,
378 "cost_per_1m_out_cached": 0,
379 "context_window": 198000,
380 "default_max_tokens": 32768,
381 "can_reason": false,
382 "supports_attachments": true,
383 "options": {}
384 },
385 {
386 "id": "grok-41-fast",
387 "name": "Grok 4.1 Fast",
388 "cost_per_1m_in": 0.25,
389 "cost_per_1m_out": 0.625,
390 "cost_per_1m_in_cached": 0,
391 "cost_per_1m_out_cached": 0,
392 "context_window": 1000000,
393 "default_max_tokens": 32768,
394 "can_reason": true,
395 "reasoning_levels": [
396 "low",
397 "medium",
398 "high"
399 ],
400 "default_reasoning_effort": "medium",
401 "supports_attachments": true,
402 "options": {}
403 },
404 {
405 "id": "grok-4-20-beta",
406 "name": "Grok 4.20 Beta",
407 "cost_per_1m_in": 2.5,
408 "cost_per_1m_out": 7.5,
409 "cost_per_1m_in_cached": 0,
410 "cost_per_1m_out_cached": 0,
411 "context_window": 2000000,
412 "default_max_tokens": 32768,
413 "can_reason": true,
414 "reasoning_levels": [
415 "low",
416 "medium",
417 "high"
418 ],
419 "default_reasoning_effort": "medium",
420 "supports_attachments": true,
421 "options": {}
422 },
423 {
424 "id": "grok-code-fast-1",
425 "name": "Grok Code Fast 1",
426 "cost_per_1m_in": 0.25,
427 "cost_per_1m_out": 1.87,
428 "cost_per_1m_in_cached": 0,
429 "cost_per_1m_out_cached": 0,
430 "context_window": 256000,
431 "default_max_tokens": 32768,
432 "can_reason": true,
433 "reasoning_levels": [
434 "low",
435 "medium",
436 "high"
437 ],
438 "default_reasoning_effort": "medium",
439 "supports_attachments": false,
440 "options": {}
441 },
442 {
443 "id": "kimi-k2-thinking",
444 "name": "Kimi K2 Thinking",
445 "cost_per_1m_in": 0.75,
446 "cost_per_1m_out": 3.2,
447 "cost_per_1m_in_cached": 0,
448 "cost_per_1m_out_cached": 0,
449 "context_window": 256000,
450 "default_max_tokens": 32768,
451 "can_reason": true,
452 "reasoning_levels": [
453 "low",
454 "medium",
455 "high"
456 ],
457 "default_reasoning_effort": "medium",
458 "supports_attachments": false,
459 "options": {}
460 },
461 {
462 "id": "kimi-k2-5",
463 "name": "Kimi K2.5",
464 "cost_per_1m_in": 0.56,
465 "cost_per_1m_out": 3.5,
466 "cost_per_1m_in_cached": 0,
467 "cost_per_1m_out_cached": 0,
468 "context_window": 256000,
469 "default_max_tokens": 32768,
470 "can_reason": true,
471 "reasoning_levels": [
472 "low",
473 "medium",
474 "high"
475 ],
476 "default_reasoning_effort": "medium",
477 "supports_attachments": true,
478 "options": {}
479 },
480 {
481 "id": "llama-3.2-3b",
482 "name": "Llama 3.2 3B",
483 "cost_per_1m_in": 0.15,
484 "cost_per_1m_out": 0.6,
485 "cost_per_1m_in_cached": 0,
486 "cost_per_1m_out_cached": 0,
487 "context_window": 128000,
488 "default_max_tokens": 32000,
489 "can_reason": false,
490 "supports_attachments": false,
491 "options": {}
492 },
493 {
494 "id": "llama-3.3-70b",
495 "name": "Llama 3.3 70B",
496 "cost_per_1m_in": 0.7,
497 "cost_per_1m_out": 2.8,
498 "cost_per_1m_in_cached": 0,
499 "cost_per_1m_out_cached": 0,
500 "context_window": 128000,
501 "default_max_tokens": 32000,
502 "can_reason": false,
503 "supports_attachments": false,
504 "options": {}
505 },
506 {
507 "id": "minimax-m21",
508 "name": "MiniMax M2.1",
509 "cost_per_1m_in": 0.35,
510 "cost_per_1m_out": 1.5,
511 "cost_per_1m_in_cached": 0,
512 "cost_per_1m_out_cached": 0,
513 "context_window": 198000,
514 "default_max_tokens": 32768,
515 "can_reason": true,
516 "reasoning_levels": [
517 "low",
518 "medium",
519 "high"
520 ],
521 "default_reasoning_effort": "medium",
522 "supports_attachments": false,
523 "options": {}
524 },
525 {
526 "id": "minimax-m25",
527 "name": "MiniMax M2.5",
528 "cost_per_1m_in": 0.34,
529 "cost_per_1m_out": 1.19,
530 "cost_per_1m_in_cached": 0,
531 "cost_per_1m_out_cached": 0,
532 "context_window": 198000,
533 "default_max_tokens": 32768,
534 "can_reason": true,
535 "reasoning_levels": [
536 "low",
537 "medium",
538 "high"
539 ],
540 "default_reasoning_effort": "medium",
541 "supports_attachments": false,
542 "options": {}
543 },
544 {
545 "id": "minimax-m27",
546 "name": "MiniMax M2.7",
547 "cost_per_1m_in": 0.375,
548 "cost_per_1m_out": 1.5,
549 "cost_per_1m_in_cached": 0,
550 "cost_per_1m_out_cached": 0,
551 "context_window": 198000,
552 "default_max_tokens": 32768,
553 "can_reason": true,
554 "reasoning_levels": [
555 "low",
556 "medium",
557 "high"
558 ],
559 "default_reasoning_effort": "medium",
560 "supports_attachments": false,
561 "options": {}
562 },
563 {
564 "id": "mistral-small-3-2-24b-instruct",
565 "name": "Mistral Small 3.2 24B Instruct",
566 "cost_per_1m_in": 0.09375,
567 "cost_per_1m_out": 0.25,
568 "cost_per_1m_in_cached": 0,
569 "cost_per_1m_out_cached": 0,
570 "context_window": 256000,
571 "default_max_tokens": 32768,
572 "can_reason": false,
573 "supports_attachments": false,
574 "options": {}
575 },
576 {
577 "id": "nvidia-nemotron-3-nano-30b-a3b",
578 "name": "NVIDIA Nemotron 3 Nano 30B",
579 "cost_per_1m_in": 0.075,
580 "cost_per_1m_out": 0.3,
581 "cost_per_1m_in_cached": 0,
582 "cost_per_1m_out_cached": 0,
583 "context_window": 128000,
584 "default_max_tokens": 32000,
585 "can_reason": false,
586 "supports_attachments": false,
587 "options": {}
588 },
589 {
590 "id": "openai-gpt-oss-120b",
591 "name": "OpenAI GPT OSS 120B",
592 "cost_per_1m_in": 0.07,
593 "cost_per_1m_out": 0.3,
594 "cost_per_1m_in_cached": 0,
595 "cost_per_1m_out_cached": 0,
596 "context_window": 128000,
597 "default_max_tokens": 32000,
598 "can_reason": false,
599 "supports_attachments": false,
600 "options": {}
601 },
602 {
603 "id": "qwen3-235b-a22b-instruct-2507",
604 "name": "Qwen 3 235B A22B Instruct 2507",
605 "cost_per_1m_in": 0.15,
606 "cost_per_1m_out": 0.75,
607 "cost_per_1m_in_cached": 0,
608 "cost_per_1m_out_cached": 0,
609 "context_window": 128000,
610 "default_max_tokens": 32000,
611 "can_reason": false,
612 "supports_attachments": false,
613 "options": {}
614 },
615 {
616 "id": "qwen3-235b-a22b-thinking-2507",
617 "name": "Qwen 3 235B A22B Thinking 2507",
618 "cost_per_1m_in": 0.45,
619 "cost_per_1m_out": 3.5,
620 "cost_per_1m_in_cached": 0,
621 "cost_per_1m_out_cached": 0,
622 "context_window": 128000,
623 "default_max_tokens": 32000,
624 "can_reason": true,
625 "reasoning_levels": [
626 "low",
627 "medium",
628 "high"
629 ],
630 "default_reasoning_effort": "medium",
631 "supports_attachments": false,
632 "options": {}
633 },
634 {
635 "id": "qwen3-coder-480b-a35b-instruct-turbo",
636 "name": "Qwen 3 Coder 480B Turbo",
637 "cost_per_1m_in": 0.35,
638 "cost_per_1m_out": 1.5,
639 "cost_per_1m_in_cached": 0,
640 "cost_per_1m_out_cached": 0,
641 "context_window": 256000,
642 "default_max_tokens": 32768,
643 "can_reason": false,
644 "supports_attachments": false,
645 "options": {}
646 },
647 {
648 "id": "qwen3-coder-480b-a35b-instruct",
649 "name": "Qwen 3 Coder 480b",
650 "cost_per_1m_in": 0.75,
651 "cost_per_1m_out": 3,
652 "cost_per_1m_in_cached": 0,
653 "cost_per_1m_out_cached": 0,
654 "context_window": 256000,
655 "default_max_tokens": 32768,
656 "can_reason": false,
657 "supports_attachments": false,
658 "options": {}
659 },
660 {
661 "id": "qwen3-next-80b",
662 "name": "Qwen 3 Next 80b",
663 "cost_per_1m_in": 0.35,
664 "cost_per_1m_out": 1.9,
665 "cost_per_1m_in_cached": 0,
666 "cost_per_1m_out_cached": 0,
667 "context_window": 256000,
668 "default_max_tokens": 32768,
669 "can_reason": false,
670 "supports_attachments": false,
671 "options": {}
672 },
673 {
674 "id": "qwen3-5-35b-a3b",
675 "name": "Qwen 3.5 35B A3B",
676 "cost_per_1m_in": 0.3125,
677 "cost_per_1m_out": 1.25,
678 "cost_per_1m_in_cached": 0,
679 "cost_per_1m_out_cached": 0,
680 "context_window": 256000,
681 "default_max_tokens": 32768,
682 "can_reason": true,
683 "reasoning_levels": [
684 "low",
685 "medium",
686 "high"
687 ],
688 "default_reasoning_effort": "medium",
689 "supports_attachments": true,
690 "options": {
691 "temperature": 1,
692 "top_p": 0.95
693 }
694 },
695 {
696 "id": "qwen3-5-9b",
697 "name": "Qwen 3.5 9B",
698 "cost_per_1m_in": 0.05,
699 "cost_per_1m_out": 0.15,
700 "cost_per_1m_in_cached": 0,
701 "cost_per_1m_out_cached": 0,
702 "context_window": 256000,
703 "default_max_tokens": 32768,
704 "can_reason": true,
705 "reasoning_levels": [
706 "low",
707 "medium",
708 "high"
709 ],
710 "default_reasoning_effort": "medium",
711 "supports_attachments": true,
712 "options": {}
713 },
714 {
715 "id": "qwen3-vl-235b-a22b",
716 "name": "Qwen3 VL 235B",
717 "cost_per_1m_in": 0.25,
718 "cost_per_1m_out": 1.5,
719 "cost_per_1m_in_cached": 0,
720 "cost_per_1m_out_cached": 0,
721 "context_window": 256000,
722 "default_max_tokens": 32768,
723 "can_reason": false,
724 "supports_attachments": true,
725 "options": {}
726 },
727 {
728 "id": "mistral-31-24b",
729 "name": "Venice Medium",
730 "cost_per_1m_in": 0.5,
731 "cost_per_1m_out": 2,
732 "cost_per_1m_in_cached": 0,
733 "cost_per_1m_out_cached": 0,
734 "context_window": 128000,
735 "default_max_tokens": 32000,
736 "can_reason": false,
737 "supports_attachments": true,
738 "options": {}
739 },
740 {
741 "id": "venice-uncensored-role-play",
742 "name": "Venice Role Play Uncensored",
743 "cost_per_1m_in": 0.5,
744 "cost_per_1m_out": 2,
745 "cost_per_1m_in_cached": 0,
746 "cost_per_1m_out_cached": 0,
747 "context_window": 128000,
748 "default_max_tokens": 32000,
749 "can_reason": false,
750 "supports_attachments": true,
751 "options": {}
752 },
753 {
754 "id": "qwen3-4b",
755 "name": "Venice Small",
756 "cost_per_1m_in": 0.05,
757 "cost_per_1m_out": 0.15,
758 "cost_per_1m_in_cached": 0,
759 "cost_per_1m_out_cached": 0,
760 "context_window": 32000,
761 "default_max_tokens": 8000,
762 "can_reason": true,
763 "reasoning_levels": [
764 "low",
765 "medium",
766 "high"
767 ],
768 "default_reasoning_effort": "medium",
769 "supports_attachments": false,
770 "options": {}
771 }
772 ]
773}