venice.json

  1{
  2  "name": "Venice AI",
  3  "id": "venice",
  4  "api_key": "$VENICE_API_KEY",
  5  "api_endpoint": "https://api.venice.ai/api/v1",
  6  "type": "openai-compat",
  7  "default_large_model_id": "claude-opus-4-6",
  8  "default_small_model_id": "minimax-m25",
  9  "models": [
 10    {
 11      "id": "claude-opus-4-5",
 12      "name": "Claude Opus 4.5",
 13      "cost_per_1m_in": 6,
 14      "cost_per_1m_out": 30,
 15      "cost_per_1m_in_cached": 0,
 16      "cost_per_1m_out_cached": 0,
 17      "context_window": 198000,
 18      "default_max_tokens": 32768,
 19      "can_reason": true,
 20      "reasoning_levels": [
 21        "low",
 22        "medium",
 23        "high"
 24      ],
 25      "default_reasoning_effort": "medium",
 26      "supports_attachments": true,
 27      "options": {}
 28    },
 29    {
 30      "id": "claude-opus-4-6",
 31      "name": "Claude Opus 4.6",
 32      "cost_per_1m_in": 6,
 33      "cost_per_1m_out": 30,
 34      "cost_per_1m_in_cached": 0,
 35      "cost_per_1m_out_cached": 0,
 36      "context_window": 1000000,
 37      "default_max_tokens": 32768,
 38      "can_reason": true,
 39      "reasoning_levels": [
 40        "low",
 41        "medium",
 42        "high"
 43      ],
 44      "default_reasoning_effort": "medium",
 45      "supports_attachments": true,
 46      "options": {}
 47    },
 48    {
 49      "id": "claude-sonnet-4-5",
 50      "name": "Claude Sonnet 4.5",
 51      "cost_per_1m_in": 3.75,
 52      "cost_per_1m_out": 18.75,
 53      "cost_per_1m_in_cached": 0,
 54      "cost_per_1m_out_cached": 0,
 55      "context_window": 198000,
 56      "default_max_tokens": 32768,
 57      "can_reason": true,
 58      "reasoning_levels": [
 59        "low",
 60        "medium",
 61        "high"
 62      ],
 63      "default_reasoning_effort": "medium",
 64      "supports_attachments": true,
 65      "options": {}
 66    },
 67    {
 68      "id": "claude-sonnet-4-6",
 69      "name": "Claude Sonnet 4.6",
 70      "cost_per_1m_in": 3.6,
 71      "cost_per_1m_out": 18,
 72      "cost_per_1m_in_cached": 0,
 73      "cost_per_1m_out_cached": 0,
 74      "context_window": 1000000,
 75      "default_max_tokens": 32768,
 76      "can_reason": true,
 77      "reasoning_levels": [
 78        "low",
 79        "medium",
 80        "high"
 81      ],
 82      "default_reasoning_effort": "medium",
 83      "supports_attachments": true,
 84      "options": {}
 85    },
 86    {
 87      "id": "deepseek-v3.2",
 88      "name": "DeepSeek V3.2",
 89      "cost_per_1m_in": 0.33,
 90      "cost_per_1m_out": 0.48,
 91      "cost_per_1m_in_cached": 0,
 92      "cost_per_1m_out_cached": 0,
 93      "context_window": 160000,
 94      "default_max_tokens": 32768,
 95      "can_reason": true,
 96      "reasoning_levels": [
 97        "low",
 98        "medium",
 99        "high"
100      ],
101      "default_reasoning_effort": "medium",
102      "supports_attachments": false,
103      "options": {}
104    },
105    {
106      "id": "zai-org-glm-4.6",
107      "name": "GLM 4.6",
108      "cost_per_1m_in": 0.85,
109      "cost_per_1m_out": 2.75,
110      "cost_per_1m_in_cached": 0,
111      "cost_per_1m_out_cached": 0,
112      "context_window": 198000,
113      "default_max_tokens": 32768,
114      "can_reason": false,
115      "supports_attachments": false,
116      "options": {}
117    },
118    {
119      "id": "zai-org-glm-4.7",
120      "name": "GLM 4.7",
121      "cost_per_1m_in": 0.55,
122      "cost_per_1m_out": 2.65,
123      "cost_per_1m_in_cached": 0,
124      "cost_per_1m_out_cached": 0,
125      "context_window": 198000,
126      "default_max_tokens": 32768,
127      "can_reason": true,
128      "reasoning_levels": [
129        "low",
130        "medium",
131        "high"
132      ],
133      "default_reasoning_effort": "medium",
134      "supports_attachments": false,
135      "options": {}
136    },
137    {
138      "id": "zai-org-glm-4.7-flash",
139      "name": "GLM 4.7 Flash",
140      "cost_per_1m_in": 0.125,
141      "cost_per_1m_out": 0.5,
142      "cost_per_1m_in_cached": 0,
143      "cost_per_1m_out_cached": 0,
144      "context_window": 128000,
145      "default_max_tokens": 32000,
146      "can_reason": true,
147      "reasoning_levels": [
148        "low",
149        "medium",
150        "high"
151      ],
152      "default_reasoning_effort": "medium",
153      "supports_attachments": false,
154      "options": {}
155    },
156    {
157      "id": "olafangensan-glm-4.7-flash-heretic",
158      "name": "GLM 4.7 Flash Heretic",
159      "cost_per_1m_in": 0.14,
160      "cost_per_1m_out": 0.8,
161      "cost_per_1m_in_cached": 0,
162      "cost_per_1m_out_cached": 0,
163      "context_window": 200000,
164      "default_max_tokens": 32768,
165      "can_reason": true,
166      "reasoning_levels": [
167        "low",
168        "medium",
169        "high"
170      ],
171      "default_reasoning_effort": "medium",
172      "supports_attachments": false,
173      "options": {}
174    },
175    {
176      "id": "zai-org-glm-5",
177      "name": "GLM 5",
178      "cost_per_1m_in": 1,
179      "cost_per_1m_out": 3.2,
180      "cost_per_1m_in_cached": 0,
181      "cost_per_1m_out_cached": 0,
182      "context_window": 198000,
183      "default_max_tokens": 32768,
184      "can_reason": true,
185      "reasoning_levels": [
186        "low",
187        "medium",
188        "high"
189      ],
190      "default_reasoning_effort": "medium",
191      "supports_attachments": false,
192      "options": {}
193    },
194    {
195      "id": "openai-gpt-4o-2024-11-20",
196      "name": "GPT-4o",
197      "cost_per_1m_in": 3.125,
198      "cost_per_1m_out": 12.5,
199      "cost_per_1m_in_cached": 0,
200      "cost_per_1m_out_cached": 0,
201      "context_window": 128000,
202      "default_max_tokens": 32000,
203      "can_reason": false,
204      "supports_attachments": true,
205      "options": {}
206    },
207    {
208      "id": "openai-gpt-4o-mini-2024-07-18",
209      "name": "GPT-4o Mini",
210      "cost_per_1m_in": 0.1875,
211      "cost_per_1m_out": 0.75,
212      "cost_per_1m_in_cached": 0,
213      "cost_per_1m_out_cached": 0,
214      "context_window": 128000,
215      "default_max_tokens": 32000,
216      "can_reason": false,
217      "supports_attachments": true,
218      "options": {}
219    },
220    {
221      "id": "openai-gpt-52",
222      "name": "GPT-5.2",
223      "cost_per_1m_in": 2.19,
224      "cost_per_1m_out": 17.5,
225      "cost_per_1m_in_cached": 0,
226      "cost_per_1m_out_cached": 0,
227      "context_window": 256000,
228      "default_max_tokens": 32768,
229      "can_reason": true,
230      "reasoning_levels": [
231        "low",
232        "medium",
233        "high"
234      ],
235      "default_reasoning_effort": "medium",
236      "supports_attachments": false,
237      "options": {}
238    },
239    {
240      "id": "openai-gpt-52-codex",
241      "name": "GPT-5.2 Codex",
242      "cost_per_1m_in": 2.19,
243      "cost_per_1m_out": 17.5,
244      "cost_per_1m_in_cached": 0,
245      "cost_per_1m_out_cached": 0,
246      "context_window": 256000,
247      "default_max_tokens": 32768,
248      "can_reason": true,
249      "reasoning_levels": [
250        "low",
251        "medium",
252        "high"
253      ],
254      "default_reasoning_effort": "medium",
255      "supports_attachments": true,
256      "options": {}
257    },
258    {
259      "id": "openai-gpt-53-codex",
260      "name": "GPT-5.3 Codex",
261      "cost_per_1m_in": 2.19,
262      "cost_per_1m_out": 17.5,
263      "cost_per_1m_in_cached": 0,
264      "cost_per_1m_out_cached": 0,
265      "context_window": 400000,
266      "default_max_tokens": 32768,
267      "can_reason": true,
268      "reasoning_levels": [
269        "low",
270        "medium",
271        "high"
272      ],
273      "default_reasoning_effort": "medium",
274      "supports_attachments": true,
275      "options": {}
276    },
277    {
278      "id": "openai-gpt-54",
279      "name": "GPT-5.4",
280      "cost_per_1m_in": 3.13,
281      "cost_per_1m_out": 18.8,
282      "cost_per_1m_in_cached": 0,
283      "cost_per_1m_out_cached": 0,
284      "context_window": 1000000,
285      "default_max_tokens": 32768,
286      "can_reason": true,
287      "reasoning_levels": [
288        "low",
289        "medium",
290        "high"
291      ],
292      "default_reasoning_effort": "medium",
293      "supports_attachments": true,
294      "options": {}
295    },
296    {
297      "id": "openai-gpt-54-pro",
298      "name": "GPT-5.4 Pro",
299      "cost_per_1m_in": 37.5,
300      "cost_per_1m_out": 225,
301      "cost_per_1m_in_cached": 0,
302      "cost_per_1m_out_cached": 0,
303      "context_window": 1000000,
304      "default_max_tokens": 32768,
305      "can_reason": true,
306      "reasoning_levels": [
307        "low",
308        "medium",
309        "high"
310      ],
311      "default_reasoning_effort": "medium",
312      "supports_attachments": true,
313      "options": {}
314    },
315    {
316      "id": "gemini-3-flash-preview",
317      "name": "Gemini 3 Flash Preview",
318      "cost_per_1m_in": 0.7,
319      "cost_per_1m_out": 3.75,
320      "cost_per_1m_in_cached": 0,
321      "cost_per_1m_out_cached": 0,
322      "context_window": 256000,
323      "default_max_tokens": 32768,
324      "can_reason": true,
325      "reasoning_levels": [
326        "low",
327        "medium",
328        "high"
329      ],
330      "default_reasoning_effort": "medium",
331      "supports_attachments": true,
332      "options": {}
333    },
334    {
335      "id": "gemini-3-pro-preview",
336      "name": "Gemini 3 Pro Preview",
337      "cost_per_1m_in": 2.5,
338      "cost_per_1m_out": 15,
339      "cost_per_1m_in_cached": 0,
340      "cost_per_1m_out_cached": 0,
341      "context_window": 198000,
342      "default_max_tokens": 32768,
343      "can_reason": true,
344      "reasoning_levels": [
345        "low",
346        "medium",
347        "high"
348      ],
349      "default_reasoning_effort": "medium",
350      "supports_attachments": true,
351      "options": {}
352    },
353    {
354      "id": "gemini-3-1-pro-preview",
355      "name": "Gemini 3.1 Pro Preview",
356      "cost_per_1m_in": 2.5,
357      "cost_per_1m_out": 15,
358      "cost_per_1m_in_cached": 0,
359      "cost_per_1m_out_cached": 0,
360      "context_window": 1000000,
361      "default_max_tokens": 32768,
362      "can_reason": true,
363      "reasoning_levels": [
364        "low",
365        "medium",
366        "high"
367      ],
368      "default_reasoning_effort": "medium",
369      "supports_attachments": true,
370      "options": {}
371    },
372    {
373      "id": "google-gemma-3-27b-it",
374      "name": "Google Gemma 3 27B Instruct",
375      "cost_per_1m_in": 0.12,
376      "cost_per_1m_out": 0.2,
377      "cost_per_1m_in_cached": 0,
378      "cost_per_1m_out_cached": 0,
379      "context_window": 198000,
380      "default_max_tokens": 32768,
381      "can_reason": false,
382      "supports_attachments": true,
383      "options": {}
384    },
385    {
386      "id": "grok-41-fast",
387      "name": "Grok 4.1 Fast",
388      "cost_per_1m_in": 0.25,
389      "cost_per_1m_out": 0.625,
390      "cost_per_1m_in_cached": 0,
391      "cost_per_1m_out_cached": 0,
392      "context_window": 1000000,
393      "default_max_tokens": 32768,
394      "can_reason": true,
395      "reasoning_levels": [
396        "low",
397        "medium",
398        "high"
399      ],
400      "default_reasoning_effort": "medium",
401      "supports_attachments": true,
402      "options": {}
403    },
404    {
405      "id": "grok-4-20-beta",
406      "name": "Grok 4.20 Beta",
407      "cost_per_1m_in": 2.5,
408      "cost_per_1m_out": 7.5,
409      "cost_per_1m_in_cached": 0,
410      "cost_per_1m_out_cached": 0,
411      "context_window": 2000000,
412      "default_max_tokens": 32768,
413      "can_reason": true,
414      "reasoning_levels": [
415        "low",
416        "medium",
417        "high"
418      ],
419      "default_reasoning_effort": "medium",
420      "supports_attachments": true,
421      "options": {}
422    },
423    {
424      "id": "grok-code-fast-1",
425      "name": "Grok Code Fast 1",
426      "cost_per_1m_in": 0.25,
427      "cost_per_1m_out": 1.87,
428      "cost_per_1m_in_cached": 0,
429      "cost_per_1m_out_cached": 0,
430      "context_window": 256000,
431      "default_max_tokens": 32768,
432      "can_reason": true,
433      "reasoning_levels": [
434        "low",
435        "medium",
436        "high"
437      ],
438      "default_reasoning_effort": "medium",
439      "supports_attachments": false,
440      "options": {}
441    },
442    {
443      "id": "kimi-k2-thinking",
444      "name": "Kimi K2 Thinking",
445      "cost_per_1m_in": 0.75,
446      "cost_per_1m_out": 3.2,
447      "cost_per_1m_in_cached": 0,
448      "cost_per_1m_out_cached": 0,
449      "context_window": 256000,
450      "default_max_tokens": 32768,
451      "can_reason": true,
452      "reasoning_levels": [
453        "low",
454        "medium",
455        "high"
456      ],
457      "default_reasoning_effort": "medium",
458      "supports_attachments": false,
459      "options": {}
460    },
461    {
462      "id": "kimi-k2-5",
463      "name": "Kimi K2.5",
464      "cost_per_1m_in": 0.56,
465      "cost_per_1m_out": 3.5,
466      "cost_per_1m_in_cached": 0,
467      "cost_per_1m_out_cached": 0,
468      "context_window": 256000,
469      "default_max_tokens": 32768,
470      "can_reason": true,
471      "reasoning_levels": [
472        "low",
473        "medium",
474        "high"
475      ],
476      "default_reasoning_effort": "medium",
477      "supports_attachments": true,
478      "options": {}
479    },
480    {
481      "id": "llama-3.2-3b",
482      "name": "Llama 3.2 3B",
483      "cost_per_1m_in": 0.15,
484      "cost_per_1m_out": 0.6,
485      "cost_per_1m_in_cached": 0,
486      "cost_per_1m_out_cached": 0,
487      "context_window": 128000,
488      "default_max_tokens": 32000,
489      "can_reason": false,
490      "supports_attachments": false,
491      "options": {}
492    },
493    {
494      "id": "llama-3.3-70b",
495      "name": "Llama 3.3 70B",
496      "cost_per_1m_in": 0.7,
497      "cost_per_1m_out": 2.8,
498      "cost_per_1m_in_cached": 0,
499      "cost_per_1m_out_cached": 0,
500      "context_window": 128000,
501      "default_max_tokens": 32000,
502      "can_reason": false,
503      "supports_attachments": false,
504      "options": {}
505    },
506    {
507      "id": "minimax-m21",
508      "name": "MiniMax M2.1",
509      "cost_per_1m_in": 0.35,
510      "cost_per_1m_out": 1.5,
511      "cost_per_1m_in_cached": 0,
512      "cost_per_1m_out_cached": 0,
513      "context_window": 198000,
514      "default_max_tokens": 32768,
515      "can_reason": true,
516      "reasoning_levels": [
517        "low",
518        "medium",
519        "high"
520      ],
521      "default_reasoning_effort": "medium",
522      "supports_attachments": false,
523      "options": {}
524    },
525    {
526      "id": "minimax-m25",
527      "name": "MiniMax M2.5",
528      "cost_per_1m_in": 0.34,
529      "cost_per_1m_out": 1.19,
530      "cost_per_1m_in_cached": 0,
531      "cost_per_1m_out_cached": 0,
532      "context_window": 198000,
533      "default_max_tokens": 32768,
534      "can_reason": true,
535      "reasoning_levels": [
536        "low",
537        "medium",
538        "high"
539      ],
540      "default_reasoning_effort": "medium",
541      "supports_attachments": false,
542      "options": {}
543    },
544    {
545      "id": "minimax-m27",
546      "name": "MiniMax M2.7",
547      "cost_per_1m_in": 0.375,
548      "cost_per_1m_out": 1.5,
549      "cost_per_1m_in_cached": 0,
550      "cost_per_1m_out_cached": 0,
551      "context_window": 198000,
552      "default_max_tokens": 32768,
553      "can_reason": true,
554      "reasoning_levels": [
555        "low",
556        "medium",
557        "high"
558      ],
559      "default_reasoning_effort": "medium",
560      "supports_attachments": false,
561      "options": {}
562    },
563    {
564      "id": "mistral-small-3-2-24b-instruct",
565      "name": "Mistral Small 3.2 24B Instruct",
566      "cost_per_1m_in": 0.09375,
567      "cost_per_1m_out": 0.25,
568      "cost_per_1m_in_cached": 0,
569      "cost_per_1m_out_cached": 0,
570      "context_window": 256000,
571      "default_max_tokens": 32768,
572      "can_reason": false,
573      "supports_attachments": false,
574      "options": {}
575    },
576    {
577      "id": "nvidia-nemotron-3-nano-30b-a3b",
578      "name": "NVIDIA Nemotron 3 Nano 30B",
579      "cost_per_1m_in": 0.075,
580      "cost_per_1m_out": 0.3,
581      "cost_per_1m_in_cached": 0,
582      "cost_per_1m_out_cached": 0,
583      "context_window": 128000,
584      "default_max_tokens": 32000,
585      "can_reason": false,
586      "supports_attachments": false,
587      "options": {}
588    },
589    {
590      "id": "openai-gpt-oss-120b",
591      "name": "OpenAI GPT OSS 120B",
592      "cost_per_1m_in": 0.07,
593      "cost_per_1m_out": 0.3,
594      "cost_per_1m_in_cached": 0,
595      "cost_per_1m_out_cached": 0,
596      "context_window": 128000,
597      "default_max_tokens": 32000,
598      "can_reason": false,
599      "supports_attachments": false,
600      "options": {}
601    },
602    {
603      "id": "qwen3-235b-a22b-instruct-2507",
604      "name": "Qwen 3 235B A22B Instruct 2507",
605      "cost_per_1m_in": 0.15,
606      "cost_per_1m_out": 0.75,
607      "cost_per_1m_in_cached": 0,
608      "cost_per_1m_out_cached": 0,
609      "context_window": 128000,
610      "default_max_tokens": 32000,
611      "can_reason": false,
612      "supports_attachments": false,
613      "options": {}
614    },
615    {
616      "id": "qwen3-235b-a22b-thinking-2507",
617      "name": "Qwen 3 235B A22B Thinking 2507",
618      "cost_per_1m_in": 0.45,
619      "cost_per_1m_out": 3.5,
620      "cost_per_1m_in_cached": 0,
621      "cost_per_1m_out_cached": 0,
622      "context_window": 128000,
623      "default_max_tokens": 32000,
624      "can_reason": true,
625      "reasoning_levels": [
626        "low",
627        "medium",
628        "high"
629      ],
630      "default_reasoning_effort": "medium",
631      "supports_attachments": false,
632      "options": {}
633    },
634    {
635      "id": "qwen3-coder-480b-a35b-instruct-turbo",
636      "name": "Qwen 3 Coder 480B Turbo",
637      "cost_per_1m_in": 0.35,
638      "cost_per_1m_out": 1.5,
639      "cost_per_1m_in_cached": 0,
640      "cost_per_1m_out_cached": 0,
641      "context_window": 256000,
642      "default_max_tokens": 32768,
643      "can_reason": false,
644      "supports_attachments": false,
645      "options": {}
646    },
647    {
648      "id": "qwen3-coder-480b-a35b-instruct",
649      "name": "Qwen 3 Coder 480b",
650      "cost_per_1m_in": 0.75,
651      "cost_per_1m_out": 3,
652      "cost_per_1m_in_cached": 0,
653      "cost_per_1m_out_cached": 0,
654      "context_window": 256000,
655      "default_max_tokens": 32768,
656      "can_reason": false,
657      "supports_attachments": false,
658      "options": {}
659    },
660    {
661      "id": "qwen3-next-80b",
662      "name": "Qwen 3 Next 80b",
663      "cost_per_1m_in": 0.35,
664      "cost_per_1m_out": 1.9,
665      "cost_per_1m_in_cached": 0,
666      "cost_per_1m_out_cached": 0,
667      "context_window": 256000,
668      "default_max_tokens": 32768,
669      "can_reason": false,
670      "supports_attachments": false,
671      "options": {}
672    },
673    {
674      "id": "qwen3-5-35b-a3b",
675      "name": "Qwen 3.5 35B A3B",
676      "cost_per_1m_in": 0.3125,
677      "cost_per_1m_out": 1.25,
678      "cost_per_1m_in_cached": 0,
679      "cost_per_1m_out_cached": 0,
680      "context_window": 256000,
681      "default_max_tokens": 32768,
682      "can_reason": true,
683      "reasoning_levels": [
684        "low",
685        "medium",
686        "high"
687      ],
688      "default_reasoning_effort": "medium",
689      "supports_attachments": true,
690      "options": {
691        "temperature": 1,
692        "top_p": 0.95
693      }
694    },
695    {
696      "id": "qwen3-5-9b",
697      "name": "Qwen 3.5 9B",
698      "cost_per_1m_in": 0.05,
699      "cost_per_1m_out": 0.15,
700      "cost_per_1m_in_cached": 0,
701      "cost_per_1m_out_cached": 0,
702      "context_window": 256000,
703      "default_max_tokens": 32768,
704      "can_reason": true,
705      "reasoning_levels": [
706        "low",
707        "medium",
708        "high"
709      ],
710      "default_reasoning_effort": "medium",
711      "supports_attachments": true,
712      "options": {}
713    },
714    {
715      "id": "qwen3-vl-235b-a22b",
716      "name": "Qwen3 VL 235B",
717      "cost_per_1m_in": 0.25,
718      "cost_per_1m_out": 1.5,
719      "cost_per_1m_in_cached": 0,
720      "cost_per_1m_out_cached": 0,
721      "context_window": 256000,
722      "default_max_tokens": 32768,
723      "can_reason": false,
724      "supports_attachments": true,
725      "options": {}
726    },
727    {
728      "id": "mistral-31-24b",
729      "name": "Venice Medium",
730      "cost_per_1m_in": 0.5,
731      "cost_per_1m_out": 2,
732      "cost_per_1m_in_cached": 0,
733      "cost_per_1m_out_cached": 0,
734      "context_window": 128000,
735      "default_max_tokens": 32000,
736      "can_reason": false,
737      "supports_attachments": true,
738      "options": {}
739    },
740    {
741      "id": "venice-uncensored-role-play",
742      "name": "Venice Role Play Uncensored",
743      "cost_per_1m_in": 0.5,
744      "cost_per_1m_out": 2,
745      "cost_per_1m_in_cached": 0,
746      "cost_per_1m_out_cached": 0,
747      "context_window": 128000,
748      "default_max_tokens": 32000,
749      "can_reason": false,
750      "supports_attachments": true,
751      "options": {}
752    },
753    {
754      "id": "qwen3-4b",
755      "name": "Venice Small",
756      "cost_per_1m_in": 0.05,
757      "cost_per_1m_out": 0.15,
758      "cost_per_1m_in_cached": 0,
759      "cost_per_1m_out_cached": 0,
760      "context_window": 32000,
761      "default_max_tokens": 8000,
762      "can_reason": true,
763      "reasoning_levels": [
764        "low",
765        "medium",
766        "high"
767      ],
768      "default_reasoning_effort": "medium",
769      "supports_attachments": false,
770      "options": {}
771    }
772  ]
773}