venice.json

  1{
  2  "name": "Venice AI",
  3  "id": "venice",
  4  "api_key": "$VENICE_API_KEY",
  5  "api_endpoint": "https://api.venice.ai/api/v1",
  6  "type": "openai-compat",
  7  "default_large_model_id": "claude-opus-4-6",
  8  "default_small_model_id": "minimax-m25",
  9  "models": [
 10    {
 11      "id": "claude-opus-4-5",
 12      "name": "Claude Opus 4.5",
 13      "cost_per_1m_in": 6,
 14      "cost_per_1m_out": 30,
 15      "cost_per_1m_in_cached": 0,
 16      "cost_per_1m_out_cached": 0,
 17      "context_window": 198000,
 18      "default_max_tokens": 32768,
 19      "can_reason": true,
 20      "reasoning_levels": [
 21        "low",
 22        "medium",
 23        "high"
 24      ],
 25      "default_reasoning_effort": "medium",
 26      "supports_attachments": true,
 27      "options": {}
 28    },
 29    {
 30      "id": "claude-opus-4-6",
 31      "name": "Claude Opus 4.6",
 32      "cost_per_1m_in": 6,
 33      "cost_per_1m_out": 30,
 34      "cost_per_1m_in_cached": 0,
 35      "cost_per_1m_out_cached": 0,
 36      "context_window": 1000000,
 37      "default_max_tokens": 32768,
 38      "can_reason": true,
 39      "reasoning_levels": [
 40        "low",
 41        "medium",
 42        "high"
 43      ],
 44      "default_reasoning_effort": "medium",
 45      "supports_attachments": true,
 46      "options": {}
 47    },
 48    {
 49      "id": "claude-sonnet-4-5",
 50      "name": "Claude Sonnet 4.5",
 51      "cost_per_1m_in": 3.75,
 52      "cost_per_1m_out": 18.75,
 53      "cost_per_1m_in_cached": 0,
 54      "cost_per_1m_out_cached": 0,
 55      "context_window": 198000,
 56      "default_max_tokens": 32768,
 57      "can_reason": true,
 58      "reasoning_levels": [
 59        "low",
 60        "medium",
 61        "high"
 62      ],
 63      "default_reasoning_effort": "medium",
 64      "supports_attachments": true,
 65      "options": {}
 66    },
 67    {
 68      "id": "claude-sonnet-4-6",
 69      "name": "Claude Sonnet 4.6",
 70      "cost_per_1m_in": 3.6,
 71      "cost_per_1m_out": 18,
 72      "cost_per_1m_in_cached": 0,
 73      "cost_per_1m_out_cached": 0,
 74      "context_window": 1000000,
 75      "default_max_tokens": 32768,
 76      "can_reason": true,
 77      "reasoning_levels": [
 78        "low",
 79        "medium",
 80        "high"
 81      ],
 82      "default_reasoning_effort": "medium",
 83      "supports_attachments": true,
 84      "options": {}
 85    },
 86    {
 87      "id": "deepseek-v3.2",
 88      "name": "DeepSeek V3.2",
 89      "cost_per_1m_in": 0.33,
 90      "cost_per_1m_out": 0.48,
 91      "cost_per_1m_in_cached": 0,
 92      "cost_per_1m_out_cached": 0,
 93      "context_window": 160000,
 94      "default_max_tokens": 32768,
 95      "can_reason": true,
 96      "reasoning_levels": [
 97        "low",
 98        "medium",
 99        "high"
100      ],
101      "default_reasoning_effort": "medium",
102      "supports_attachments": false,
103      "options": {}
104    },
105    {
106      "id": "zai-org-glm-4.6",
107      "name": "GLM 4.6",
108      "cost_per_1m_in": 0.85,
109      "cost_per_1m_out": 2.75,
110      "cost_per_1m_in_cached": 0,
111      "cost_per_1m_out_cached": 0,
112      "context_window": 198000,
113      "default_max_tokens": 32768,
114      "can_reason": false,
115      "supports_attachments": false,
116      "options": {}
117    },
118    {
119      "id": "zai-org-glm-4.7",
120      "name": "GLM 4.7",
121      "cost_per_1m_in": 0.55,
122      "cost_per_1m_out": 2.65,
123      "cost_per_1m_in_cached": 0,
124      "cost_per_1m_out_cached": 0,
125      "context_window": 198000,
126      "default_max_tokens": 32768,
127      "can_reason": true,
128      "reasoning_levels": [
129        "low",
130        "medium",
131        "high"
132      ],
133      "default_reasoning_effort": "medium",
134      "supports_attachments": false,
135      "options": {}
136    },
137    {
138      "id": "zai-org-glm-4.7-flash",
139      "name": "GLM 4.7 Flash",
140      "cost_per_1m_in": 0.125,
141      "cost_per_1m_out": 0.5,
142      "cost_per_1m_in_cached": 0,
143      "cost_per_1m_out_cached": 0,
144      "context_window": 128000,
145      "default_max_tokens": 32000,
146      "can_reason": true,
147      "reasoning_levels": [
148        "low",
149        "medium",
150        "high"
151      ],
152      "default_reasoning_effort": "medium",
153      "supports_attachments": false,
154      "options": {}
155    },
156    {
157      "id": "olafangensan-glm-4.7-flash-heretic",
158      "name": "GLM 4.7 Flash Heretic",
159      "cost_per_1m_in": 0.14,
160      "cost_per_1m_out": 0.8,
161      "cost_per_1m_in_cached": 0,
162      "cost_per_1m_out_cached": 0,
163      "context_window": 200000,
164      "default_max_tokens": 32768,
165      "can_reason": true,
166      "reasoning_levels": [
167        "low",
168        "medium",
169        "high"
170      ],
171      "default_reasoning_effort": "medium",
172      "supports_attachments": false,
173      "options": {}
174    },
175    {
176      "id": "zai-org-glm-5",
177      "name": "GLM 5",
178      "cost_per_1m_in": 1,
179      "cost_per_1m_out": 3.2,
180      "cost_per_1m_in_cached": 0,
181      "cost_per_1m_out_cached": 0,
182      "context_window": 198000,
183      "default_max_tokens": 32768,
184      "can_reason": true,
185      "reasoning_levels": [
186        "low",
187        "medium",
188        "high"
189      ],
190      "default_reasoning_effort": "medium",
191      "supports_attachments": false,
192      "options": {}
193    },
194    {
195      "id": "openai-gpt-4o-2024-11-20",
196      "name": "GPT-4o",
197      "cost_per_1m_in": 3.125,
198      "cost_per_1m_out": 12.5,
199      "cost_per_1m_in_cached": 0,
200      "cost_per_1m_out_cached": 0,
201      "context_window": 128000,
202      "default_max_tokens": 32000,
203      "can_reason": false,
204      "supports_attachments": true,
205      "options": {}
206    },
207    {
208      "id": "openai-gpt-4o-mini-2024-07-18",
209      "name": "GPT-4o Mini",
210      "cost_per_1m_in": 0.1875,
211      "cost_per_1m_out": 0.75,
212      "cost_per_1m_in_cached": 0,
213      "cost_per_1m_out_cached": 0,
214      "context_window": 128000,
215      "default_max_tokens": 32000,
216      "can_reason": false,
217      "supports_attachments": true,
218      "options": {}
219    },
220    {
221      "id": "openai-gpt-52",
222      "name": "GPT-5.2",
223      "cost_per_1m_in": 2.19,
224      "cost_per_1m_out": 17.5,
225      "cost_per_1m_in_cached": 0,
226      "cost_per_1m_out_cached": 0,
227      "context_window": 256000,
228      "default_max_tokens": 32768,
229      "can_reason": true,
230      "reasoning_levels": [
231        "low",
232        "medium",
233        "high"
234      ],
235      "default_reasoning_effort": "medium",
236      "supports_attachments": false,
237      "options": {}
238    },
239    {
240      "id": "openai-gpt-52-codex",
241      "name": "GPT-5.2 Codex",
242      "cost_per_1m_in": 2.19,
243      "cost_per_1m_out": 17.5,
244      "cost_per_1m_in_cached": 0,
245      "cost_per_1m_out_cached": 0,
246      "context_window": 256000,
247      "default_max_tokens": 32768,
248      "can_reason": true,
249      "reasoning_levels": [
250        "low",
251        "medium",
252        "high"
253      ],
254      "default_reasoning_effort": "medium",
255      "supports_attachments": true,
256      "options": {}
257    },
258    {
259      "id": "openai-gpt-53-codex",
260      "name": "GPT-5.3 Codex",
261      "cost_per_1m_in": 2.19,
262      "cost_per_1m_out": 17.5,
263      "cost_per_1m_in_cached": 0,
264      "cost_per_1m_out_cached": 0,
265      "context_window": 400000,
266      "default_max_tokens": 32768,
267      "can_reason": true,
268      "reasoning_levels": [
269        "low",
270        "medium",
271        "high"
272      ],
273      "default_reasoning_effort": "medium",
274      "supports_attachments": true,
275      "options": {}
276    },
277    {
278      "id": "openai-gpt-54",
279      "name": "GPT-5.4",
280      "cost_per_1m_in": 3.13,
281      "cost_per_1m_out": 18.8,
282      "cost_per_1m_in_cached": 0,
283      "cost_per_1m_out_cached": 0,
284      "context_window": 1000000,
285      "default_max_tokens": 32768,
286      "can_reason": true,
287      "reasoning_levels": [
288        "low",
289        "medium",
290        "high"
291      ],
292      "default_reasoning_effort": "medium",
293      "supports_attachments": true,
294      "options": {}
295    },
296    {
297      "id": "openai-gpt-54-pro",
298      "name": "GPT-5.4 Pro",
299      "cost_per_1m_in": 37.5,
300      "cost_per_1m_out": 225,
301      "cost_per_1m_in_cached": 0,
302      "cost_per_1m_out_cached": 0,
303      "context_window": 1000000,
304      "default_max_tokens": 32768,
305      "can_reason": true,
306      "reasoning_levels": [
307        "low",
308        "medium",
309        "high"
310      ],
311      "default_reasoning_effort": "medium",
312      "supports_attachments": true,
313      "options": {}
314    },
315    {
316      "id": "gemini-3-flash-preview",
317      "name": "Gemini 3 Flash Preview",
318      "cost_per_1m_in": 0.7,
319      "cost_per_1m_out": 3.75,
320      "cost_per_1m_in_cached": 0,
321      "cost_per_1m_out_cached": 0,
322      "context_window": 256000,
323      "default_max_tokens": 32768,
324      "can_reason": true,
325      "reasoning_levels": [
326        "low",
327        "medium",
328        "high"
329      ],
330      "default_reasoning_effort": "medium",
331      "supports_attachments": true,
332      "options": {}
333    },
334    {
335      "id": "gemini-3-1-pro-preview",
336      "name": "Gemini 3.1 Pro Preview",
337      "cost_per_1m_in": 2.5,
338      "cost_per_1m_out": 15,
339      "cost_per_1m_in_cached": 0,
340      "cost_per_1m_out_cached": 0,
341      "context_window": 1000000,
342      "default_max_tokens": 32768,
343      "can_reason": true,
344      "reasoning_levels": [
345        "low",
346        "medium",
347        "high"
348      ],
349      "default_reasoning_effort": "medium",
350      "supports_attachments": true,
351      "options": {}
352    },
353    {
354      "id": "google-gemma-3-27b-it",
355      "name": "Google Gemma 3 27B Instruct",
356      "cost_per_1m_in": 0.12,
357      "cost_per_1m_out": 0.2,
358      "cost_per_1m_in_cached": 0,
359      "cost_per_1m_out_cached": 0,
360      "context_window": 198000,
361      "default_max_tokens": 32768,
362      "can_reason": false,
363      "supports_attachments": true,
364      "options": {}
365    },
366    {
367      "id": "grok-41-fast",
368      "name": "Grok 4.1 Fast",
369      "cost_per_1m_in": 0.25,
370      "cost_per_1m_out": 0.625,
371      "cost_per_1m_in_cached": 0,
372      "cost_per_1m_out_cached": 0,
373      "context_window": 1000000,
374      "default_max_tokens": 32768,
375      "can_reason": true,
376      "reasoning_levels": [
377        "low",
378        "medium",
379        "high"
380      ],
381      "default_reasoning_effort": "medium",
382      "supports_attachments": true,
383      "options": {}
384    },
385    {
386      "id": "grok-4-20-beta",
387      "name": "Grok 4.20 Beta",
388      "cost_per_1m_in": 2.5,
389      "cost_per_1m_out": 7.5,
390      "cost_per_1m_in_cached": 0,
391      "cost_per_1m_out_cached": 0,
392      "context_window": 2000000,
393      "default_max_tokens": 32768,
394      "can_reason": true,
395      "reasoning_levels": [
396        "low",
397        "medium",
398        "high"
399      ],
400      "default_reasoning_effort": "medium",
401      "supports_attachments": true,
402      "options": {}
403    },
404    {
405      "id": "grok-code-fast-1",
406      "name": "Grok Code Fast 1",
407      "cost_per_1m_in": 0.25,
408      "cost_per_1m_out": 1.87,
409      "cost_per_1m_in_cached": 0,
410      "cost_per_1m_out_cached": 0,
411      "context_window": 256000,
412      "default_max_tokens": 32768,
413      "can_reason": true,
414      "reasoning_levels": [
415        "low",
416        "medium",
417        "high"
418      ],
419      "default_reasoning_effort": "medium",
420      "supports_attachments": false,
421      "options": {}
422    },
423    {
424      "id": "kimi-k2-thinking",
425      "name": "Kimi K2 Thinking",
426      "cost_per_1m_in": 0.75,
427      "cost_per_1m_out": 3.2,
428      "cost_per_1m_in_cached": 0,
429      "cost_per_1m_out_cached": 0,
430      "context_window": 256000,
431      "default_max_tokens": 32768,
432      "can_reason": true,
433      "reasoning_levels": [
434        "low",
435        "medium",
436        "high"
437      ],
438      "default_reasoning_effort": "medium",
439      "supports_attachments": false,
440      "options": {}
441    },
442    {
443      "id": "kimi-k2-5",
444      "name": "Kimi K2.5",
445      "cost_per_1m_in": 0.56,
446      "cost_per_1m_out": 3.5,
447      "cost_per_1m_in_cached": 0,
448      "cost_per_1m_out_cached": 0,
449      "context_window": 256000,
450      "default_max_tokens": 32768,
451      "can_reason": true,
452      "reasoning_levels": [
453        "low",
454        "medium",
455        "high"
456      ],
457      "default_reasoning_effort": "medium",
458      "supports_attachments": true,
459      "options": {}
460    },
461    {
462      "id": "llama-3.2-3b",
463      "name": "Llama 3.2 3B",
464      "cost_per_1m_in": 0.15,
465      "cost_per_1m_out": 0.6,
466      "cost_per_1m_in_cached": 0,
467      "cost_per_1m_out_cached": 0,
468      "context_window": 128000,
469      "default_max_tokens": 32000,
470      "can_reason": false,
471      "supports_attachments": false,
472      "options": {}
473    },
474    {
475      "id": "llama-3.3-70b",
476      "name": "Llama 3.3 70B",
477      "cost_per_1m_in": 0.7,
478      "cost_per_1m_out": 2.8,
479      "cost_per_1m_in_cached": 0,
480      "cost_per_1m_out_cached": 0,
481      "context_window": 128000,
482      "default_max_tokens": 32000,
483      "can_reason": false,
484      "supports_attachments": false,
485      "options": {}
486    },
487    {
488      "id": "minimax-m21",
489      "name": "MiniMax M2.1",
490      "cost_per_1m_in": 0.35,
491      "cost_per_1m_out": 1.5,
492      "cost_per_1m_in_cached": 0,
493      "cost_per_1m_out_cached": 0,
494      "context_window": 198000,
495      "default_max_tokens": 32768,
496      "can_reason": true,
497      "reasoning_levels": [
498        "low",
499        "medium",
500        "high"
501      ],
502      "default_reasoning_effort": "medium",
503      "supports_attachments": false,
504      "options": {}
505    },
506    {
507      "id": "minimax-m25",
508      "name": "MiniMax M2.5",
509      "cost_per_1m_in": 0.34,
510      "cost_per_1m_out": 1.19,
511      "cost_per_1m_in_cached": 0,
512      "cost_per_1m_out_cached": 0,
513      "context_window": 198000,
514      "default_max_tokens": 32768,
515      "can_reason": true,
516      "reasoning_levels": [
517        "low",
518        "medium",
519        "high"
520      ],
521      "default_reasoning_effort": "medium",
522      "supports_attachments": false,
523      "options": {}
524    },
525    {
526      "id": "minimax-m27",
527      "name": "MiniMax M2.7",
528      "cost_per_1m_in": 0.375,
529      "cost_per_1m_out": 1.5,
530      "cost_per_1m_in_cached": 0,
531      "cost_per_1m_out_cached": 0,
532      "context_window": 198000,
533      "default_max_tokens": 32768,
534      "can_reason": true,
535      "reasoning_levels": [
536        "low",
537        "medium",
538        "high"
539      ],
540      "default_reasoning_effort": "medium",
541      "supports_attachments": false,
542      "options": {}
543    },
544    {
545      "id": "mistral-small-3-2-24b-instruct",
546      "name": "Mistral Small 3.2 24B Instruct",
547      "cost_per_1m_in": 0.09375,
548      "cost_per_1m_out": 0.25,
549      "cost_per_1m_in_cached": 0,
550      "cost_per_1m_out_cached": 0,
551      "context_window": 256000,
552      "default_max_tokens": 32768,
553      "can_reason": false,
554      "supports_attachments": false,
555      "options": {}
556    },
557    {
558      "id": "nvidia-nemotron-3-nano-30b-a3b",
559      "name": "NVIDIA Nemotron 3 Nano 30B",
560      "cost_per_1m_in": 0.075,
561      "cost_per_1m_out": 0.3,
562      "cost_per_1m_in_cached": 0,
563      "cost_per_1m_out_cached": 0,
564      "context_window": 128000,
565      "default_max_tokens": 32000,
566      "can_reason": false,
567      "supports_attachments": false,
568      "options": {}
569    },
570    {
571      "id": "openai-gpt-oss-120b",
572      "name": "OpenAI GPT OSS 120B",
573      "cost_per_1m_in": 0.07,
574      "cost_per_1m_out": 0.3,
575      "cost_per_1m_in_cached": 0,
576      "cost_per_1m_out_cached": 0,
577      "context_window": 128000,
578      "default_max_tokens": 32000,
579      "can_reason": false,
580      "supports_attachments": false,
581      "options": {}
582    },
583    {
584      "id": "qwen3-235b-a22b-instruct-2507",
585      "name": "Qwen 3 235B A22B Instruct 2507",
586      "cost_per_1m_in": 0.15,
587      "cost_per_1m_out": 0.75,
588      "cost_per_1m_in_cached": 0,
589      "cost_per_1m_out_cached": 0,
590      "context_window": 128000,
591      "default_max_tokens": 32000,
592      "can_reason": false,
593      "supports_attachments": false,
594      "options": {}
595    },
596    {
597      "id": "qwen3-235b-a22b-thinking-2507",
598      "name": "Qwen 3 235B A22B Thinking 2507",
599      "cost_per_1m_in": 0.45,
600      "cost_per_1m_out": 3.5,
601      "cost_per_1m_in_cached": 0,
602      "cost_per_1m_out_cached": 0,
603      "context_window": 128000,
604      "default_max_tokens": 32000,
605      "can_reason": true,
606      "reasoning_levels": [
607        "low",
608        "medium",
609        "high"
610      ],
611      "default_reasoning_effort": "medium",
612      "supports_attachments": false,
613      "options": {}
614    },
615    {
616      "id": "qwen3-coder-480b-a35b-instruct-turbo",
617      "name": "Qwen 3 Coder 480B Turbo",
618      "cost_per_1m_in": 0.35,
619      "cost_per_1m_out": 1.5,
620      "cost_per_1m_in_cached": 0,
621      "cost_per_1m_out_cached": 0,
622      "context_window": 256000,
623      "default_max_tokens": 32768,
624      "can_reason": false,
625      "supports_attachments": false,
626      "options": {}
627    },
628    {
629      "id": "qwen3-coder-480b-a35b-instruct",
630      "name": "Qwen 3 Coder 480b",
631      "cost_per_1m_in": 0.75,
632      "cost_per_1m_out": 3,
633      "cost_per_1m_in_cached": 0,
634      "cost_per_1m_out_cached": 0,
635      "context_window": 256000,
636      "default_max_tokens": 32768,
637      "can_reason": false,
638      "supports_attachments": false,
639      "options": {}
640    },
641    {
642      "id": "qwen3-next-80b",
643      "name": "Qwen 3 Next 80b",
644      "cost_per_1m_in": 0.35,
645      "cost_per_1m_out": 1.9,
646      "cost_per_1m_in_cached": 0,
647      "cost_per_1m_out_cached": 0,
648      "context_window": 256000,
649      "default_max_tokens": 32768,
650      "can_reason": false,
651      "supports_attachments": false,
652      "options": {}
653    },
654    {
655      "id": "qwen3-5-35b-a3b",
656      "name": "Qwen 3.5 35B A3B",
657      "cost_per_1m_in": 0.3125,
658      "cost_per_1m_out": 1.25,
659      "cost_per_1m_in_cached": 0,
660      "cost_per_1m_out_cached": 0,
661      "context_window": 256000,
662      "default_max_tokens": 32768,
663      "can_reason": true,
664      "reasoning_levels": [
665        "low",
666        "medium",
667        "high"
668      ],
669      "default_reasoning_effort": "medium",
670      "supports_attachments": true,
671      "options": {
672        "temperature": 1,
673        "top_p": 0.95
674      }
675    },
676    {
677      "id": "qwen3-5-9b",
678      "name": "Qwen 3.5 9B",
679      "cost_per_1m_in": 0.05,
680      "cost_per_1m_out": 0.15,
681      "cost_per_1m_in_cached": 0,
682      "cost_per_1m_out_cached": 0,
683      "context_window": 256000,
684      "default_max_tokens": 32768,
685      "can_reason": true,
686      "reasoning_levels": [
687        "low",
688        "medium",
689        "high"
690      ],
691      "default_reasoning_effort": "medium",
692      "supports_attachments": true,
693      "options": {}
694    },
695    {
696      "id": "qwen3-vl-235b-a22b",
697      "name": "Qwen3 VL 235B",
698      "cost_per_1m_in": 0.25,
699      "cost_per_1m_out": 1.5,
700      "cost_per_1m_in_cached": 0,
701      "cost_per_1m_out_cached": 0,
702      "context_window": 256000,
703      "default_max_tokens": 32768,
704      "can_reason": false,
705      "supports_attachments": true,
706      "options": {}
707    },
708    {
709      "id": "mistral-31-24b",
710      "name": "Venice Medium",
711      "cost_per_1m_in": 0.5,
712      "cost_per_1m_out": 2,
713      "cost_per_1m_in_cached": 0,
714      "cost_per_1m_out_cached": 0,
715      "context_window": 128000,
716      "default_max_tokens": 32000,
717      "can_reason": false,
718      "supports_attachments": true,
719      "options": {}
720    },
721    {
722      "id": "venice-uncensored-role-play",
723      "name": "Venice Role Play Uncensored",
724      "cost_per_1m_in": 0.5,
725      "cost_per_1m_out": 2,
726      "cost_per_1m_in_cached": 0,
727      "cost_per_1m_out_cached": 0,
728      "context_window": 128000,
729      "default_max_tokens": 32000,
730      "can_reason": false,
731      "supports_attachments": true,
732      "options": {}
733    },
734    {
735      "id": "qwen3-4b",
736      "name": "Venice Small",
737      "cost_per_1m_in": 0.05,
738      "cost_per_1m_out": 0.15,
739      "cost_per_1m_in_cached": 0,
740      "cost_per_1m_out_cached": 0,
741      "context_window": 32000,
742      "default_max_tokens": 8000,
743      "can_reason": true,
744      "reasoning_levels": [
745        "low",
746        "medium",
747        "high"
748      ],
749      "default_reasoning_effort": "medium",
750      "supports_attachments": false,
751      "options": {}
752    }
753  ]
754}