1{
2 "name": "Venice AI",
3 "id": "venice",
4 "type": "openai-compat",
5 "api_key": "$VENICE_API_KEY",
6 "api_endpoint": "https://api.venice.ai/api/v1",
7 "default_large_model_id": "qwen3-235b:strip_thinking_response=true",
8 "default_small_model_id": "mistral-31-24b",
9 "models": [
10 {
11 "id": "qwen3-235b:strip_thinking_response=true",
12 "name": "Venice Large (qwen3-235b)",
13 "cost_per_1m_in": 1.5,
14 "cost_per_1m_out": 6,
15 "cost_per_1m_in_cached": 0,
16 "cost_per_1m_out_cached": 0,
17 "context_window": 131072,
18 "default_max_tokens": 50000,
19 "can_reason": true,
20 "supports_attachments": false
21 },
22 {
23 "id": "qwen3-4b:strip_thinking_response=true",
24 "name": "Venice Small (qwen3-4b)",
25 "cost_per_1m_in": 0.15,
26 "cost_per_1m_out": 0.6,
27 "cost_per_1m_in_cached": 0,
28 "cost_per_1m_out_cached": 0,
29 "context_window": 32768,
30 "default_max_tokens": 25000,
31 "can_reason": true,
32 "supports_attachments": false
33 },
34 {
35 "id": "mistral-31-24b",
36 "name": "Venice Medium (mistral-31-24b)",
37 "cost_per_1m_in": 0.5,
38 "cost_per_1m_out": 2,
39 "cost_per_1m_in_cached": 0,
40 "cost_per_1m_out_cached": 0,
41 "context_window": 131072,
42 "default_max_tokens": 50000,
43 "can_reason": false,
44 "supports_attachments": true
45 },
46 {
47 "id": "llama-3.2-3b",
48 "name": "Llama 3.2 3B",
49 "cost_per_1m_in": 0.15,
50 "cost_per_1m_out": 0.6,
51 "cost_per_1m_in_cached": 0,
52 "cost_per_1m_out_cached": 0,
53 "context_window": 131072,
54 "default_max_tokens": 25000,
55 "can_reason": false,
56 "supports_attachments": false
57 },
58 {
59 "id": "llama-3.3-70b",
60 "name": "Llama 3.3 70B",
61 "cost_per_1m_in": 0.7,
62 "cost_per_1m_out": 2.8,
63 "cost_per_1m_in_cached": 0,
64 "cost_per_1m_out_cached": 0,
65 "context_window": 65536,
66 "default_max_tokens": 32000,
67 "can_reason": false,
68 "supports_attachments": false
69 }
70 ]
71}