1{
2 "name": "io.net",
3 "id": "ionet",
4 "api_key": "$IONET_API_KEY",
5 "api_endpoint": "https://api.intelligence.io.solutions/api/v1",
6 "type": "openai-compat",
7 "default_large_model_id": "zai-org/GLM-4.7",
8 "default_small_model_id": "zai-org/GLM-4.7-Flash",
9 "models": [
10 {
11 "id": "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar",
12 "name": "Intel: Qwen3 Coder 480B A35B Instruct INT4 Mixed AR",
13 "cost_per_1m_in": 0.22,
14 "cost_per_1m_out": 0.95,
15 "cost_per_1m_in_cached": 0.11,
16 "cost_per_1m_out_cached": 0.44,
17 "context_window": 106000,
18 "default_max_tokens": 10600,
19 "can_reason": false,
20 "supports_attachments": false,
21 "options": {}
22 },
23 {
24 "id": "meta-llama/Llama-3.3-70B-Instruct",
25 "name": "Meta: Llama 3.3 70B Instruct",
26 "cost_per_1m_in": 0.1,
27 "cost_per_1m_out": 0.32,
28 "cost_per_1m_in_cached": 0.05,
29 "cost_per_1m_out_cached": 0.2,
30 "context_window": 128000,
31 "default_max_tokens": 12800,
32 "can_reason": true,
33 "reasoning_levels": [
34 "low",
35 "medium",
36 "high"
37 ],
38 "default_reasoning_effort": "medium",
39 "supports_attachments": false,
40 "options": {}
41 },
42 {
43 "id": "mistralai/Mistral-Large-Instruct-2411",
44 "name": "Mistral: Mistral Large Instruct 2411",
45 "cost_per_1m_in": 2,
46 "cost_per_1m_out": 6,
47 "cost_per_1m_in_cached": 1,
48 "cost_per_1m_out_cached": 4,
49 "context_window": 128000,
50 "default_max_tokens": 12800,
51 "can_reason": false,
52 "supports_attachments": true,
53 "options": {}
54 },
55 {
56 "id": "moonshotai/Kimi-K2-Instruct-0905",
57 "name": "MoonshotAI: Kimi K2 Instruct 0905",
58 "cost_per_1m_in": 0.39,
59 "cost_per_1m_out": 1.9,
60 "cost_per_1m_in_cached": 0.195,
61 "cost_per_1m_out_cached": 0.78,
62 "context_window": 262144,
63 "default_max_tokens": 26214,
64 "can_reason": false,
65 "supports_attachments": false,
66 "options": {}
67 },
68 {
69 "id": "moonshotai/Kimi-K2-Thinking",
70 "name": "MoonshotAI: Kimi K2 Thinking",
71 "cost_per_1m_in": 0.32,
72 "cost_per_1m_out": 0.48,
73 "cost_per_1m_in_cached": 0.16,
74 "cost_per_1m_out_cached": 0.64,
75 "context_window": 262144,
76 "default_max_tokens": 26214,
77 "can_reason": true,
78 "reasoning_levels": [
79 "low",
80 "medium",
81 "high"
82 ],
83 "default_reasoning_effort": "medium",
84 "supports_attachments": false,
85 "options": {}
86 },
87 {
88 "id": "Qwen/Qwen3-Next-80B-A3B-Instruct",
89 "name": "Qwen: Qwen3 Next 80B A3B Instruct",
90 "cost_per_1m_in": 0.06,
91 "cost_per_1m_out": 0.6,
92 "cost_per_1m_in_cached": 0.03,
93 "cost_per_1m_out_cached": 0.12,
94 "context_window": 262144,
95 "default_max_tokens": 26214,
96 "can_reason": false,
97 "supports_attachments": false,
98 "options": {}
99 },
100 {
101 "id": "zai-org/GLM-4.6",
102 "name": "Z.ai: GLM 4.6",
103 "cost_per_1m_in": 0.35,
104 "cost_per_1m_out": 1.5,
105 "cost_per_1m_in_cached": 0.175,
106 "cost_per_1m_out_cached": 0.7,
107 "context_window": 200000,
108 "default_max_tokens": 20000,
109 "can_reason": true,
110 "reasoning_levels": [
111 "low",
112 "medium",
113 "high"
114 ],
115 "default_reasoning_effort": "medium",
116 "supports_attachments": false,
117 "options": {}
118 },
119 {
120 "id": "zai-org/GLM-4.7",
121 "name": "Z.ai: GLM 4.7",
122 "cost_per_1m_in": 0.3,
123 "cost_per_1m_out": 1.4,
124 "cost_per_1m_in_cached": 0.15,
125 "cost_per_1m_out_cached": 0.6,
126 "context_window": 202752,
127 "default_max_tokens": 20275,
128 "can_reason": true,
129 "reasoning_levels": [
130 "low",
131 "medium",
132 "high"
133 ],
134 "default_reasoning_effort": "medium",
135 "supports_attachments": false,
136 "options": {}
137 },
138 {
139 "id": "zai-org/GLM-4.7-Flash",
140 "name": "Z.ai: GLM 4.7 Flash",
141 "cost_per_1m_in": 0.07,
142 "cost_per_1m_out": 0.4,
143 "cost_per_1m_in_cached": 0.035,
144 "cost_per_1m_out_cached": 0.14,
145 "context_window": 200000,
146 "default_max_tokens": 20000,
147 "can_reason": true,
148 "reasoning_levels": [
149 "low",
150 "medium",
151 "high"
152 ],
153 "default_reasoning_effort": "medium",
154 "supports_attachments": false,
155 "options": {}
156 },
157 {
158 "id": "zai-org/GLM-5",
159 "name": "Z.ai: GLM 5",
160 "cost_per_1m_in": 0.94,
161 "cost_per_1m_out": 3,
162 "cost_per_1m_in_cached": 0.47,
163 "cost_per_1m_out_cached": 0,
164 "context_window": 202752,
165 "default_max_tokens": 20275,
166 "can_reason": true,
167 "reasoning_levels": [
168 "low",
169 "medium",
170 "high"
171 ],
172 "default_reasoning_effort": "medium",
173 "supports_attachments": false,
174 "options": {}
175 }
176 ]
177}