ionet.json

  1{
  2  "name": "io.net",
  3  "id": "ionet",
  4  "api_key": "$IONET_API_KEY",
  5  "api_endpoint": "https://api.intelligence.io.solutions/api/v1",
  6  "type": "openai-compat",
  7  "default_large_model_id": "zai-org/GLM-4.7",
  8  "default_small_model_id": "zai-org/GLM-4.7-Flash",
  9  "models": [
 10    {
 11      "id": "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar",
 12      "name": "Intel: Qwen3 Coder 480B A35B Instruct INT4 Mixed AR",
 13      "cost_per_1m_in": 0.22,
 14      "cost_per_1m_out": 0.95,
 15      "cost_per_1m_in_cached": 0.11,
 16      "cost_per_1m_out_cached": 0.44,
 17      "context_window": 106000,
 18      "default_max_tokens": 10600,
 19      "can_reason": false,
 20      "supports_attachments": false,
 21      "options": {}
 22    },
 23    {
 24      "id": "meta-llama/Llama-3.3-70B-Instruct",
 25      "name": "Meta: Llama 3.3 70B Instruct",
 26      "cost_per_1m_in": 0.1,
 27      "cost_per_1m_out": 0.32,
 28      "cost_per_1m_in_cached": 0.05,
 29      "cost_per_1m_out_cached": 0.2,
 30      "context_window": 128000,
 31      "default_max_tokens": 12800,
 32      "can_reason": true,
 33      "reasoning_levels": [
 34        "low",
 35        "medium",
 36        "high"
 37      ],
 38      "default_reasoning_effort": "medium",
 39      "supports_attachments": false,
 40      "options": {}
 41    },
 42    {
 43      "id": "mistralai/Mistral-Large-Instruct-2411",
 44      "name": "Mistral: Mistral Large Instruct 2411",
 45      "cost_per_1m_in": 2,
 46      "cost_per_1m_out": 6,
 47      "cost_per_1m_in_cached": 1,
 48      "cost_per_1m_out_cached": 4,
 49      "context_window": 128000,
 50      "default_max_tokens": 12800,
 51      "can_reason": false,
 52      "supports_attachments": true,
 53      "options": {}
 54    },
 55    {
 56      "id": "moonshotai/Kimi-K2-Instruct-0905",
 57      "name": "MoonshotAI: Kimi K2 Instruct 0905",
 58      "cost_per_1m_in": 0.39,
 59      "cost_per_1m_out": 1.9,
 60      "cost_per_1m_in_cached": 0.195,
 61      "cost_per_1m_out_cached": 0.78,
 62      "context_window": 262144,
 63      "default_max_tokens": 26214,
 64      "can_reason": false,
 65      "supports_attachments": false,
 66      "options": {}
 67    },
 68    {
 69      "id": "moonshotai/Kimi-K2-Thinking",
 70      "name": "MoonshotAI: Kimi K2 Thinking",
 71      "cost_per_1m_in": 0.32,
 72      "cost_per_1m_out": 0.48,
 73      "cost_per_1m_in_cached": 0.16,
 74      "cost_per_1m_out_cached": 0.64,
 75      "context_window": 262144,
 76      "default_max_tokens": 26214,
 77      "can_reason": true,
 78      "reasoning_levels": [
 79        "low",
 80        "medium",
 81        "high"
 82      ],
 83      "default_reasoning_effort": "medium",
 84      "supports_attachments": false,
 85      "options": {}
 86    },
 87    {
 88      "id": "Qwen/Qwen3-Next-80B-A3B-Instruct",
 89      "name": "Qwen: Qwen3 Next 80B A3B Instruct",
 90      "cost_per_1m_in": 0.06,
 91      "cost_per_1m_out": 0.6,
 92      "cost_per_1m_in_cached": 0.03,
 93      "cost_per_1m_out_cached": 0.12,
 94      "context_window": 262144,
 95      "default_max_tokens": 26214,
 96      "can_reason": false,
 97      "supports_attachments": false,
 98      "options": {}
 99    },
100    {
101      "id": "zai-org/GLM-4.6",
102      "name": "Z.ai: GLM 4.6",
103      "cost_per_1m_in": 0.35,
104      "cost_per_1m_out": 1.5,
105      "cost_per_1m_in_cached": 0.175,
106      "cost_per_1m_out_cached": 0.7,
107      "context_window": 200000,
108      "default_max_tokens": 20000,
109      "can_reason": true,
110      "reasoning_levels": [
111        "low",
112        "medium",
113        "high"
114      ],
115      "default_reasoning_effort": "medium",
116      "supports_attachments": false,
117      "options": {}
118    },
119    {
120      "id": "zai-org/GLM-4.7",
121      "name": "Z.ai: GLM 4.7",
122      "cost_per_1m_in": 0.3,
123      "cost_per_1m_out": 1.4,
124      "cost_per_1m_in_cached": 0.15,
125      "cost_per_1m_out_cached": 0.6,
126      "context_window": 202752,
127      "default_max_tokens": 20275,
128      "can_reason": true,
129      "reasoning_levels": [
130        "low",
131        "medium",
132        "high"
133      ],
134      "default_reasoning_effort": "medium",
135      "supports_attachments": false,
136      "options": {}
137    },
138    {
139      "id": "zai-org/GLM-4.7-Flash",
140      "name": "Z.ai: GLM 4.7 Flash",
141      "cost_per_1m_in": 0.07,
142      "cost_per_1m_out": 0.4,
143      "cost_per_1m_in_cached": 0.035,
144      "cost_per_1m_out_cached": 0.14,
145      "context_window": 200000,
146      "default_max_tokens": 20000,
147      "can_reason": true,
148      "reasoning_levels": [
149        "low",
150        "medium",
151        "high"
152      ],
153      "default_reasoning_effort": "medium",
154      "supports_attachments": false,
155      "options": {}
156    },
157    {
158      "id": "zai-org/GLM-5",
159      "name": "Z.ai: GLM 5",
160      "cost_per_1m_in": 0.94,
161      "cost_per_1m_out": 3,
162      "cost_per_1m_in_cached": 0.47,
163      "cost_per_1m_out_cached": 0,
164      "context_window": 202752,
165      "default_max_tokens": 20275,
166      "can_reason": true,
167      "reasoning_levels": [
168        "low",
169        "medium",
170        "high"
171      ],
172      "default_reasoning_effort": "medium",
173      "supports_attachments": false,
174      "options": {}
175    }
176  ]
177}