From 0d9e7e2ebd55bd0bcb22148dffbb042002c05795 Mon Sep 17 00:00:00 2001 From: Andrey Nering Date: Tue, 10 Mar 2026 11:39:59 -0300 Subject: [PATCH] chore: run generators --- internal/providers/configs/aihubmix.json | 18 +- internal/providers/configs/huggingface.json | 13 - internal/providers/configs/ionet.json | 27 +- internal/providers/configs/openrouter.json | 384 +++++++++++--------- internal/providers/configs/synthetic.json | 12 +- internal/providers/configs/vercel.json | 137 +++---- 6 files changed, 308 insertions(+), 283 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 484dd5c56c70c92ccb82455180eb25dbaa325c5c..24c372a2e67428243ab2d6135bcb166184e1ef9a 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -719,7 +719,7 @@ "id": "deepseek-v3.2-speciale", "name": "DeepSeek V3.2 Speciale", "cost_per_1m_in": 0.58, - "cost_per_1m_out": 1.680028, + "cost_per_1m_out": 1.68003, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 128000, @@ -809,9 +809,9 @@ "id": "doubao-seed-1-8", "name": "Doubao Seed 1.8", "cost_per_1m_in": 0.10959, - "cost_per_1m_out": 0.273975, + "cost_per_1m_out": 0.27398, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.021918, + "cost_per_1m_out_cached": 0.02192, "context_window": 256000, "default_max_tokens": 64000, "can_reason": true, @@ -849,7 +849,7 @@ "cost_per_1m_in": 0.09041, "cost_per_1m_out": 0.54246, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.018082, + "cost_per_1m_out_cached": 0.01808, "context_window": 256000, "default_max_tokens": 32000, "can_reason": true, @@ -865,10 +865,10 @@ { "id": "doubao-seed-2-0-mini", "name": "Doubao Seed 2.0 Mini", - "cost_per_1m_in": 0.030136, + "cost_per_1m_in": 0.03014, "cost_per_1m_out": 0.30136, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.006027, + "cost_per_1m_out_cached": 0.00603, "context_window": 256000, "default_max_tokens": 32000, "can_reason": true, @@ -1390,10 +1390,10 @@ { "id": "glm-4.7", "name": "GLM 4.7", - "cost_per_1m_in": 0.273974, - "cost_per_1m_out": 1.095896, + "cost_per_1m_in": 0.27397, + "cost_per_1m_out": 1.0959, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.054795, + "cost_per_1m_out_cached": 0.0548, "context_window": 200000, "default_max_tokens": 20000, "can_reason": true, diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 98519041158498e208d8f1a5964a445189a06181..8c9c824797c3b69e0a414ca8671721e65b8b93d9 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -137,19 +137,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct:groq", - "name": "meta-llama/Llama-4-Maverick-17B-128E-Instruct (groq)", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.6, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "meta-llama/Llama-4-Scout-17B-16E-Instruct:groq", "name": "meta-llama/Llama-4-Scout-17B-16E-Instruct (groq)", diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index 3e73945b1e9b95f5722827d0fde41188ff5833a0..8b5b5b314c97b48b8beb8cc0ae0c8ded802867ba 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -23,10 +23,10 @@ { "id": "meta-llama/Llama-3.3-70B-Instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.32, - "cost_per_1m_in_cached": 0.049999999999999996, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_in_cached": 0.05, + "cost_per_1m_out_cached": 0.2, "context_window": 128000, "default_max_tokens": 12800, "can_reason": true, @@ -139,7 +139,7 @@ "id": "zai-org/GLM-4.7-Flash", "name": "Z.ai: GLM 4.7 Flash", "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.035, "cost_per_1m_out_cached": 0.14, "context_window": 200000, @@ -153,6 +153,25 @@ "default_reasoning_effort": "medium", "supports_attachments": false, "options": {} + }, + { + "id": "zai-org/GLM-5", + "name": "Z.ai: GLM 5", + "cost_per_1m_in": 0.94, + "cost_per_1m_out": 3, + "cost_per_1m_in_cached": 0.47, + "cost_per_1m_out_cached": 0, + "context_window": 202752, + "default_max_tokens": 20275, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} } ] } diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 739d5b345b76c4ca2573879e72e0ce250fa36985..a0fa260fad21915ea1cc80f0b9c48b7c8e5a3e66 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -23,7 +23,7 @@ { "id": "allenai/olmo-3.1-32b-instruct", "name": "AllenAI: Olmo 3.1 32B Instruct", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -94,8 +94,8 @@ { "id": "amazon/nova-pro-v1", "name": "Amazon: Nova Pro 1.0", - "cost_per_1m_in": 0.7999999999999999, - "cost_per_1m_out": 3.1999999999999997, + "cost_per_1m_in": 0.8, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 300000, @@ -120,7 +120,7 @@ { "id": "anthropic/claude-3.5-haiku", "name": "Anthropic: Claude 3.5 Haiku", - "cost_per_1m_in": 0.7999999999999999, + "cost_per_1m_in": 0.8, "cost_per_1m_out": 4, "cost_per_1m_in_cached": 1, "cost_per_1m_out_cached": 0.08, @@ -151,7 +151,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 32000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -187,7 +187,7 @@ "cost_per_1m_in": 1, "cost_per_1m_out": 5, "cost_per_1m_in_cached": 1.25, - "cost_per_1m_out_cached": 0.09999999999999999, + "cost_per_1m_out_cached": 0.1, "context_window": 200000, "default_max_tokens": 32000, "can_reason": true, @@ -470,8 +470,8 @@ { "id": "bytedance-seed/seed-2.0-mini", "name": "ByteDance Seed: Seed-2.0-Mini", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -515,7 +515,7 @@ { "id": "deepseek/deepseek-chat", "name": "DeepSeek: DeepSeek V3", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 1.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -547,10 +547,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.19999999999999998, - "cost_per_1m_out": 0.7999999999999999, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.09999999999999999, + "cost_per_1m_out_cached": 0.1, "context_window": 163840, "default_max_tokens": 32768, "can_reason": true, @@ -566,12 +566,12 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.7899999999999999, + "cost_per_1m_in": 0.23, + "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1300000002, + "cost_per_1m_out_cached": 0.115, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -586,9 +586,9 @@ "id": "deepseek/deepseek-v3.1-terminus:exacto", "name": "DeepSeek: DeepSeek V3.1 Terminus (exacto)", "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.7899999999999999, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.16799999999999998, + "cost_per_1m_out_cached": 0.168, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -609,7 +609,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -661,10 +661,10 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.44999999999999996, - "cost_per_1m_out": 2.1500000000000004, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.22499999999999998, + "cost_per_1m_out_cached": 0.225, "context_window": 163840, "default_max_tokens": 32768, "can_reason": true, @@ -693,10 +693,10 @@ { "id": "google/gemini-2.0-flash-001", "name": "Google: Gemini 2.0 Flash", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, - "cost_per_1m_in_cached": 0.08333333333333334, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, + "cost_per_1m_in_cached": 0.08333, + "cost_per_1m_out_cached": 0.025, "context_window": 1048576, "default_max_tokens": 4096, "can_reason": false, @@ -721,7 +721,7 @@ "name": "Google: Gemini 2.5 Flash", "cost_per_1m_in": 0.3, "cost_per_1m_out": 2.5, - "cost_per_1m_in_cached": 0.08333333333333334, + "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.03, "context_window": 1048576, "default_max_tokens": 32767, @@ -738,9 +738,9 @@ { "id": "google/gemini-2.5-flash-lite", "name": "Google: Gemini 2.5 Flash Lite", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, - "cost_per_1m_in_cached": 0.08333333333333334, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, + "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.01, "context_window": 1048576, "default_max_tokens": 32767, @@ -757,9 +757,9 @@ { "id": "google/gemini-2.5-flash-lite-preview-09-2025", "name": "Google: Gemini 2.5 Flash Lite Preview 09-2025", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, - "cost_per_1m_in_cached": 0.08333333333333334, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, + "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.01, "context_window": 1048576, "default_max_tokens": 32768, @@ -835,10 +835,10 @@ "name": "Google: Gemini 3 Flash Preview", "cost_per_1m_in": 0.5, "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.08333333333333334, - "cost_per_1m_out_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.08333, + "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -855,7 +855,7 @@ "cost_per_1m_in": 2, "cost_per_1m_out": 12, "cost_per_1m_in_cached": 0.375, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 1048576, "default_max_tokens": 32768, "can_reason": true, @@ -873,8 +873,8 @@ "name": "Google: Gemini 3.1 Flash Lite Preview", "cost_per_1m_in": 0.25, "cost_per_1m_out": 1.5, - "cost_per_1m_in_cached": 0.08333333333333334, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_in_cached": 0.08333, + "cost_per_1m_out_cached": 0.025, "context_window": 1048576, "default_max_tokens": 32768, "can_reason": true, @@ -893,7 +893,7 @@ "cost_per_1m_in": 2, "cost_per_1m_out": 12, "cost_per_1m_in_cached": 0.375, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 1048576, "default_max_tokens": 32768, "can_reason": true, @@ -912,7 +912,7 @@ "cost_per_1m_in": 2, "cost_per_1m_out": 12, "cost_per_1m_in_cached": 0.375, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 1048576, "default_max_tokens": 32768, "can_reason": true, @@ -957,7 +957,7 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_out_cached": 0.025, "context_window": 128000, "default_max_tokens": 16000, "can_reason": false, @@ -970,7 +970,7 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_out_cached": 0.025, "context_window": 128000, "default_max_tokens": 25000, "can_reason": true, @@ -989,7 +989,7 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 0.75, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_out_cached": 0.025, "context_window": 128000, "default_max_tokens": 16000, "can_reason": false, @@ -1012,10 +1012,10 @@ { "id": "meituan/longcat-flash-chat", "name": "Meituan: LongCat Flash Chat", - "cost_per_1m_in": 0.19999999999999998, - "cost_per_1m_out": 0.7999999999999999, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 131072, "default_max_tokens": 65536, "can_reason": false, @@ -1038,12 +1038,12 @@ { "id": "meta-llama/llama-3.1-70b-instruct", "name": "Meta: Llama 3.1 70B Instruct", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 8192, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1100,6 +1100,25 @@ "supports_attachments": true, "options": {} }, + { + "id": "minimax/minimax-m1", + "name": "MiniMax: MiniMax M1", + "cost_per_1m_in": 0.44, + "cost_per_1m_out": 1.76, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 20000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", @@ -1146,7 +1165,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, "context_window": 204800, - "default_max_tokens": 65550, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -1200,7 +1219,7 @@ "id": "mistralai/codestral-2508", "name": "Mistral: Codestral 2508", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.8999999999999999, + "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 256000, @@ -1212,7 +1231,7 @@ { "id": "mistralai/devstral-2512", "name": "Mistral: Devstral 2 2512", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1225,7 +1244,7 @@ { "id": "mistralai/devstral-medium", "name": "Mistral: Devstral Medium", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1238,7 +1257,7 @@ { "id": "mistralai/devstral-small", "name": "Mistral: Devstral Small 1.1", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1251,8 +1270,8 @@ { "id": "mistralai/ministral-14b-2512", "name": "Mistral: Ministral 3 14B 2512", - "cost_per_1m_in": 0.19999999999999998, - "cost_per_1m_out": 0.19999999999999998, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -1264,8 +1283,8 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1303,7 +1322,7 @@ { "id": "mistralai/mistral-medium-3", "name": "Mistral: Mistral Medium 3", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1316,7 +1335,7 @@ { "id": "mistralai/mistral-medium-3.1", "name": "Mistral: Mistral Medium 3.1", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1355,12 +1374,12 @@ { "id": "mistralai/mistral-small-3.2-24b-instruct", "name": "Mistral: Mistral Small 3.2 24B", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.18, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": true, "options": {} @@ -1368,7 +1387,7 @@ { "id": "mistralai/mistral-small-creative", "name": "Mistral: Mistral Small Creative", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1420,7 +1439,7 @@ { "id": "mistralai/mistral-saba", "name": "Mistral: Saba", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1433,7 +1452,7 @@ { "id": "mistralai/voxtral-small-24b-2507", "name": "Mistral: Voxtral Small 24B 2507", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1446,12 +1465,12 @@ { "id": "moonshotai/kimi-k2", "name": "MoonshotAI: Kimi K2 0711", - "cost_per_1m_in": 0.5700000000000001, - "cost_per_1m_out": 2.3, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, "supports_attachments": false, "options": {} @@ -1472,10 +1491,10 @@ { "id": "moonshotai/kimi-k2-0905:exacto", "name": "MoonshotAI: Kimi K2 0905 (exacto)", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1.15, + "cost_per_1m_out": 8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, @@ -1488,7 +1507,7 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.15, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1504,10 +1523,10 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 2.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 26214, "can_reason": true, @@ -1536,8 +1555,8 @@ { "id": "nvidia/llama-3.3-nemotron-super-49b-v1.5", "name": "NVIDIA: Llama 3.3 Nemotron Super 49B V1.5", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1555,8 +1574,8 @@ { "id": "nvidia/nemotron-3-nano-30b-a3b", "name": "NVIDIA: Nemotron 3 Nano 30B A3B", - "cost_per_1m_in": 0.049999999999999996, - "cost_per_1m_out": 0.19999999999999998, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -1715,10 +1734,10 @@ { "id": "openai/gpt-4.1-mini", "name": "OpenAI: GPT-4.1 Mini", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 1.5999999999999999, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 1.6, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.09999999999999999, + "cost_per_1m_out_cached": 0.1, "context_window": 1047576, "default_max_tokens": 104757, "can_reason": false, @@ -1728,8 +1747,8 @@ { "id": "openai/gpt-4.1-nano", "name": "OpenAI: GPT-4.1 Nano", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, "context_window": 1047576, @@ -1940,8 +1959,8 @@ { "id": "openai/gpt-5-nano", "name": "OpenAI: GPT-5 Nano", - "cost_per_1m_in": 0.049999999999999996, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.01, "context_window": 400000, @@ -1981,7 +2000,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0.125, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2051,7 +2070,7 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.024999999999999998, + "cost_per_1m_out_cached": 0.025, "context_window": 400000, "default_max_tokens": 50000, "can_reason": true, @@ -2091,7 +2110,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.175, "context_window": 128000, - "default_max_tokens": 8192, + "default_max_tokens": 16000, "can_reason": false, "supports_attachments": true, "options": {} @@ -2207,12 +2226,12 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.49, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.45, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.025, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -2474,7 +2493,7 @@ { "id": "prime-intellect/intellect-3", "name": "Prime Intellect: INTELLECT-3", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -2538,8 +2557,8 @@ { "id": "qwen/qwen-vl-max", "name": "Qwen: Qwen VL Max", - "cost_per_1m_in": 0.7999999999999999, - "cost_per_1m_out": 3.1999999999999997, + "cost_per_1m_in": 0.8, + "cost_per_1m_out": 3.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2554,7 +2573,7 @@ "cost_per_1m_in": 1.04, "cost_per_1m_out": 4.16, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.20800000000000002, + "cost_per_1m_out_cached": 0.208, "context_window": 32768, "default_max_tokens": 4096, "can_reason": false, @@ -2564,7 +2583,7 @@ { "id": "qwen/qwen-plus", "name": "Qwen: Qwen-Plus", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.08, @@ -2580,7 +2599,7 @@ "cost_per_1m_in": 0.0325, "cost_per_1m_out": 0.13, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.006500000000000001, + "cost_per_1m_out_cached": 0.0065, "context_window": 131072, "default_max_tokens": 4096, "can_reason": false, @@ -2591,7 +2610,7 @@ "id": "qwen/qwen-2.5-7b-instruct", "name": "Qwen: Qwen2.5 7B Instruct", "cost_per_1m_in": 0.04, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.04, "context_window": 32768, @@ -2603,8 +2622,8 @@ { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.22749999999999998, - "cost_per_1m_out": 0.9099999999999999, + "cost_per_1m_in": 0.2275, + "cost_per_1m_out": 0.91, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2622,8 +2641,8 @@ { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", - "cost_per_1m_in": 0.45499999999999996, - "cost_per_1m_out": 1.8199999999999998, + "cost_per_1m_in": 0.455, + "cost_per_1m_out": 1.82, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2641,10 +2660,10 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.19999999999999998, - "cost_per_1m_out": 0.7999999999999999, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2692,12 +2711,12 @@ { "id": "qwen/qwen3-30b-a3b-instruct-2507", "name": "Qwen: Qwen3 30B A3B Instruct 2507", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false, "options": {} @@ -2705,8 +2724,8 @@ { "id": "qwen/qwen3-32b", "name": "Qwen: Qwen3 32B", - "cost_per_1m_in": 0.10400000000000001, - "cost_per_1m_out": 0.41600000000000004, + "cost_per_1m_in": 0.104, + "cost_per_1m_out": 0.416, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2744,7 +2763,7 @@ "id": "qwen/qwen3-8b", "name": "Qwen: Qwen3 8B", "cost_per_1m_in": 0.117, - "cost_per_1m_out": 0.45499999999999996, + "cost_per_1m_out": 0.455, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2762,12 +2781,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.28, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -2776,7 +2795,7 @@ "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", "cost_per_1m_in": 0.22, - "cost_per_1m_out": 1.7999999999999998, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2789,7 +2808,7 @@ "id": "qwen/qwen3-coder:exacto", "name": "Qwen: Qwen3 Coder 480B A35B (exacto)", "cost_per_1m_in": 0.22, - "cost_per_1m_out": 1.7999999999999998, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.022, "context_window": 262144, @@ -2930,8 +2949,8 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 1.9, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 1.04, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2975,12 +2994,12 @@ { "id": "qwen/qwen3-vl-30b-a3b-thinking", "name": "Qwen: Qwen3 VL 30B A3B Thinking", - "cost_per_1m_in": 0.29, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2994,8 +3013,8 @@ { "id": "qwen/qwen3-vl-32b-instruct", "name": "Qwen: Qwen3 VL 32B Instruct", - "cost_per_1m_in": 0.10400000000000001, - "cost_per_1m_out": 0.41600000000000004, + "cost_per_1m_in": 0.104, + "cost_per_1m_out": 0.416, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -3040,7 +3059,7 @@ "id": "qwen/qwen3.5-397b-a17b", "name": "Qwen: Qwen3.5 397B A17B", "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3.5999999999999996, + "cost_per_1m_out": 3.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -3096,10 +3115,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3131,11 +3150,30 @@ "supports_attachments": true, "options": {} }, + { + "id": "qwen/qwen3.5-9b", + "name": "Qwen: Qwen3.5-9B", + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.15, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 26214, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true, + "options": {} + }, { "id": "qwen/qwen3.5-flash-02-23", "name": "Qwen: Qwen3.5-Flash", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 1000000, @@ -3166,7 +3204,7 @@ { "id": "stepfun/step-3.5-flash", "name": "StepFun: Step 3.5 Flash", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.02, @@ -3223,7 +3261,7 @@ { "id": "thedrummer/rocinante-12b", "name": "TheDrummer: Rocinante 12B", - "cost_per_1m_in": 0.16999999999999998, + "cost_per_1m_in": 0.17, "cost_per_1m_out": 0.43, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -3236,8 +3274,8 @@ { "id": "thedrummer/unslopnemo-12b", "name": "TheDrummer: UnslopNemo 12B", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 32768, @@ -3250,7 +3288,7 @@ "id": "alibaba/tongyi-deepresearch-30b-a3b", "name": "Tongyi DeepResearch 30B A3B", "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.44999999999999996, + "cost_per_1m_out": 0.45, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.09, "context_window": 131072, @@ -3306,8 +3344,8 @@ { "id": "z-ai/glm-4-32b", "name": "Z.ai: GLM 4 32B ", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 128000, @@ -3377,7 +3415,7 @@ "id": "z-ai/glm-4.5v", "name": "Z.ai: GLM 4.5V", "cost_per_1m_in": 0.6, - "cost_per_1m_out": 1.7999999999999998, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 65536, @@ -3395,12 +3433,12 @@ { "id": "z-ai/glm-4.6", "name": "Z.ai: GLM 4.6", - "cost_per_1m_in": 0.39, - "cost_per_1m_out": 1.9, + "cost_per_1m_in": 0.55, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.11, "context_window": 204800, - "default_max_tokens": 102400, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3414,12 +3452,12 @@ { "id": "z-ai/glm-4.6:exacto", "name": "Z.ai: GLM 4.6 (exacto)", - "cost_per_1m_in": 0.44, - "cost_per_1m_out": 1.76, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, - "context_window": 204800, - "default_max_tokens": 65536, + "context_window": 200000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -3434,7 +3472,7 @@ "id": "z-ai/glm-4.6v", "name": "Z.ai: GLM 4.6V", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.8999999999999999, + "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -3453,7 +3491,7 @@ "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", "cost_per_1m_in": 0.54, - "cost_per_1m_out": 1.9800000000000002, + "cost_per_1m_out": 1.98, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.099, "context_window": 204800, @@ -3471,7 +3509,7 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.43, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -3491,9 +3529,9 @@ "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", "cost_per_1m_in": 0.95, - "cost_per_1m_out": 2.5500000000000003, + "cost_per_1m_out": 2.55, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.19999999999999998, + "cost_per_1m_out_cached": 0.2, "context_window": 204800, "default_max_tokens": 65536, "can_reason": true, @@ -3509,10 +3547,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3522,10 +3560,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3535,10 +3573,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3554,10 +3592,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3592,10 +3630,10 @@ { "id": "x-ai/grok-4-fast", "name": "xAI: Grok 4 Fast", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.049999999999999996, + "cost_per_1m_out_cached": 0.05, "context_window": 2000000, "default_max_tokens": 15000, "can_reason": true, @@ -3611,10 +3649,10 @@ { "id": "x-ai/grok-4.1-fast", "name": "xAI: Grok 4.1 Fast", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.049999999999999996, + "cost_per_1m_out_cached": 0.05, "context_window": 2000000, "default_max_tokens": 15000, "can_reason": true, @@ -3630,7 +3668,7 @@ { "id": "x-ai/grok-code-fast-1", "name": "xAI: Grok Code Fast 1", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.02, diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index f5917d65e81b66c7ee81523832966e534c42a721..4b893ff14aa6b9de45e43be4779697cee1e8f9f3 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -62,7 +62,7 @@ "id": "hf:zai-org/GLM-4.7", "name": "GLM 4.7", "cost_per_1m_in": 0.55, - "cost_per_1m_out": 2.1900000000000004, + "cost_per_1m_out": 2.19, "cost_per_1m_in_cached": 0.55, "cost_per_1m_out_cached": 0.55, "context_window": 202752, @@ -81,7 +81,7 @@ "id": "hf:zai-org/GLM-4.7-Flash", "name": "GLM 4.7 Flash", "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.06, "cost_per_1m_out_cached": 0.06, "context_window": 196608, @@ -265,10 +265,10 @@ { "id": "hf:openai/gpt-oss-120b", "name": "gpt oss 120b", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, - "cost_per_1m_in_cached": 0.09999999999999999, - "cost_per_1m_out_cached": 0.09999999999999999, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, + "cost_per_1m_in_cached": 0.1, + "cost_per_1m_out_cached": 0.1, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index 7155d7371ffa4a2f7fa4ac4c0f1b2f110e580bc4..43b3b5facce3a868e8b39bcea2f05d7044f0e8a9 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -23,7 +23,7 @@ { "id": "anthropic/claude-3.5-haiku", "name": "Claude 3.5 Haiku", - "cost_per_1m_in": 0.7999999999999999, + "cost_per_1m_in": 0.8, "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0.08, "cost_per_1m_out_cached": 1, @@ -86,7 +86,7 @@ "name": "Claude Haiku 4.5", "cost_per_1m_in": 1, "cost_per_1m_out": 5, - "cost_per_1m_in_cached": 0.09999999999999999, + "cost_per_1m_in_cached": 0.1, "cost_per_1m_out_cached": 1.25, "context_window": 200000, "default_max_tokens": 8000, @@ -257,25 +257,6 @@ "supports_attachments": true, "options": {} }, - { - "id": "openai/codex-mini", - "name": "Codex Mini", - "cost_per_1m_in": 1.5, - "cost_per_1m_out": 6, - "cost_per_1m_in_cached": 0.375, - "cost_per_1m_out_cached": 0, - "context_window": 200000, - "default_max_tokens": 8000, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} - }, { "id": "cohere/command-a", "name": "Command A", @@ -357,7 +338,7 @@ "id": "deepseek/deepseek-v3.1", "name": "DeepSeek-V3.1", "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.7899999999999999, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 163840, @@ -388,7 +369,7 @@ { "id": "mistral/devstral-small", "name": "Devstral Small 1.1", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -414,7 +395,7 @@ { "id": "zai/glm-4.5-air", "name": "GLM 4.5 Air", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0.03, "cost_per_1m_out_cached": 0, @@ -434,7 +415,7 @@ "id": "zai/glm-4.5v", "name": "GLM 4.5V", "cost_per_1m_in": 0.6, - "cost_per_1m_out": 1.7999999999999998, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 65536, @@ -452,8 +433,8 @@ { "id": "zai/glm-4.6", "name": "GLM 4.6", - "cost_per_1m_in": 0.44999999999999996, - "cost_per_1m_out": 1.7999999999999998, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0.11, "cost_per_1m_out_cached": 0, "context_window": 200000, @@ -491,7 +472,7 @@ "id": "zai/glm-4.7-flash", "name": "GLM 4.7 Flash", "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 200000, @@ -510,7 +491,7 @@ "id": "zai/glm-4.7-flashx", "name": "GLM 4.7 FlashX", "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.01, "cost_per_1m_out_cached": 0, "context_window": 200000, @@ -528,7 +509,7 @@ { "id": "zai/glm-5", "name": "GLM 5", - "cost_per_1m_in": 0.7999999999999999, + "cost_per_1m_in": 0.8, "cost_per_1m_out": 2.56, "cost_per_1m_in_cached": 0.16, "cost_per_1m_out_cached": 0, @@ -567,8 +548,8 @@ "id": "zai/glm-4.6v", "name": "GLM-4.6V", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.8999999999999999, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_out": 0.9, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8000, @@ -644,7 +625,7 @@ "name": "GPT 5.1 Codex Mini", "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, - "cost_per_1m_in_cached": 0.024999999999999998, + "cost_per_1m_in_cached": 0.025, "cost_per_1m_out_cached": 0, "context_window": 400000, "default_max_tokens": 8000, @@ -839,9 +820,9 @@ { "id": "openai/gpt-4.1-mini", "name": "GPT-4.1 mini", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 1.5999999999999999, - "cost_per_1m_in_cached": 0.09999999999999999, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 1.6, + "cost_per_1m_in_cached": 0.1, "cost_per_1m_out_cached": 0, "context_window": 1047576, "default_max_tokens": 8000, @@ -852,8 +833,8 @@ { "id": "openai/gpt-4.1-nano", "name": "GPT-4.1 nano", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.03, "cost_per_1m_out_cached": 0, "context_window": 1047576, @@ -929,8 +910,8 @@ { "id": "openai/gpt-5-nano", "name": "GPT-5 nano", - "cost_per_1m_in": 0.049999999999999996, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.01, "cost_per_1m_out_cached": 0, "context_window": 400000, @@ -1062,8 +1043,8 @@ { "id": "google/gemini-2.5-flash-lite", "name": "Gemini 2.5 Flash Lite", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.01, "cost_per_1m_out_cached": 0, "context_window": 1048576, @@ -1081,8 +1062,8 @@ { "id": "google/gemini-2.5-flash-lite-preview-09-2025", "name": "Gemini 2.5 Flash Lite Preview 09-2025", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.01, "cost_per_1m_out_cached": 0, "context_window": 1048576, @@ -1140,7 +1121,7 @@ "name": "Gemini 3 Flash", "cost_per_1m_in": 0.5, "cost_per_1m_out": 3, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 1000000, "default_max_tokens": 8000, @@ -1159,7 +1140,7 @@ "name": "Gemini 3 Pro Preview", "cost_per_1m_in": 2, "cost_per_1m_out": 12, - "cost_per_1m_in_cached": 0.19999999999999998, + "cost_per_1m_in_cached": 0.2, "cost_per_1m_out_cached": 0, "context_window": 1000000, "default_max_tokens": 8000, @@ -1197,7 +1178,7 @@ "name": "Gemini 3.1 Pro Preview", "cost_per_1m_in": 2, "cost_per_1m_out": 12, - "cost_per_1m_in_cached": 0.19999999999999998, + "cost_per_1m_in_cached": 0.2, "cost_per_1m_out_cached": 0, "context_window": 1000000, "default_max_tokens": 8000, @@ -1298,9 +1279,9 @@ { "id": "xai/grok-4-fast-non-reasoning", "name": "Grok 4 Fast Non-Reasoning", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 2000000, "default_max_tokens": 8000, @@ -1311,9 +1292,9 @@ { "id": "xai/grok-4-fast-reasoning", "name": "Grok 4 Fast Reasoning", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 2000000, "default_max_tokens": 8000, @@ -1330,9 +1311,9 @@ { "id": "xai/grok-4.1-fast-non-reasoning", "name": "Grok 4.1 Fast Non-Reasoning", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 2000000, "default_max_tokens": 8000, @@ -1343,9 +1324,9 @@ { "id": "xai/grok-4.1-fast-reasoning", "name": "Grok 4.1 Fast Reasoning", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.5, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 2000000, "default_max_tokens": 8000, @@ -1362,7 +1343,7 @@ { "id": "xai/grok-code-fast-1", "name": "Grok Code Fast 1", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0.02, "cost_per_1m_out_cached": 0, @@ -1381,7 +1362,7 @@ { "id": "prime-intellect/intellect-3", "name": "INTELLECT 3", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1415,7 +1396,7 @@ "name": "Kimi K2 Thinking", "cost_per_1m_in": 0.47, "cost_per_1m_out": 2, - "cost_per_1m_in_cached": 0.14100000000000001, + "cost_per_1m_in_cached": 0.141, "cost_per_1m_out_cached": 0, "context_window": 216144, "default_max_tokens": 8000, @@ -1483,8 +1464,8 @@ { "id": "meta/llama-3.1-70b", "name": "Llama 3.1 70B Instruct", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1497,7 +1478,7 @@ "id": "meta/llama-3.1-8b", "name": "Llama 3.1 8B Instruct", "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.049999999999999996, + "cost_per_1m_out": 0.05, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1608,7 +1589,7 @@ "name": "Mercury 2", "cost_per_1m_in": 0.25, "cost_per_1m_out": 0.75, - "cost_per_1m_in_cached": 0.024999999999999998, + "cost_per_1m_in_cached": 0.025, "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8000, @@ -1765,8 +1746,8 @@ { "id": "mistral/ministral-8b", "name": "Ministral 8B", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 128000, @@ -1779,7 +1760,7 @@ "id": "mistral/codestral", "name": "Mistral Codestral", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.8999999999999999, + "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 128000, @@ -1791,7 +1772,7 @@ { "id": "mistral/mistral-medium", "name": "Mistral Medium 3.1", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1804,7 +1785,7 @@ { "id": "mistral/mistral-small", "name": "Mistral Small", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1817,7 +1798,7 @@ { "id": "nvidia/nemotron-nano-12b-v2-vl", "name": "Nvidia Nemotron Nano 12B V2 VL", - "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_in": 0.2, "cost_per_1m_out": 0.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1881,7 +1862,7 @@ { "id": "alibaba/qwen-3-32b", "name": "Qwen 3 32B", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, @@ -1938,8 +1919,8 @@ { "id": "alibaba/qwen3.5-flash", "name": "Qwen 3.5 Flash", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0.001, "cost_per_1m_out_cached": 0.125, "context_window": 1000000, @@ -1957,7 +1938,7 @@ { "id": "alibaba/qwen3.5-plus", "name": "Qwen 3.5 Plus", - "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_in": 0.4, "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0.04, "cost_per_1m_out_cached": 0.5, @@ -1977,7 +1958,7 @@ "id": "alibaba/qwen3-235b-a22b-thinking", "name": "Qwen3 235B A22B Thinking 2507", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 2.9000000000000004, + "cost_per_1m_out": 2.9, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262114, @@ -1995,8 +1976,8 @@ { "id": "alibaba/qwen3-coder", "name": "Qwen3 Coder 480B A35B Instruct", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 1.5999999999999999, + "cost_per_1m_in": 0.4, + "cost_per_1m_out": 1.6, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, @@ -2023,7 +2004,7 @@ "name": "Qwen3 Coder Plus", "cost_per_1m_in": 1, "cost_per_1m_out": 5, - "cost_per_1m_in_cached": 0.19999999999999998, + "cost_per_1m_in_cached": 0.2, "cost_per_1m_out_cached": 0, "context_window": 1000000, "default_max_tokens": 8000, @@ -2119,7 +2100,7 @@ "name": "Seed 1.6", "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, - "cost_per_1m_in_cached": 0.049999999999999996, + "cost_per_1m_in_cached": 0.05, "cost_per_1m_out_cached": 0, "context_window": 256000, "default_max_tokens": 8000, @@ -2175,7 +2156,7 @@ { "id": "openai/gpt-oss-120b", "name": "gpt-oss-120b", - "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0,