diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 3e971b02a3c398f8ef43566a5a4559ed5dd2974c..79e39c69379381335142a956587eff67c2c8a42c 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -3115,13 +3115,31 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "qwen3.6-27b", + "name": "Qwen3.6 27B", + "cost_per_1m_in": 0.422, + "cost_per_1m_out": 2.532, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 254000, + "default_max_tokens": 64000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "qwen3.6-35b-a3b", "name": "Qwen3.6 35B A3B", "cost_per_1m_in": 0.254, "cost_per_1m_out": 1.524, - "cost_per_1m_in_cached": 0.3175, - "cost_per_1m_out_cached": 0.0254, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.254, "context_window": 254000, "default_max_tokens": 64000, "can_reason": true, @@ -3246,6 +3264,30 @@ "default_max_tokens": 25600, "can_reason": false, "supports_attachments": true + }, + { + "id": "xiaomi-mimo-v2.5", + "name": "Xiaomi Mimo V2.5", + "cost_per_1m_in": 0.44, + "cost_per_1m_out": 2.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.088, + "context_window": 256000, + "default_max_tokens": 25600, + "can_reason": false, + "supports_attachments": true + }, + { + "id": "xiaomi-mimo-v2.5-pro", + "name": "Xiaomi Mimo V2.5 Pro", + "cost_per_1m_in": 1.1, + "cost_per_1m_out": 3.3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.22, + "context_window": 1000000, + "default_max_tokens": 100000, + "can_reason": false, + "supports_attachments": false } ], "default_headers": { diff --git a/internal/providers/configs/cortecs.json b/internal/providers/configs/cortecs.json index d2b03ee29ca5a72f4db35e64c64bff464ea94598..da477264a60fa531910d7d7d234b41dad86db877 100644 --- a/internal/providers/configs/cortecs.json +++ b/internal/providers/configs/cortecs.json @@ -7,6 +7,24 @@ "default_large_model_id": "qwen3-coder-30b-a3b-instruct", "default_small_model_id": "glm-4.7-flash", "models": [ + { + "id": "kimi-k2.6", + "name": "Kimi K2.6", + "cost_per_1m_in": 0.6936, + "cost_per_1m_out": 3.0345, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 25600, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "claude-opus4-7", "name": "Claude Opus 4.7", @@ -82,12 +100,12 @@ { "id": "qwen3.5-9b", "name": "Qwen3.5 9B", - "cost_per_1m_in": 0.13311, - "cost_per_1m_out": 0.17748, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "context_window": 262000, + "default_max_tokens": 26200, "can_reason": true, "reasoning_levels": [ "low", diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index e9629ca154aec5c892150c4d470f0edd18da166b..623d3280957d340ee1cfcd0cbb16df3a6b2833b8 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -91,6 +91,18 @@ "can_reason": false, "supports_attachments": false }, + { + "id": "meta-llama/Llama-3.3-70B-Instruct:groq", + "name": "meta-llama/Llama-3.3-70B-Instruct (groq)", + "cost_per_1m_in": 0.59, + "cost_per_1m_out": 0.79, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 8192, + "can_reason": false, + "supports_attachments": false + }, { "id": "meta-llama/Llama-4-Scout-17B-16E-Instruct:groq", "name": "meta-llama/Llama-4-Scout-17B-16E-Instruct (groq)", diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index ba537e0934311ada3f4ee6eb66dff9dec4935be8..7e3761316313945b54969c649fbb5efe69ecc236 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -108,7 +108,7 @@ "name": "MoonshotAI: Kimi K2.6", "cost_per_1m_in": 0.8, "cost_per_1m_out": 4, - "cost_per_1m_in_cached": 0.2, + "cost_per_1m_in_cached": 0.25, "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 26214, diff --git a/internal/providers/configs/opencode-zen.json b/internal/providers/configs/opencode-zen.json index 21bc23ead58d7ee48a1480c7b11a7d35be61941b..9eb7d975df9fd9b564f8d592f778fdd0b338f4c9 100644 --- a/internal/providers/configs/opencode-zen.json +++ b/internal/providers/configs/opencode-zen.json @@ -541,6 +541,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "hy3-preview-free", + "name": "Hy3 preview Free", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 64000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "kimi-k2.5", "name": "Kimi K2.5", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index a87333e5ea7ea8dcb9202c47507c76c6d19442ff..7eb41eac11082525abe537a7cbe5c288a1e00dc4 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -129,7 +129,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 64000, + "default_max_tokens": 32000, "can_reason": true, "reasoning_levels": [ "low", @@ -586,12 +586,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.5, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.25, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -607,9 +607,9 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.27, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 81920, "can_reason": true, "reasoning_levels": [ "low", @@ -807,7 +807,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32768, + "default_max_tokens": 32767, "can_reason": true, "reasoning_levels": [ "low", @@ -874,10 +874,10 @@ { "id": "google/gemma-4-26b-a4b-it", "name": "Google: Gemma 4 26B A4B ", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.09, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.05, + "cost_per_1m_out_cached": 0.06, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -988,12 +988,12 @@ { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.59, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.6, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false }, @@ -1210,12 +1210,12 @@ { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": false, "supports_attachments": true }, @@ -1444,10 +1444,10 @@ { "id": "moonshotai/kimi-k2.6", "name": "MoonshotAI: Kimi K2.6", - "cost_per_1m_in": 0.8, + "cost_per_1m_in": 0.95, "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.2, + "cost_per_1m_out_cached": 0.16, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1849,9 +1849,9 @@ "cost_per_1m_in": 0.25, "cost_per_1m_out": 2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0.03, "context_window": 400000, - "default_max_tokens": 64000, + "default_max_tokens": 40000, "can_reason": true, "reasoning_levels": [ "low", @@ -1867,9 +1867,9 @@ "cost_per_1m_in": 0.05, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0.005, "context_window": 400000, - "default_max_tokens": 40000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -2188,12 +2188,12 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.14, + "cost_per_1m_in": 0.045, + "cost_per_1m_out": 0.18, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2578,10 +2578,10 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.1, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2638,12 +2638,12 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 1.56, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.08, - "context_window": 131072, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0, + "context_window": 81920, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -2692,12 +2692,12 @@ { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.27, + "cost_per_1m_in": 0.2925, + "cost_per_1m_out": 1.4625, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 160000, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false }, @@ -2743,7 +2743,7 @@ "cost_per_1m_in": 0.15, "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0.1, "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, @@ -2818,8 +2818,8 @@ { "id": "qwen/qwen3-next-80b-a3b-thinking", "name": "Qwen: Qwen3 Next 80B A3B Thinking", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.0975, + "cost_per_1m_out": 0.78, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2974,10 +2974,10 @@ { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 2.08, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3010,12 +3010,12 @@ { "id": "qwen/qwen3.5-35b-a3b", "name": "Qwen: Qwen3.5-35B-A3B", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.225, + "cost_per_1m_out": 1.8, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.225, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3127,6 +3127,24 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "tencent/hy3-preview:free", + "name": "Tencent: Hy3 preview (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 131072, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "thedrummer/rocinante-12b", "name": "TheDrummer: Rocinante 12B", @@ -3190,12 +3208,12 @@ { "id": "xiaomi/mimo-v2-flash", "name": "Xiaomi: MiMo-V2-Flash", - "cost_per_1m_in": 0.1, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.29, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.02, + "cost_per_1m_out_cached": 0.045, "context_window": 262144, - "default_max_tokens": 16000, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3297,7 +3315,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 48000, + "default_max_tokens": 49152, "can_reason": true, "reasoning_levels": [ "low", @@ -3385,9 +3403,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 12000, "can_reason": true, "reasoning_levels": [ "low", @@ -3400,12 +3418,12 @@ { "id": "z-ai/glm-4.7", "name": "Z.ai: GLM 4.7", - "cost_per_1m_in": 0.38, - "cost_per_1m_out": 1.74, + "cost_per_1m_in": 0.45, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 202752, - "default_max_tokens": 20275, + "cost_per_1m_out_cached": 0.11, + "context_window": 204800, + "default_max_tokens": 102400, "can_reason": true, "reasoning_levels": [ "low", @@ -3505,6 +3523,18 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "inclusionai/ling-2.6-1t:free", + "name": "inclusionAI: Ling-2.6-1T (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 16384, + "can_reason": false, + "supports_attachments": false + }, { "id": "inclusionai/ling-2.6-flash:free", "name": "inclusionAI: Ling-2.6-flash (free)", @@ -3520,10 +3550,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3532,10 +3562,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 5, - "cost_per_1m_out": 25, + "cost_per_1m_in": 3, + "cost_per_1m_out": 15, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0.75, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -3544,10 +3574,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.5, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.075, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3562,10 +3592,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.5, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.075, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index c74301aad382f4db23648d71cf5cd8130313d0dd..8d39ef99aba1c3c88414b0d86312c574c35f7f14 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -520,8 +520,8 @@ { "id": "kimi-k2-6", "name": "Kimi K2.6", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 3.5, + "cost_per_1m_in": 0.7448, + "cost_per_1m_out": 4.655, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 256000,