diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 79e39c69379381335142a956587eff67c2c8a42c..384bf12b2ecbb49115dc47a6d714faa2b6f1cc95 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -817,6 +817,60 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "deepseek-v4-flash", + "name": "DeepSeek V4 Flash", + "cost_per_1m_in": 0.154, + "cost_per_1m_out": 0.308, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.0308, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek-v4-flash-think", + "name": "DeepSeek V4 Flash Thinking", + "cost_per_1m_in": 0.154, + "cost_per_1m_out": 0.308, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.0308, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek-v4-pro", + "name": "DeepSeek V4 Pro", + "cost_per_1m_in": 1.914, + "cost_per_1m_out": 3.828, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.17226, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "doubao-seed-1-6", "name": "Doubao Seed 1.6", diff --git a/internal/providers/configs/avian.json b/internal/providers/configs/avian.json index 6bfc405fde0882c2f3c837372f46679d4f96e288..ebd98e2d3654c6a3bc2de88866620905578fc27a 100644 --- a/internal/providers/configs/avian.json +++ b/internal/providers/configs/avian.json @@ -25,6 +25,42 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "deepseek/deepseek-v4-flash", + "name": "DeepSeek V4 Flash", + "cost_per_1m_in": 0.105, + "cost_per_1m_out": 0.21, + "cost_per_1m_in_cached": 0.021, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 393216, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek/deepseek-v4-pro", + "name": "DeepSeek V4 Pro", + "cost_per_1m_in": 1.305, + "cost_per_1m_out": 2.61, + "cost_per_1m_in_cached": 0.10875, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 393216, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "z-ai/glm-5", "name": "GLM-5", diff --git a/internal/providers/configs/chutes.json b/internal/providers/configs/chutes.json index 816247ee47ea2f186a93607e881b5584762a60a2..7236f2143abcba0bd93de044377c67b067a6cfef 100644 --- a/internal/providers/configs/chutes.json +++ b/internal/providers/configs/chutes.json @@ -433,6 +433,24 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "Qwen/Qwen3.6-27B-TEE", + "name": "Qwen3.6-27B-TEE", + "cost_per_1m_in": 0.195, + "cost_per_1m_out": 1.56, + "cost_per_1m_in_cached": 0.0975, + "cost_per_1m_out_cached": 0, + "context_window": 262144, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "unsloth/gemma-3-27b-it", "name": "gemma-3-27b-it", @@ -445,6 +463,24 @@ "can_reason": false, "supports_attachments": true }, + { + "id": "google/gemma-4-31B-turbo-TEE", + "name": "gemma-4-31B-turbo-TEE", + "cost_per_1m_in": 0.13, + "cost_per_1m_out": 0.38, + "cost_per_1m_in_cached": 0.065, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 65536, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "openai/gpt-oss-120b-TEE", "name": "gpt-oss-120b-TEE", diff --git a/internal/providers/configs/cortecs.json b/internal/providers/configs/cortecs.json index da477264a60fa531910d7d7d234b41dad86db877..390dbd4687ec70d6e73112bfe102a5aa7589dee0 100644 --- a/internal/providers/configs/cortecs.json +++ b/internal/providers/configs/cortecs.json @@ -7,6 +7,24 @@ "default_large_model_id": "qwen3-coder-30b-a3b-instruct", "default_small_model_id": "glm-4.7-flash", "models": [ + { + "id": "gpt-5.4", + "name": "GPT 5.4", + "cost_per_1m_in": 2.601, + "cost_per_1m_out": 13.872, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1050000, + "default_max_tokens": 105000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "kimi-k2.6", "name": "Kimi K2.6", diff --git a/internal/providers/configs/opencode-go.json b/internal/providers/configs/opencode-go.json index 3253827804cab95a238e848cda89558d2d0dc956..320ce23ef9116aad80574b542825ffdd007acab0 100644 --- a/internal/providers/configs/opencode-go.json +++ b/internal/providers/configs/opencode-go.json @@ -7,6 +7,42 @@ "default_large_model_id": "minimax-m2.7", "default_small_model_id": "minimax-m2.7", "models": [ + { + "id": "deepseek-v4-flash", + "name": "DeepSeek V4 Flash", + "cost_per_1m_in": 0.14, + "cost_per_1m_out": 0.28, + "cost_per_1m_in_cached": 0.03, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek-v4-pro", + "name": "DeepSeek V4 Pro", + "cost_per_1m_in": 1.74, + "cost_per_1m_out": 3.48, + "cost_per_1m_in_cached": 0.14, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 384000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "glm-5", "name": "GLM-5", diff --git a/internal/providers/configs/opencode-zen.json b/internal/providers/configs/opencode-zen.json index 9eb7d975df9fd9b564f8d592f778fdd0b338f4c9..51570294a17c2f0cca08626ac701a74d39eddb8f 100644 --- a/internal/providers/configs/opencode-zen.json +++ b/internal/providers/configs/opencode-zen.json @@ -487,6 +487,42 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "gpt-5.5", + "name": "GPT-5.5", + "cost_per_1m_in": 5, + "cost_per_1m_out": 30, + "cost_per_1m_in_cached": 0.5, + "cost_per_1m_out_cached": 0, + "context_window": 1050000, + "default_max_tokens": 130000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, + { + "id": "gpt-5.5-pro", + "name": "GPT-5.5 Pro", + "cost_per_1m_in": 30, + "cost_per_1m_out": 180, + "cost_per_1m_in_cached": 30, + "cost_per_1m_out_cached": 0, + "context_window": 1050000, + "default_max_tokens": 128000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "gemini-3-flash", "name": "Gemini 3 Flash", diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 7eb41eac11082525abe537a7cbe5c288a1e00dc4..fd9ac55715c51690eb2c39e72aa83396c281e183 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -129,7 +129,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 32000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -538,10 +538,10 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 1.12, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.135, + "cost_per_1m_out_cached": 0, "context_window": 163840, "default_max_tokens": 81920, "can_reason": false, @@ -550,10 +550,10 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.79, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 1.7, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.13, + "cost_per_1m_out_cached": 0, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -607,9 +607,45 @@ "cost_per_1m_in": 0.27, "cost_per_1m_out": 0.41, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 81920, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek/deepseek-v4-flash", + "name": "DeepSeek: DeepSeek V4 Flash", + "cost_per_1m_in": 0.14, + "cost_per_1m_out": 0.28, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.028, + "context_window": 1048576, + "default_max_tokens": 192000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek/deepseek-v4-pro", + "name": "DeepSeek: DeepSeek V4 Pro", + "cost_per_1m_in": 1.74, + "cost_per_1m_out": 3.48, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.145, + "context_window": 1048576, + "default_max_tokens": 192000, "can_reason": true, "reasoning_levels": [ "low", @@ -807,7 +843,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -968,8 +1004,8 @@ "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, - "context_window": 256000, - "default_max_tokens": 40000, + "context_window": 262144, + "default_max_tokens": 72000, "can_reason": false, "supports_attachments": false }, @@ -1025,11 +1061,11 @@ "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", "cost_per_1m_in": 0.255, - "cost_per_1m_out": 1.02, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 204800, - "default_max_tokens": 65536, + "cost_per_1m_out_cached": 0.03, + "context_window": 196608, + "default_max_tokens": 98304, "can_reason": true, "reasoning_levels": [ "low", @@ -1198,12 +1234,12 @@ { "id": "mistralai/ministral-14b-2512", "name": "Mistral: Ministral 3 14B 2512", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.2, + "cost_per_1m_in": 0.35, + "cost_per_1m_out": 0.35, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.02, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": true }, @@ -1426,10 +1462,10 @@ { "id": "moonshotai/kimi-k2.5", "name": "MoonshotAI: Kimi K2.5", - "cost_per_1m_in": 0.45, - "cost_per_1m_out": 2.25, + "cost_per_1m_in": 0.57, + "cost_per_1m_out": 2.85, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.07, + "cost_per_1m_out_cached": 0.095, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -1444,10 +1480,10 @@ { "id": "moonshotai/kimi-k2.6", "name": "MoonshotAI: Kimi K2.6", - "cost_per_1m_in": 0.95, + "cost_per_1m_in": 0.8, "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.16, + "cost_per_1m_out_cached": 0.25, "context_window": 262144, "default_max_tokens": 131072, "can_reason": true, @@ -2149,15 +2185,51 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "openai/gpt-5.5", + "name": "OpenAI: GPT-5.5", + "cost_per_1m_in": 5, + "cost_per_1m_out": 30, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.5, + "context_window": 1050000, + "default_max_tokens": 64000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, + { + "id": "openai/gpt-5.5-pro", + "name": "OpenAI: GPT-5.5 Pro", + "cost_per_1m_in": 30, + "cost_per_1m_out": 180, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1050000, + "default_max_tokens": 64000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.49, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2188,12 +2260,12 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.045, - "cost_per_1m_out": 0.18, + "cost_per_1m_in": 0.05, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.05, "context_window": 131072, - "default_max_tokens": 16384, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2542,12 +2614,12 @@ { "id": "qwen/qwen3-14b", "name": "Qwen: Qwen3 14B", - "cost_per_1m_in": 0.12, - "cost_per_1m_out": 0.24, + "cost_per_1m_in": 0.2275, + "cost_per_1m_out": 0.91, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 40960, - "default_max_tokens": 20480, + "context_window": 131072, + "default_max_tokens": 4096, "can_reason": true, "reasoning_levels": [ "low", @@ -2578,12 +2650,12 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.2, - "cost_per_1m_out": 0.8, + "cost_per_1m_in": 0.22, + "cost_per_1m_out": 0.88, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": false }, @@ -2629,21 +2701,21 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false }, { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 1.56, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 81920, - "default_max_tokens": 16384, + "cost_per_1m_out_cached": 0.08, + "context_window": 131072, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2794,12 +2866,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.09, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.07, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, "supports_attachments": false }, @@ -2836,12 +2908,12 @@ { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.26, - "cost_per_1m_out": 1.04, + "cost_per_1m_in": 0.25, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, + "context_window": 256000, + "default_max_tokens": 8192, "can_reason": false, "supports_attachments": true }, @@ -2938,12 +3010,12 @@ { "id": "qwen/qwen3.5-397b-a17b", "name": "Qwen: Qwen3.5 397B A17B", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 3.6, + "cost_per_1m_in": 0.39, + "cost_per_1m_out": 2.34, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.3, + "cost_per_1m_out_cached": 0.195, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -2992,10 +3064,10 @@ { "id": "qwen/qwen3.5-27b", "name": "Qwen: Qwen3.5-27B", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 2.16, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.27, + "cost_per_1m_out_cached": 0, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3208,12 +3280,12 @@ { "id": "xiaomi/mimo-v2-flash", "name": "Xiaomi: MiMo-V2-Flash", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.29, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.045, + "cost_per_1m_out_cached": 0.02, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 16000, "can_reason": true, "reasoning_levels": [ "low", @@ -3315,7 +3387,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 48000, "can_reason": true, "reasoning_levels": [ "low", @@ -3403,9 +3475,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.9, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.05, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 12000, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3436,12 +3508,12 @@ { "id": "z-ai/glm-4.7-flash", "name": "Z.ai: GLM 4.7 Flash", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.4, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.43, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.01, + "cost_per_1m_out_cached": 0, "context_window": 202752, - "default_max_tokens": 20275, + "default_max_tokens": 101376, "can_reason": true, "reasoning_levels": [ "low", @@ -3574,10 +3646,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -3592,10 +3664,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 0.5, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.15, + "cost_per_1m_out_cached": 0.075, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index 8d39ef99aba1c3c88414b0d86312c574c35f7f14..f00553f444140ea04893418851e34a38d8058e1f 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -5,7 +5,7 @@ "api_endpoint": "https://api.venice.ai/api/v1", "type": "openai-compat", "default_large_model_id": "claude-opus-4-6-fast", - "default_small_model_id": "mistral-small-2603", + "default_small_model_id": "deepseek-v4-flash", "models": [ { "id": "claude-opus-4-5", @@ -127,6 +127,42 @@ "can_reason": true, "supports_attachments": false }, + { + "id": "deepseek-v4-flash", + "name": "DeepSeek V4 Flash", + "cost_per_1m_in": 0.175, + "cost_per_1m_out": 0.35, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek-v4-pro", + "name": "DeepSeek V4 Pro", + "cost_per_1m_in": 2.175, + "cost_per_1m_out": 4.35, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 32768, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "zai-org-glm-4.6", "name": "GLM 4.6", @@ -373,6 +409,42 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "openai-gpt-55", + "name": "GPT-5.5", + "cost_per_1m_in": 6.25, + "cost_per_1m_out": 37.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 131072, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, + { + "id": "openai-gpt-55-pro", + "name": "GPT-5.5 Pro", + "cost_per_1m_in": 37.5, + "cost_per_1m_out": 225, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 128000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "gemini-3-flash-preview", "name": "Gemini 3 Flash Preview", @@ -849,6 +921,22 @@ "can_reason": true, "supports_attachments": true }, + { + "id": "qwen3-6-27b", + "name": "Qwen3.6 27B", + "cost_per_1m_in": 0.325, + "cost_per_1m_out": 3.25, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 256000, + "default_max_tokens": 65536, + "can_reason": true, + "supports_attachments": true, + "options": { + "temperature": 1, + "top_p": 0.95 + } + }, { "id": "arcee-trinity-large-thinking", "name": "Trinity Large Thinking", diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index 7692ac7c0aa4a7801b28eaef4a9625ac6e1c473b..ec1f8ca44e06ffd913139bbad2faa8fe0283f5e9 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -313,6 +313,42 @@ "default_reasoning_effort": "medium", "supports_attachments": false }, + { + "id": "deepseek/deepseek-v4-flash", + "name": "DeepSeek V4 Flash", + "cost_per_1m_in": 0.14, + "cost_per_1m_out": 0.28, + "cost_per_1m_in_cached": 0.028, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, + { + "id": "deepseek/deepseek-v4-pro", + "name": "DeepSeek V4 Pro", + "cost_per_1m_in": 1.74, + "cost_per_1m_out": 3.48, + "cost_per_1m_in_cached": 0.145, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false + }, { "id": "deepseek/deepseek-r1", "name": "DeepSeek-R1", @@ -847,6 +883,42 @@ "default_reasoning_effort": "medium", "supports_attachments": true }, + { + "id": "openai/gpt-5.5", + "name": "GPT 5.5", + "cost_per_1m_in": 5, + "cost_per_1m_out": 30, + "cost_per_1m_in_cached": 0.5, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, + { + "id": "openai/gpt-5.5-pro", + "name": "GPT 5.5 Pro", + "cost_per_1m_in": 30, + "cost_per_1m_out": 180, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1000000, + "default_max_tokens": 8000, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": true + }, { "id": "openai/gpt-oss-20b", "name": "GPT OSS 120B",