From 1a40e6e80af38e2256fb92caaba72056dc1bb2a6 Mon Sep 17 00:00:00 2001 From: actions-user Date: Fri, 26 Sep 2025 02:33:30 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/openrouter.json | 262 ++++++++++++--------- 1 file changed, 157 insertions(+), 105 deletions(-) diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index 0ae5b9b80c362f05c2f7e0562497244ccf730c9a..27d1476f116c7f7bcc933fe316f57778728f12f2 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -197,7 +197,7 @@ "cost_per_1m_in_cached": 18.75, "cost_per_1m_out_cached": 1.5, "context_window": 200000, - "default_max_tokens": 16000, + "default_max_tokens": 20000, "can_reason": true, "has_reasoning_efforts": true, "supports_attachments": true @@ -283,8 +283,8 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.28, - "cost_per_1m_out": 1.1400000000000001, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1.13, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 163840, @@ -309,8 +309,8 @@ { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.27, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 163840, @@ -372,17 +372,17 @@ "supports_attachments": false }, { - "id": "google/gemini-flash-1.5", - "name": "Google: Gemini 1.5 Flash ", - "cost_per_1m_in": 0.075, - "cost_per_1m_out": 0.3, - "cost_per_1m_in_cached": 0.1583, - "cost_per_1m_out_cached": 0.01875, - "context_window": 1000000, - "default_max_tokens": 4096, - "can_reason": false, - "has_reasoning_efforts": false, - "supports_attachments": true + "id": "deepseek/deepseek-r1-distill-llama-70b", + "name": "DeepSeek: R1 Distill Llama 70B", + "cost_per_1m_in": 0.75, + "cost_per_1m_out": 0.9900000000000001, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 65536, + "can_reason": true, + "has_reasoning_efforts": true, + "supports_attachments": false }, { "id": "google/gemini-flash-1.5-8b", @@ -397,19 +397,6 @@ "has_reasoning_efforts": false, "supports_attachments": true }, - { - "id": "google/gemini-pro-1.5", - "name": "Google: Gemini 1.5 Pro", - "cost_per_1m_in": 1.25, - "cost_per_1m_out": 5, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 2000000, - "default_max_tokens": 4096, - "can_reason": false, - "has_reasoning_efforts": false, - "supports_attachments": true - }, { "id": "google/gemini-2.0-flash-001", "name": "Google: Gemini 2.0 Flash", @@ -488,6 +475,32 @@ "has_reasoning_efforts": true, "supports_attachments": true }, + { + "id": "google/gemini-2.5-flash-lite-preview-09-2025", + "name": "Google: Gemini 2.5 Flash Lite Preview 09-2025", + "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_out": 0.39999999999999997, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 1048576, + "default_max_tokens": 32768, + "can_reason": true, + "has_reasoning_efforts": true, + "supports_attachments": true + }, + { + "id": "google/gemini-2.5-flash-preview-09-2025", + "name": "Google: Gemini 2.5 Flash Preview 09-2025", + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.5, + "cost_per_1m_in_cached": 0.3833, + "cost_per_1m_out_cached": 0.075, + "context_window": 1048576, + "default_max_tokens": 32768, + "can_reason": true, + "has_reasoning_efforts": true, + "supports_attachments": true + }, { "id": "google/gemini-2.5-pro", "name": "Google: Gemini 2.5 Pro", @@ -595,12 +608,12 @@ { "id": "meta-llama/llama-3.1-405b-instruct", "name": "Meta: Llama 3.1 405B Instruct", - "cost_per_1m_in": 3, - "cost_per_1m_out": 3, + "cost_per_1m_in": 0.7999999999999999, + "cost_per_1m_out": 0.7999999999999999, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 13107, + "context_window": 32768, + "default_max_tokens": 8192, "can_reason": false, "has_reasoning_efforts": false, "supports_attachments": false @@ -608,12 +621,12 @@ { "id": "meta-llama/llama-3.1-70b-instruct", "name": "Meta: Llama 3.1 70B Instruct", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.28, + "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_out": 0.39999999999999997, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 8192, + "default_max_tokens": 13107, "can_reason": false, "has_reasoning_efforts": false, "supports_attachments": false @@ -631,6 +644,19 @@ "has_reasoning_efforts": false, "supports_attachments": false }, + { + "id": "meta-llama/llama-3.2-3b-instruct", + "name": "Meta: Llama 3.2 3B Instruct", + "cost_per_1m_in": 0.03, + "cost_per_1m_out": 0.049999999999999996, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 32768, + "default_max_tokens": 16000, + "can_reason": false, + "has_reasoning_efforts": false, + "supports_attachments": false + }, { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", @@ -673,12 +699,12 @@ { "id": "meta-llama/llama-4-maverick", "name": "Meta: Llama 4 Maverick", - "cost_per_1m_in": 0.22, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.85, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 1048576, - "default_max_tokens": 104857, + "default_max_tokens": 8192, "can_reason": false, "has_reasoning_efforts": false, "supports_attachments": true @@ -1112,6 +1138,19 @@ "has_reasoning_efforts": false, "supports_attachments": false }, + { + "id": "mistralai/pixtral-12b", + "name": "Mistral: Pixtral 12B", + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 0.15, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 13107, + "can_reason": false, + "has_reasoning_efforts": false, + "supports_attachments": true + }, { "id": "mistralai/pixtral-large-2411", "name": "Mistral: Pixtral Large 2411", @@ -1204,23 +1243,23 @@ "supports_attachments": false }, { - "id": "nousresearch/hermes-3-llama-3.1-70b", - "name": "Nous: Hermes 3 70B Instruct", - "cost_per_1m_in": 0.39999999999999997, - "cost_per_1m_out": 0.39999999999999997, + "id": "nousresearch/hermes-4-405b", + "name": "Nous: Hermes 4 405B", + "cost_per_1m_in": 0.24999987999999998, + "cost_per_1m_out": 0.999999888, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 12288, - "default_max_tokens": 1228, - "can_reason": false, - "has_reasoning_efforts": false, + "context_window": 131072, + "default_max_tokens": 13107, + "can_reason": true, + "has_reasoning_efforts": true, "supports_attachments": false }, { - "id": "nousresearch/hermes-4-405b", - "name": "Nous: Hermes 4 405B", - "cost_per_1m_in": 0.24999987999999998, - "cost_per_1m_out": 0.999999888, + "id": "nousresearch/hermes-4-70b", + "name": "Nous: Hermes 4 70B", + "cost_per_1m_in": 0.11, + "cost_per_1m_out": 0.38, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1544,8 +1583,8 @@ { "id": "openai/gpt-oss-120b", "name": "OpenAI: gpt-oss-120b", - "cost_per_1m_in": 0.049999999999999996, - "cost_per_1m_out": 0.44999999999999996, + "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_out": 0.49, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1659,16 +1698,16 @@ "supports_attachments": true }, { - "id": "qwen/qwq-32b", - "name": "Qwen: QwQ 32B", - "cost_per_1m_in": 0.15, - "cost_per_1m_out": 0.58, + "id": "qwen/qwen-2.5-72b-instruct", + "name": "Qwen2.5 72B Instruct", + "cost_per_1m_in": 0.12, + "cost_per_1m_out": 0.39, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 65536, - "can_reason": true, - "has_reasoning_efforts": true, + "context_window": 32768, + "default_max_tokens": 8192, + "can_reason": false, + "has_reasoning_efforts": false, "supports_attachments": false }, { @@ -1762,6 +1801,19 @@ "has_reasoning_efforts": true, "supports_attachments": false }, + { + "id": "qwen/qwen3-235b-a22b:free", + "name": "Qwen: Qwen3 235B A22B (free)", + "cost_per_1m_in": 0, + "cost_per_1m_out": 0, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 131072, + "default_max_tokens": 13107, + "can_reason": false, + "has_reasoning_efforts": false, + "supports_attachments": false + }, { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", @@ -1778,12 +1830,12 @@ { "id": "qwen/qwen3-235b-a22b-thinking-2507", "name": "Qwen: Qwen3 235B A22B Thinking 2507", - "cost_per_1m_in": 0.13, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.09999999999999999, + "cost_per_1m_out": 0.39, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": true, "has_reasoning_efforts": true, "supports_attachments": false @@ -1791,8 +1843,8 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.28, + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.44999999999999996, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -1817,12 +1869,12 @@ { "id": "qwen/qwen3-30b-a3b-thinking-2507", "name": "Qwen: Qwen3 30B A3B Thinking 2507", - "cost_per_1m_in": 0.08, - "cost_per_1m_out": 0.29, + "cost_per_1m_in": 0.09, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 65536, "can_reason": true, "has_reasoning_efforts": true, "supports_attachments": false @@ -1861,7 +1913,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, "has_reasoning_efforts": false, "supports_attachments": false @@ -1869,12 +1921,12 @@ { "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1, + "cost_per_1m_in": 0.29, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "has_reasoning_efforts": false, "supports_attachments": false @@ -1970,19 +2022,6 @@ "has_reasoning_efforts": true, "supports_attachments": true }, - { - "id": "qwen/qwen3-vl-235b-a22b-thinking", - "name": "Qwen: Qwen3 VL 235B A22B Thinking", - "cost_per_1m_in": 0.7, - "cost_per_1m_out": 8.399999999999999, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 16384, - "can_reason": true, - "has_reasoning_efforts": true, - "supports_attachments": true - }, { "id": "stepfun-ai/step3", "name": "StepFun: Step3", @@ -2051,12 +2090,12 @@ { "id": "z-ai/glm-4.5", "name": "Z.AI: GLM 4.5", - "cost_per_1m_in": 0.41, - "cost_per_1m_out": 1.6500000000000001, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.11, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 48000, "can_reason": true, "has_reasoning_efforts": true, "supports_attachments": false @@ -2090,12 +2129,12 @@ { "id": "z-ai/glm-4.5v", "name": "Z.AI: GLM 4.5V", - "cost_per_1m_in": 0.6, + "cost_per_1m_in": 0.5, "cost_per_1m_out": 1.7999999999999998, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 65536, - "default_max_tokens": 8192, + "default_max_tokens": 32768, "can_reason": true, "has_reasoning_efforts": true, "supports_attachments": true @@ -2103,10 +2142,10 @@ { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -2116,10 +2155,10 @@ { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, @@ -2129,10 +2168,10 @@ { "id": "x-ai/grok-3-mini", "name": "xAI: Grok 3 Mini", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.5, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.075, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -2142,10 +2181,10 @@ { "id": "x-ai/grok-3-mini-beta", "name": "xAI: Grok 3 Mini Beta", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 0.5, + "cost_per_1m_in": 0.6, + "cost_per_1m_out": 4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.075, + "cost_per_1m_out_cached": 0.15, "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, @@ -2165,6 +2204,19 @@ "has_reasoning_efforts": true, "supports_attachments": true }, + { + "id": "x-ai/grok-4-fast", + "name": "xAI: Grok 4 Fast", + "cost_per_1m_in": 0.19999999999999998, + "cost_per_1m_out": 0.5, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.049999999999999996, + "context_window": 2000000, + "default_max_tokens": 15000, + "can_reason": true, + "has_reasoning_efforts": true, + "supports_attachments": true + }, { "id": "x-ai/grok-4-fast:free", "name": "xAI: Grok 4 Fast (free)",