From 41c653aebfd20c585b11f0ca2a0f0138e05ddf58 Mon Sep 17 00:00:00 2001 From: Charm <124303983+charmcli@users.noreply.github.com> Date: Wed, 7 Jan 2026 02:44:43 +0000 Subject: [PATCH] chore: auto-update generated files --- internal/providers/configs/openrouter.json | 207 ++++++++------------- internal/providers/configs/synthetic.json | 4 +- 2 files changed, 80 insertions(+), 131 deletions(-) diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index e3c2d2267d98089b5080addb38fdd3bd8b5be2b8..1c571abe8409c4e915b7e6c1dd6d52727d3fb4be 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -253,7 +253,7 @@ "cost_per_1m_in_cached": 18.75, "cost_per_1m_out_cached": 1.5, "context_window": 200000, - "default_max_tokens": 20000, + "default_max_tokens": 16000, "can_reason": true, "reasoning_levels": [ "low", @@ -541,19 +541,13 @@ { "id": "deepseek/deepseek-chat-v3-0324", "name": "DeepSeek: DeepSeek V3 0324", - "cost_per_1m_in": 0.77, - "cost_per_1m_out": 0.77, + "cost_per_1m_in": 0.5, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", + "default_max_tokens": 16384, + "can_reason": false, "supports_attachments": false, "options": {} }, @@ -579,12 +573,12 @@ { "id": "deepseek/deepseek-v3.1-terminus", "name": "DeepSeek: DeepSeek V3.1 Terminus", - "cost_per_1m_in": 0.21, - "cost_per_1m_out": 0.7899999999999999, + "cost_per_1m_in": 0.22999999999999998, + "cost_per_1m_out": 0.8999999999999999, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.16799999999999998, + "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -617,12 +611,12 @@ { "id": "deepseek/deepseek-v3.2", "name": "DeepSeek: DeepSeek V3.2", - "cost_per_1m_in": 0.56, - "cost_per_1m_out": 1.68, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 0.39, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -636,12 +630,12 @@ { "id": "deepseek/deepseek-v3.2-exp", "name": "DeepSeek: DeepSeek V3.2 Exp", - "cost_per_1m_in": 0.27, - "cost_per_1m_out": 0.41, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.32, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.16799999999999998, "context_window": 163840, - "default_max_tokens": 32768, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -674,12 +668,12 @@ { "id": "deepseek/deepseek-r1-0528", "name": "DeepSeek: R1 0528", - "cost_per_1m_in": 0.7999999999999999, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_out": 1.75, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 163840, - "default_max_tokens": 16384, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -1062,45 +1056,6 @@ "supports_attachments": true, "options": {} }, - { - "id": "microsoft/phi-3-medium-128k-instruct", - "name": "Microsoft: Phi-3 Medium 128K Instruct", - "cost_per_1m_in": 1, - "cost_per_1m_out": 1, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 12800, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, - { - "id": "microsoft/phi-3-mini-128k-instruct", - "name": "Microsoft: Phi-3 Mini 128K Instruct", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 12800, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, - { - "id": "microsoft/phi-3.5-mini-128k-instruct", - "name": "Microsoft: Phi-3.5 Mini 128K Instruct", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 128000, - "default_max_tokens": 12800, - "can_reason": false, - "supports_attachments": false, - "options": {} - }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", @@ -1124,7 +1079,7 @@ "id": "minimax/minimax-m2.1", "name": "MiniMax: MiniMax M2.1", "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0.375, "cost_per_1m_out_cached": 0.03, "context_window": 204800, @@ -1402,12 +1357,12 @@ { "id": "mistralai/mistral-small-24b-instruct-2501", "name": "Mistral: Mistral Small 3", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.3, + "cost_per_1m_in": 0.03, + "cost_per_1m_out": 0.11, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 32768, - "default_max_tokens": 1024, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -1614,7 +1569,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": true, "reasoning_levels": [ "low", @@ -1733,25 +1688,6 @@ "supports_attachments": false, "options": {} }, - { - "id": "nousresearch/hermes-4-70b", - "name": "Nous: Hermes 4 70B", - "cost_per_1m_in": 0.11, - "cost_per_1m_out": 0.38, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 65536, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} - }, { "id": "openai/codex-mini", "name": "OpenAI: Codex Mini", @@ -1842,9 +1778,9 @@ "cost_per_1m_in": 0.09999999999999999, "cost_per_1m_out": 0.39999999999999997, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.03, + "cost_per_1m_out_cached": 0.024999999999999998, "context_window": 1047576, - "default_max_tokens": 104757, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": true, "options": {} @@ -1855,7 +1791,7 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, @@ -2286,12 +2222,12 @@ { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.016, - "cost_per_1m_out": 0.06, + "cost_per_1m_in": 0.02, + "cost_per_1m_out": 0.09999999999999999, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -2650,12 +2586,12 @@ { "id": "qwen/qwen3-235b-a22b", "name": "Qwen: Qwen3 235B A22B", - "cost_per_1m_in": 0.22, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.18, + "cost_per_1m_out": 0.54, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 131072, - "default_max_tokens": 13107, + "context_window": 40960, + "default_max_tokens": 20480, "can_reason": true, "reasoning_levels": [ "low", @@ -2669,12 +2605,12 @@ { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.09999999999999999, - "cost_per_1m_out": 0.09999999999999999, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.55, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": false, "supports_attachments": false, "options": {} @@ -2701,8 +2637,8 @@ { "id": "qwen/qwen3-30b-a3b", "name": "Qwen: Qwen3 30B A3B", - "cost_per_1m_in": 0.09, - "cost_per_1m_out": 0.44999999999999996, + "cost_per_1m_in": 0.08, + "cost_per_1m_out": 0.28, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2787,6 +2723,25 @@ "supports_attachments": false, "options": {} }, + { + "id": "qwen/qwen3-8b", + "name": "Qwen: Qwen3 8B", + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 40960, + "default_max_tokens": 4096, + "can_reason": true, + "reasoning_levels": [ + "low", + "medium", + "high" + ], + "default_reasoning_effort": "medium", + "supports_attachments": false, + "options": {} + }, { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", @@ -2803,12 +2758,12 @@ { "id": "qwen/qwen3-coder", "name": "Qwen: Qwen3 Coder 480B A35B", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.3, + "cost_per_1m_in": 0.28, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 32768, + "default_max_tokens": 26214, "can_reason": false, "supports_attachments": false, "options": {} @@ -2881,12 +2836,12 @@ { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.06, - "cost_per_1m_out": 0.6, + "cost_per_1m_in": 0.15, + "cost_per_1m_out": 1.5, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 16384, "can_reason": false, "supports_attachments": false, "options": {} @@ -2911,21 +2866,15 @@ "options": {} }, { - "id": "qwen/qwen3-vl-235b-a22b-thinking", - "name": "Qwen: Qwen3 VL 235B A22B Thinking", - "cost_per_1m_in": 0.44999999999999996, - "cost_per_1m_out": 3.5, + "id": "qwen/qwen3-vl-235b-a22b-instruct", + "name": "Qwen: Qwen3 VL 235B A22B Instruct", + "cost_per_1m_in": 0.39999999999999997, + "cost_per_1m_out": 1.5999999999999999, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 131072, - "can_reason": true, - "reasoning_levels": [ - "low", - "medium", - "high" - ], - "default_reasoning_effort": "medium", + "context_window": 131072, + "default_max_tokens": 16384, + "can_reason": false, "supports_attachments": true, "options": {} }, @@ -3279,9 +3228,9 @@ "cost_per_1m_in": 0.3, "cost_per_1m_out": 0.8999999999999999, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.055, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 16384, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3295,10 +3244,10 @@ { "id": "z-ai/glm-4.7", "name": "Z.AI: GLM 4.7", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.2, - "cost_per_1m_in_cached": 0.11, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_in": 0.54, + "cost_per_1m_out": 1.9800000000000002, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0.099, "context_window": 204800, "default_max_tokens": 65536, "can_reason": true, diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index c156b250d063fa66f55756e404c7587971612025..0065c008f47ead00483d467bcc373ba7196a377e 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -4,7 +4,7 @@ "api_key": "$SYNTHETIC_API_KEY", "api_endpoint": "https://api.synthetic.new/openai/v1", "type": "openai-compat", - "default_large_model_id": "hf:zai-org/GLM-4.7", + "default_large_model_id": "hf:zai-org/GLM-4.6", "default_small_model_id": "hf:deepseek-ai/DeepSeek-V3.1-Terminus", "models": [ { @@ -290,4 +290,4 @@ "options": {} } ] -} +} \ No newline at end of file