From b827bdec4020557754852088956afc13bca0e49e Mon Sep 17 00:00:00 2001 From: Andrey Nering Date: Wed, 1 Apr 2026 11:32:15 -0300 Subject: [PATCH] chore: re-generate all --- internal/providers/configs/aihubmix.json | 579 +++++-------- internal/providers/configs/avian.json | 12 +- internal/providers/configs/copilot.json | 75 +- internal/providers/configs/cortecs.json | 267 ++---- internal/providers/configs/huggingface.json | 76 +- internal/providers/configs/ionet.json | 42 +- internal/providers/configs/openrouter.json | 867 ++++++++------------ internal/providers/configs/synthetic.json | 54 +- internal/providers/configs/venice.json | 126 +-- internal/providers/configs/vercel.json | 456 ++++------ internal/providers/configs/xai.json | 30 +- 11 files changed, 898 insertions(+), 1686 deletions(-) diff --git a/internal/providers/configs/aihubmix.json b/internal/providers/configs/aihubmix.json index 2b63e075cb709e4c90f24b0c6df685bd5cf37d9a..bfcfa9b3553dac26bdee6f7f404e9124e78676f3 100644 --- a/internal/providers/configs/aihubmix.json +++ b/internal/providers/configs/aihubmix.json @@ -17,8 +17,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "AiHubmix-Phi-4-reasoning", @@ -36,8 +35,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "ByteDance-Seed/Seed-OSS-36B-Instruct", @@ -49,8 +47,7 @@ "context_window": 256000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "DeepSeek-R1", @@ -62,8 +59,7 @@ "context_window": 1638000, "default_max_tokens": 163800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3", @@ -75,8 +71,7 @@ "context_window": 1638000, "default_max_tokens": 163800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3-Fast", @@ -88,8 +83,7 @@ "context_window": 32000, "default_max_tokens": 3200, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3.1-Fast", @@ -101,8 +95,7 @@ "context_window": 163000, "default_max_tokens": 16300, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3.1-Terminus", @@ -114,8 +107,7 @@ "context_window": 160000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3.1-Think", @@ -133,8 +125,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3.2-Exp", @@ -146,8 +137,7 @@ "context_window": 163000, "default_max_tokens": 16300, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "DeepSeek-V3.2-Exp-Think", @@ -165,8 +155,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "ERNIE-X1.1-Preview", @@ -184,8 +173,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Kimi-K2-0905", @@ -197,8 +185,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "aihub-Phi-4-mini-instruct", @@ -210,8 +197,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "aihub-Phi-4-multimodal-instruct", @@ -223,8 +209,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic-opus-4-6", @@ -242,8 +227,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-3-5-haiku", @@ -255,8 +239,7 @@ "context_window": 200000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-3-5-sonnet", @@ -268,8 +251,7 @@ "context_window": 200000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-3-5-sonnet-20240620", @@ -281,8 +263,7 @@ "context_window": 200000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-3-7-sonnet", @@ -300,8 +281,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-haiku-4-5", @@ -319,8 +299,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-0", @@ -332,8 +311,7 @@ "context_window": 200000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-1", @@ -351,8 +329,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-5", @@ -370,8 +347,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-5-think", @@ -389,8 +365,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-6", @@ -408,8 +383,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-6-think", @@ -427,8 +401,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-0", @@ -446,8 +419,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-5", @@ -465,8 +437,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-5-think", @@ -484,8 +455,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-6", @@ -503,8 +473,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-6-think", @@ -522,8 +491,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "coding-glm-4.6-free", @@ -541,8 +509,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2", @@ -560,8 +527,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2-free", @@ -579,8 +545,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.1", @@ -598,8 +563,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.1-free", @@ -617,8 +581,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.5", @@ -636,8 +599,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.5-free", @@ -655,8 +617,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.5-highspeed", @@ -674,8 +635,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.7", @@ -693,8 +653,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.7-free", @@ -712,8 +671,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "coding-minimax-m2.7-highspeed", @@ -731,8 +689,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-math-v2", @@ -750,8 +707,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3.2", @@ -769,8 +725,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3.2-fast", @@ -788,8 +743,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3.2-speciale", @@ -807,8 +761,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3.2-think", @@ -826,8 +779,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "doubao-seed-1-6", @@ -839,8 +791,7 @@ "context_window": 256000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-1-6-flash", @@ -852,8 +803,7 @@ "context_window": 256000, "default_max_tokens": 33000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-1-6-lite", @@ -865,8 +815,7 @@ "context_window": 256000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-1-6-thinking", @@ -878,8 +827,7 @@ "context_window": 256000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-1-8", @@ -897,8 +845,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-2-0-code-preview", @@ -916,8 +863,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-2-0-lite", @@ -935,8 +881,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-2-0-mini", @@ -954,8 +899,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "doubao-seed-2-0-pro", @@ -973,8 +917,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ernie-4.5", @@ -986,8 +929,7 @@ "context_window": 160000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ernie-4.5-turbo-latest", @@ -999,8 +941,7 @@ "context_window": 135000, "default_max_tokens": 12000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ernie-4.5-turbo-vl", @@ -1012,8 +953,7 @@ "context_window": 139000, "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ernie-5.0-thinking-exp", @@ -1031,8 +971,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ernie-5.0-thinking-preview", @@ -1050,8 +989,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "ernie-x1-turbo", @@ -1069,8 +1007,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gemini-2.0-flash", @@ -1082,8 +1019,7 @@ "context_window": 1048576, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.0-flash-free", @@ -1095,8 +1031,7 @@ "context_window": 1048576, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash", @@ -1108,8 +1043,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-lite", @@ -1121,8 +1055,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-lite-nothink", @@ -1134,8 +1067,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-lite-preview-09-2025", @@ -1147,8 +1079,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-lite-preview-09-2025-nothink", @@ -1160,8 +1091,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-nothink", @@ -1173,8 +1103,7 @@ "context_window": 1047576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-preview-05-20-nothink", @@ -1186,8 +1115,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-preview-05-20-search", @@ -1199,8 +1127,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-preview-09-2025", @@ -1212,8 +1139,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash-search", @@ -1225,8 +1151,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro", @@ -1244,8 +1169,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro-preview-05-06", @@ -1263,8 +1187,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro-preview-06-05", @@ -1282,8 +1205,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro-search", @@ -1301,8 +1223,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-flash-preview", @@ -1320,8 +1241,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-flash-preview-free", @@ -1339,8 +1259,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-flash-preview-search", @@ -1358,8 +1277,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3.1-flash-lite-preview", @@ -1377,8 +1295,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3.1-flash-lite-preview-nothink", @@ -1396,8 +1313,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3.1-pro-preview", @@ -1415,8 +1331,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3.1-pro-preview-customtools", @@ -1434,8 +1349,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "glm-4.5v", @@ -1447,8 +1361,7 @@ "context_window": 64000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "glm-4.6", @@ -1466,8 +1379,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-4.6v", @@ -1479,8 +1391,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "glm-4.7", @@ -1498,8 +1409,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-5", @@ -1517,8 +1427,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-5-turbo", @@ -1536,8 +1445,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4.1", @@ -1549,8 +1457,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-free", @@ -1562,8 +1469,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-mini", @@ -1575,8 +1481,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-mini-free", @@ -1588,8 +1493,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-nano", @@ -1601,8 +1505,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-nano-free", @@ -1614,8 +1517,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o", @@ -1627,8 +1529,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o-2024-11-20", @@ -1640,8 +1541,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o-audio-preview", @@ -1653,8 +1553,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4o-free", @@ -1666,8 +1565,7 @@ "context_window": 1047576, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o-mini", @@ -1679,8 +1577,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o-mini-search-preview", @@ -1692,8 +1589,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4o-search-preview", @@ -1705,8 +1601,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5", @@ -1724,8 +1619,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-chat-latest", @@ -1737,8 +1631,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-codex", @@ -1756,8 +1649,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-mini", @@ -1775,8 +1667,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-nano", @@ -1794,8 +1685,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-pro", @@ -1813,8 +1703,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1", @@ -1832,8 +1721,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-chat-latest", @@ -1845,8 +1733,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex", @@ -1864,8 +1751,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex-max", @@ -1883,8 +1769,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex-mini", @@ -1902,8 +1787,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2", @@ -1921,8 +1805,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-chat-latest", @@ -1934,8 +1817,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-codex", @@ -1953,8 +1835,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-high", @@ -1972,8 +1853,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-low", @@ -1991,8 +1871,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-pro", @@ -2010,8 +1889,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.3-chat-latest", @@ -2023,8 +1901,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.3-codex", @@ -2042,8 +1919,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4", @@ -2061,8 +1937,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-high", @@ -2080,8 +1955,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-low", @@ -2099,8 +1973,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-mini", @@ -2112,8 +1985,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-nano", @@ -2131,8 +2003,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-pro", @@ -2150,8 +2021,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-oss-120b", @@ -2169,8 +2039,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-oss-20b", @@ -2188,8 +2057,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "grok-4", @@ -2207,8 +2075,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-1-fast-non-reasoning", @@ -2220,8 +2087,7 @@ "context_window": 2000000, "default_max_tokens": 200000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-1-fast-reasoning", @@ -2239,8 +2105,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-fast-non-reasoning", @@ -2252,8 +2117,7 @@ "context_window": 2000000, "default_max_tokens": 30000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-fast-reasoning", @@ -2271,8 +2135,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4.20-beta-0309-non-reasoning", @@ -2290,8 +2153,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4.20-beta-0309-reasoning", @@ -2309,8 +2171,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4.20-multi-agent-beta-0309", @@ -2328,8 +2189,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-code-fast-1", @@ -2347,8 +2207,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "jina-deepsearch-v1", @@ -2366,8 +2225,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "kat-dev", @@ -2379,8 +2237,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-for-coding-free", @@ -2398,8 +2255,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-0711", @@ -2411,8 +2267,7 @@ "context_window": 131000, "default_max_tokens": 13100, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-thinking", @@ -2430,8 +2285,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-turbo-preview", @@ -2443,8 +2297,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2.5", @@ -2462,8 +2315,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-4-maverick", @@ -2475,8 +2327,7 @@ "context_window": 1048576, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-4-scout", @@ -2488,8 +2339,7 @@ "context_window": 131000, "default_max_tokens": 13100, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mimo-v2-flash-free", @@ -2501,8 +2351,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mimo-v2-omni", @@ -2514,8 +2363,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mimo-v2-pro", @@ -2527,8 +2375,7 @@ "context_window": 1000000, "default_max_tokens": 100000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "minimax-m2", @@ -2546,8 +2393,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m2.1", @@ -2565,8 +2411,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m2.5", @@ -2584,8 +2429,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m2.5-highspeed", @@ -2603,8 +2447,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m2.7", @@ -2622,8 +2465,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-large-3", @@ -2635,8 +2477,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nvidia-nemotron-3-super-120b-a12b", @@ -2654,8 +2495,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "o3", @@ -2673,8 +2513,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "o3-mini", @@ -2692,8 +2531,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "o3-pro", @@ -2711,8 +2549,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "o4-mini", @@ -2730,8 +2567,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-235b-a22b", @@ -2743,8 +2579,7 @@ "context_window": 131100, "default_max_tokens": 13110, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-235b-a22b-instruct-2507", @@ -2756,8 +2591,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-235b-a22b-thinking-2507", @@ -2775,8 +2609,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-coder-30b-a3b-instruct", @@ -2788,8 +2621,7 @@ "context_window": 2000000, "default_max_tokens": 262000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-480b-a35b-instruct", @@ -2801,8 +2633,7 @@ "context_window": 262000, "default_max_tokens": 26200, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-flash", @@ -2814,8 +2645,7 @@ "context_window": 256000, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-next", @@ -2827,8 +2657,7 @@ "context_window": 2000000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-plus", @@ -2840,8 +2669,7 @@ "context_window": 1048576, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-plus-2025-07-22", @@ -2853,8 +2681,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-max", @@ -2866,8 +2693,7 @@ "context_window": 262144, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-max-2026-01-23", @@ -2885,8 +2711,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-next-80b-a3b-instruct", @@ -2898,8 +2723,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-next-80b-a3b-thinking", @@ -2917,8 +2741,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-235b-a22b-instruct", @@ -2930,8 +2753,7 @@ "context_window": 131000, "default_max_tokens": 33000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-235b-a22b-thinking", @@ -2949,8 +2771,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-30b-a3b-instruct", @@ -2962,8 +2783,7 @@ "context_window": 128000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-30b-a3b-thinking", @@ -2981,8 +2801,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-flash", @@ -2994,8 +2813,7 @@ "context_window": 254000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-flash-2026-01-22", @@ -3007,8 +2825,7 @@ "context_window": 254000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-plus", @@ -3020,8 +2837,7 @@ "context_window": 256000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-122b-a10b", @@ -3039,8 +2855,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-27b", @@ -3058,8 +2873,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-35b-a3b", @@ -3077,8 +2891,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-397b-a17b", @@ -3096,8 +2909,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-flash", @@ -3115,8 +2927,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.5-plus", @@ -3134,8 +2945,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3.6-plus-preview-free", @@ -3153,8 +2963,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "step-3.5-flash", @@ -3166,8 +2975,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "step-3.5-flash-free", @@ -3179,8 +2987,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true } ], "default_headers": { diff --git a/internal/providers/configs/avian.json b/internal/providers/configs/avian.json index 476f81feddd6e482904efa6d0569326567824562..e64d00442c5aca4692fa957499ae204dcf637577 100644 --- a/internal/providers/configs/avian.json +++ b/internal/providers/configs/avian.json @@ -23,8 +23,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-5", @@ -36,8 +35,7 @@ "context_window": 204800, "default_max_tokens": 131072, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2.5", @@ -49,8 +47,7 @@ "context_window": 262144, "default_max_tokens": 262144, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.5", @@ -62,8 +59,7 @@ "context_window": 196608, "default_max_tokens": 131072, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] } diff --git a/internal/providers/configs/copilot.json b/internal/providers/configs/copilot.json index b5e15efe6b77fbb3f03f472c6c038bff41a8ad51..026bb7fffda5e91657763d8fb61689220d265125 100644 --- a/internal/providers/configs/copilot.json +++ b/internal/providers/configs/copilot.json @@ -16,8 +16,7 @@ "context_window": 200000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4.5", @@ -29,8 +28,7 @@ "context_window": 200000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4.6", @@ -42,8 +40,7 @@ "context_window": 200000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4.6-fast", @@ -55,8 +52,7 @@ "context_window": 200000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4", @@ -68,8 +64,7 @@ "context_window": 216000, "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4.5", @@ -81,8 +76,7 @@ "context_window": 200000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4.6", @@ -94,8 +88,7 @@ "context_window": 200000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro", @@ -107,8 +100,7 @@ "context_window": 128000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-flash-preview", @@ -120,8 +112,7 @@ "context_window": 128000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3.1-pro-preview", @@ -133,8 +124,7 @@ "context_window": 128000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "goldeneye-free-auto", @@ -146,8 +136,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-3.5-turbo-0613", @@ -159,8 +148,7 @@ "context_window": 16384, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4-0125-preview", @@ -172,8 +160,7 @@ "context_window": 128000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4-0613", @@ -185,8 +172,7 @@ "context_window": 32768, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-5-mini", @@ -198,8 +184,7 @@ "context_window": 264000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1", @@ -211,8 +196,7 @@ "context_window": 264000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex", @@ -224,8 +208,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex-max", @@ -237,8 +220,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.1-codex-mini", @@ -250,8 +232,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2", @@ -263,8 +244,7 @@ "context_window": 264000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.2-codex", @@ -276,8 +256,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.3-codex", @@ -289,8 +268,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4", @@ -302,8 +280,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5.4-mini", @@ -315,8 +292,7 @@ "context_window": 400000, "default_max_tokens": 128000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-code-fast-1", @@ -328,8 +304,7 @@ "context_window": 128000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] } diff --git a/internal/providers/configs/cortecs.json b/internal/providers/configs/cortecs.json index eb7c86fdc66ef74f5ecea1159b002dabb0c86f3a..eb086a34c61d6a2626ef69d93d9df83777ef1c98 100644 --- a/internal/providers/configs/cortecs.json +++ b/internal/providers/configs/cortecs.json @@ -23,8 +23,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-4.6", @@ -42,8 +41,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-chat-v3.1", @@ -61,8 +59,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen-2.5-72b-instruct", @@ -74,8 +71,7 @@ "context_window": 33000, "default_max_tokens": 3300, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3.5-397b-a17b", @@ -93,8 +89,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3.2", @@ -112,8 +107,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-small-2603", @@ -131,8 +125,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "minimax-m2.5", @@ -150,8 +143,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "claude-4-6-sonnet", @@ -169,8 +161,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "glm-4.7-flash", @@ -182,8 +173,7 @@ "context_window": 203000, "default_max_tokens": 20300, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2.5", @@ -201,8 +191,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "claude-opus4-6", @@ -220,8 +209,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "minimax-m2", @@ -239,8 +227,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-4.7", @@ -258,8 +245,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m2.1", @@ -277,8 +263,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "llama-guard-3-8b", @@ -290,8 +275,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-vl-235b-a22b", @@ -309,8 +293,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral-small-creative", @@ -322,8 +305,7 @@ "context_window": 32000, "default_max_tokens": 3200, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia-nemotron-3-nano-30b-a3b", @@ -341,8 +323,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "claude-opus4-5", @@ -360,8 +341,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-next-80b-a3b-thinking", @@ -379,8 +359,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "holo2-30b-a3b", @@ -398,8 +377,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "devstral-2512", @@ -411,8 +389,7 @@ "context_window": 262000, "default_max_tokens": 20000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nova-2-lite", @@ -430,8 +407,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-oss-safeguard-120b", @@ -449,8 +425,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-large-2512", @@ -462,8 +437,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ministral-8b-2512", @@ -475,8 +449,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ministral-3b-2512", @@ -488,8 +461,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "ministral-14b-2512", @@ -501,8 +473,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "kimi-k2-thinking", @@ -520,8 +491,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "intellect-3", @@ -539,8 +509,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-5.1", @@ -558,8 +527,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nemotron-nano-v2-12b", @@ -577,8 +545,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-haiku-4-5", @@ -596,8 +563,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-4-5-sonnet", @@ -615,8 +581,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "magistral-medium-2509", @@ -634,8 +599,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "magistral-small-2509", @@ -653,8 +617,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "hermes-4-70b", @@ -666,8 +629,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-5", @@ -685,8 +647,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-235b-a22b-thinking-2507", @@ -704,8 +665,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-oss-120b", @@ -723,8 +683,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-30b-a3b-instruct-2507", @@ -742,8 +701,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-30b-a3b-thinking-2507", @@ -761,8 +719,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-480b-a35b-instruct", @@ -780,8 +737,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-oss-20b", @@ -799,8 +755,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-instruct", @@ -812,8 +767,7 @@ "context_window": 131000, "default_max_tokens": 13100, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-4.5", @@ -831,8 +785,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "glm-4.5-air", @@ -850,8 +803,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-7b-instruct-v0.3", @@ -863,8 +815,7 @@ "context_window": 127000, "default_max_tokens": 12700, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-large-2402", @@ -882,8 +833,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "pixtral-large-2502", @@ -901,8 +851,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral-small-3.2-24b-instruct-2506", @@ -914,8 +863,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-32b", @@ -933,8 +881,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-235b-a22b-instruct-2507", @@ -952,8 +899,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-30b-a3b-instruct", @@ -971,8 +917,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4.1", @@ -990,8 +935,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-mini", @@ -1009,8 +953,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-4.1-nano", @@ -1028,8 +971,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nova-micro-v1", @@ -1047,8 +989,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nova-lite-v1", @@ -1066,8 +1007,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nova-pro-v1", @@ -1085,8 +1025,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4", @@ -1104,8 +1043,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-3-7-sonnet", @@ -1123,8 +1061,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-3.1-nemotron-ultra-253b-v1", @@ -1142,8 +1079,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "llama-4-maverick", @@ -1155,8 +1091,7 @@ "context_window": 1050000, "default_max_tokens": 105000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-v3-0324", @@ -1174,8 +1109,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-small-2503", @@ -1187,8 +1121,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral-small-2506", @@ -1200,8 +1133,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.0-flash-001", @@ -1213,8 +1145,7 @@ "context_window": 1048576, "default_max_tokens": 104857, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.0-flash-lite-001", @@ -1226,8 +1157,7 @@ "context_window": 1048576, "default_max_tokens": 104857, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-flash", @@ -1245,8 +1175,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-2.5-pro", @@ -1264,8 +1193,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemma-3-27b-it", @@ -1283,8 +1211,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "deepseek-r1-distill-llama-70b", @@ -1302,8 +1229,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-r1-0528", @@ -1321,8 +1247,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "codestral-2508", @@ -1334,8 +1259,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "llama-3.3-70b-instruct", @@ -1353,8 +1277,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4o", @@ -1372,8 +1295,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-mini", @@ -1391,8 +1313,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gpt-5-nano", @@ -1410,8 +1331,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral-large-2411", @@ -1429,8 +1349,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hermes-4-405b", @@ -1442,8 +1361,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-nemo-instruct-2407", @@ -1455,8 +1373,7 @@ "context_window": 131072, "default_max_tokens": 11800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "devstral-medium-2507", @@ -1468,8 +1385,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "devstral-small-2507", @@ -1481,8 +1397,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-medium-2508", @@ -1500,8 +1415,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-3.1-405b-instruct", @@ -1519,8 +1433,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "gpt-4o-mini", @@ -1538,8 +1451,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-3.1-8b-instruct", @@ -1557,8 +1469,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] } diff --git a/internal/providers/configs/huggingface.json b/internal/providers/configs/huggingface.json index 9d4189a85dc7fb887bc7e572944703b65805a999..b0558ca3e4ad2352da047fc8d61bcdd62b687131 100644 --- a/internal/providers/configs/huggingface.json +++ b/internal/providers/configs/huggingface.json @@ -17,8 +17,7 @@ "context_window": 204800, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "MiniMaxAI/MiniMax-M2.5:fireworks-ai", @@ -30,8 +29,7 @@ "context_window": 204800, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Qwen/Qwen3-32B:groq", @@ -43,8 +41,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Qwen/Qwen3-8B:fireworks-ai", @@ -56,8 +53,7 @@ "context_window": 40960, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Qwen/Qwen3-VL-30B-A3B-Instruct:fireworks-ai", @@ -69,8 +65,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Qwen/Qwen3-VL-30B-A3B-Thinking:fireworks-ai", @@ -82,8 +77,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepcogito/cogito-671b-v2.1:fireworks-ai", @@ -95,21 +89,7 @@ "context_window": 163840, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} - }, - { - "id": "deepseek-ai/DeepSeek-V3.1:fireworks-ai", - "name": "deepseek-ai/DeepSeek-V3.1 (fireworks-ai)", - "cost_per_1m_in": 0, - "cost_per_1m_out": 0, - "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 163840, - "default_max_tokens": 8192, - "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek-ai/DeepSeek-V3.2:fireworks-ai", @@ -121,8 +101,7 @@ "context_window": 163840, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/Llama-3.3-70B-Instruct:groq", @@ -134,8 +113,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/Llama-4-Scout-17B-16E-Instruct:groq", @@ -147,8 +125,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2-Instruct-0905:fireworks-ai", @@ -160,8 +137,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2-Instruct-0905:groq", @@ -173,8 +149,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2-Thinking:fireworks-ai", @@ -186,8 +161,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2.5:fireworks-ai", @@ -199,8 +173,7 @@ "context_window": 262144, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-120b:fireworks-ai", @@ -212,8 +185,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-120b:groq", @@ -225,8 +197,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b:fireworks-ai", @@ -238,8 +209,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b:groq", @@ -251,8 +221,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-safeguard-20b:groq", @@ -264,8 +233,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-4.7:fireworks-ai", @@ -277,8 +245,7 @@ "context_window": 202752, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-5:fireworks-ai", @@ -290,8 +257,7 @@ "context_window": 202752, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false } ], "default_headers": { diff --git a/internal/providers/configs/ionet.json b/internal/providers/configs/ionet.json index 6554d316b8ea63cf2eae1bf6ff6b3364407271de..bcb4d9db58ffa78d1f3a56c5964564dc042cd923 100644 --- a/internal/providers/configs/ionet.json +++ b/internal/providers/configs/ionet.json @@ -17,8 +17,7 @@ "context_window": 106000, "default_max_tokens": 10600, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/Llama-3.3-70B-Instruct", @@ -36,8 +35,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "MiniMaxAI/MiniMax-M2.5", @@ -49,8 +47,7 @@ "context_window": 196600, "default_max_tokens": 19660, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/Mistral-Large-Instruct-2411", @@ -62,8 +59,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "moonshotai/Kimi-K2-Instruct-0905", @@ -75,8 +71,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2-Thinking", @@ -94,8 +89,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/Kimi-K2.5", @@ -113,8 +107,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-120b", @@ -132,8 +125,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b", @@ -151,8 +143,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "Qwen/Qwen3-Next-80B-A3B-Instruct", @@ -164,8 +155,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-4.6", @@ -183,8 +173,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-4.7", @@ -202,8 +191,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-4.7-Flash", @@ -221,8 +209,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org/GLM-5", @@ -240,8 +227,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] } diff --git a/internal/providers/configs/openrouter.json b/internal/providers/configs/openrouter.json index b95fd568a4e774d0539b0c09f483efac857a621c..77c505b7ab24576cab5f09735900d12ba21ab95f 100644 --- a/internal/providers/configs/openrouter.json +++ b/internal/providers/configs/openrouter.json @@ -17,8 +17,7 @@ "context_window": 256000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "allenai/olmo-3.1-32b-instruct", @@ -30,8 +29,7 @@ "context_window": 65536, "default_max_tokens": 6553, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "amazon/nova-2-lite-v1", @@ -49,8 +47,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "amazon/nova-lite-v1", @@ -62,8 +59,7 @@ "context_window": 300000, "default_max_tokens": 2560, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "amazon/nova-micro-v1", @@ -75,8 +71,7 @@ "context_window": 128000, "default_max_tokens": 2560, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "amazon/nova-premier-v1", @@ -88,8 +83,7 @@ "context_window": 1000000, "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "amazon/nova-pro-v1", @@ -101,8 +95,7 @@ "context_window": 300000, "default_max_tokens": 2560, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3-haiku", @@ -114,8 +107,7 @@ "context_window": 200000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.5-haiku", @@ -127,8 +119,7 @@ "context_window": 200000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.5-sonnet", @@ -140,8 +131,7 @@ "context_window": 200000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.7-sonnet", @@ -151,7 +141,7 @@ "cost_per_1m_in_cached": 3.75, "cost_per_1m_out_cached": 0.3, "context_window": 200000, - "default_max_tokens": 32000, + "default_max_tokens": 64000, "can_reason": true, "reasoning_levels": [ "low", @@ -159,8 +149,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.7-sonnet:thinking", @@ -178,8 +167,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-haiku-4.5", @@ -197,8 +185,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4", @@ -216,8 +203,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.1", @@ -235,8 +221,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.5", @@ -254,8 +239,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.6", @@ -273,8 +257,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4", @@ -292,8 +275,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4.5", @@ -311,8 +293,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4.6", @@ -330,8 +311,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "arcee-ai/trinity-large-preview:free", @@ -343,8 +323,7 @@ "context_window": 131000, "default_max_tokens": 13100, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "arcee-ai/trinity-mini", @@ -362,8 +341,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "arcee-ai/trinity-mini:free", @@ -381,8 +359,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "arcee-ai/virtuoso-large", @@ -394,8 +371,7 @@ "context_window": 131072, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "baidu/ernie-4.5-21b-a3b", @@ -407,8 +383,7 @@ "context_window": 120000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "baidu/ernie-4.5-vl-28b-a3b", @@ -426,8 +401,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "bytedance-seed/seed-1.6", @@ -445,8 +419,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "bytedance-seed/seed-1.6-flash", @@ -464,8 +437,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "bytedance-seed/seed-2.0-lite", @@ -483,8 +455,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "bytedance-seed/seed-2.0-mini", @@ -502,8 +473,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "cohere/command-r-08-2024", @@ -515,8 +485,7 @@ "context_window": 128000, "default_max_tokens": 2000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "cohere/command-r-plus-08-2024", @@ -528,8 +497,7 @@ "context_window": 128000, "default_max_tokens": 2000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-chat", @@ -541,8 +509,7 @@ "context_window": 64000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-chat-v3-0324", @@ -554,16 +521,15 @@ "context_window": 163840, "default_max_tokens": 81920, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-chat-v3.1", "name": "DeepSeek: DeepSeek V3.1", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 1.7, + "cost_per_1m_in": 0.21, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.13, "context_window": 163840, "default_max_tokens": 16384, "can_reason": true, @@ -573,8 +539,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.1-terminus", @@ -592,8 +557,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.2", @@ -611,8 +575,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.2-exp", @@ -630,8 +593,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-r1", @@ -649,8 +611,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-r1-0528", @@ -668,8 +629,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "essentialai/rnj-1-instruct", @@ -681,8 +641,7 @@ "context_window": 32768, "default_max_tokens": 3276, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "google/gemini-2.0-flash-001", @@ -694,8 +653,7 @@ "context_window": 1048576, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.0-flash-lite-001", @@ -707,8 +665,7 @@ "context_window": 1048576, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-flash", @@ -726,8 +683,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-flash-lite", @@ -745,8 +701,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-flash-lite-preview-09-2025", @@ -764,8 +719,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-pro", @@ -783,8 +737,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-pro-preview-05-06", @@ -802,8 +755,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-pro-preview", @@ -821,8 +773,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3-flash-preview", @@ -832,7 +783,7 @@ "cost_per_1m_in_cached": 0.08333, "cost_per_1m_out_cached": 0.05, "context_window": 1048576, - "default_max_tokens": 32767, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -840,8 +791,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3.1-flash-lite-preview", @@ -859,8 +809,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3.1-pro-preview", @@ -878,8 +827,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3.1-pro-preview-customtools", @@ -897,8 +845,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "inception/mercury", @@ -910,8 +857,7 @@ "context_window": 128000, "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "inception/mercury-2", @@ -929,8 +875,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "inception/mercury-coder", @@ -942,8 +887,7 @@ "context_window": 128000, "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kwaipilot/kat-coder-pro-v2", @@ -952,11 +896,10 @@ "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.06, - "context_window": 262144, - "default_max_tokens": 72000, + "context_window": 256000, + "default_max_tokens": 40000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meituan/longcat-flash-chat", @@ -968,8 +911,7 @@ "context_window": 131072, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/llama-3.1-70b-instruct", @@ -979,23 +921,21 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/llama-3.3-70b-instruct", "name": "Meta: Llama 3.3 70B Instruct", - "cost_per_1m_in": 0.88, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.59, + "cost_per_1m_out": 0.79, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 1024, + "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta-llama/llama-3.3-70b-instruct:free", @@ -1007,8 +947,19 @@ "context_window": 65536, "default_max_tokens": 6553, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false + }, + { + "id": "meta-llama/llama-4-maverick", + "name": "Meta: Llama 4 Maverick", + "cost_per_1m_in": 0.35, + "cost_per_1m_out": 1.15, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 524288, + "default_max_tokens": 4096, + "can_reason": false, + "supports_attachments": true }, { "id": "meta-llama/llama-4-scout", @@ -1020,18 +971,17 @@ "context_window": 1310720, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "minimax/minimax-m2", "name": "MiniMax: MiniMax M2", - "cost_per_1m_in": 0.3, - "cost_per_1m_out": 1.2, + "cost_per_1m_in": 0.255, + "cost_per_1m_out": 1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.03, - "context_window": 204800, - "default_max_tokens": 65536, + "context_window": 196608, + "default_max_tokens": 98304, "can_reason": true, "reasoning_levels": [ "low", @@ -1039,8 +989,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.1", @@ -1058,18 +1007,17 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.5", "name": "MiniMax: MiniMax M2.5", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.4, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 1.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.06, + "cost_per_1m_out_cached": 0.03, "context_window": 204800, - "default_max_tokens": 65536, + "default_max_tokens": 65550, "can_reason": true, "reasoning_levels": [ "low", @@ -1077,8 +1025,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.5:free", @@ -1096,8 +1043,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.7", @@ -1115,8 +1061,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/mistral-large", @@ -1128,8 +1073,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/mistral-large-2407", @@ -1141,8 +1085,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/mistral-large-2411", @@ -1154,8 +1097,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/codestral-2508", @@ -1167,8 +1109,7 @@ "context_window": 256000, "default_max_tokens": 25600, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/devstral-2512", @@ -1180,8 +1121,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/devstral-medium", @@ -1193,8 +1133,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/devstral-small", @@ -1206,34 +1145,31 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/ministral-14b-2512", "name": "Mistral: Ministral 3 14B 2512", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.35, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.02, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/ministral-3b-2512", "name": "Mistral: Ministral 3 3B 2512", - "cost_per_1m_in": 0.35, - "cost_per_1m_out": 0.35, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.1, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.01, "context_window": 131072, - "default_max_tokens": 65536, + "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/ministral-8b-2512", @@ -1245,8 +1181,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-large-2512", @@ -1258,8 +1193,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-medium-3", @@ -1271,8 +1205,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-medium-3.1", @@ -1284,8 +1217,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-nemo", @@ -1297,8 +1229,19 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false + }, + { + "id": "mistralai/mistral-small-24b-instruct-2501", + "name": "Mistral: Mistral Small 3", + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, + "cost_per_1m_in_cached": 0, + "cost_per_1m_out_cached": 0, + "context_window": 32768, + "default_max_tokens": 1024, + "can_reason": false, + "supports_attachments": false }, { "id": "mistralai/mistral-small-3.2-24b-instruct", @@ -1310,8 +1253,7 @@ "context_window": 256000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-small-2603", @@ -1329,8 +1271,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-small-creative", @@ -1342,8 +1283,7 @@ "context_window": 32768, "default_max_tokens": 3276, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/mixtral-8x22b-instruct", @@ -1355,8 +1295,7 @@ "context_window": 65536, "default_max_tokens": 6553, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/mixtral-8x7b-instruct", @@ -1368,8 +1307,7 @@ "context_window": 32768, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/pixtral-large-2411", @@ -1381,8 +1319,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistralai/mistral-saba", @@ -1394,8 +1331,7 @@ "context_window": 32768, "default_max_tokens": 3276, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistralai/voxtral-small-24b-2507", @@ -1407,8 +1343,7 @@ "context_window": 32000, "default_max_tokens": 3200, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2", @@ -1420,21 +1355,19 @@ "context_window": 131072, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-0905", "name": "MoonshotAI: Kimi K2 0905", - "cost_per_1m_in": 0.6, - "cost_per_1m_out": 2.5, + "cost_per_1m_in": 1, + "cost_per_1m_out": 3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.5, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-thinking", @@ -1452,8 +1385,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2.5", @@ -1471,8 +1403,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nvidia/llama-3.1-nemotron-70b-instruct", @@ -1484,8 +1415,7 @@ "context_window": 131072, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/llama-3.3-nemotron-super-49b-v1.5", @@ -1503,8 +1433,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-3-nano-30b-a3b", @@ -1522,8 +1451,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-3-nano-30b-a3b:free", @@ -1541,8 +1469,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-3-super-120b-a12b", @@ -1560,8 +1487,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-3-super-120b-a12b:free", @@ -1579,8 +1505,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-nano-12b-v2-vl:free", @@ -1598,8 +1523,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nvidia/nemotron-nano-9b-v2", @@ -1617,8 +1541,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia/nemotron-nano-9b-v2:free", @@ -1636,8 +1559,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nex-agi/deepseek-v3.1-nex-n1", @@ -1649,8 +1571,7 @@ "context_window": 131072, "default_max_tokens": 81920, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-audio", @@ -1662,8 +1583,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-audio-mini", @@ -1675,8 +1595,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-4-turbo", @@ -1688,8 +1607,7 @@ "context_window": 128000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4-1106-preview", @@ -1701,8 +1619,7 @@ "context_window": 128000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-4-turbo-preview", @@ -1714,8 +1631,7 @@ "context_window": 128000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-4.1", @@ -1727,8 +1643,7 @@ "context_window": 1047576, "default_max_tokens": 104757, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4.1-mini", @@ -1740,8 +1655,7 @@ "context_window": 1047576, "default_max_tokens": 104757, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4.1-nano", @@ -1749,12 +1663,11 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.025, + "cost_per_1m_out_cached": 0.03, "context_window": 1047576, - "default_max_tokens": 16384, + "default_max_tokens": 104757, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o", @@ -1762,12 +1675,11 @@ "cost_per_1m_in": 2.5, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 1.25, + "cost_per_1m_out_cached": 0, "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-2024-05-13", @@ -1779,8 +1691,7 @@ "context_window": 128000, "default_max_tokens": 2048, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-2024-08-06", @@ -1792,8 +1703,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-2024-11-20", @@ -1805,8 +1715,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o:extended", @@ -1818,8 +1727,7 @@ "context_window": 128000, "default_max_tokens": 32000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-audio-preview", @@ -1831,8 +1739,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-4o-mini", @@ -1844,8 +1751,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-mini-2024-07-18", @@ -1857,8 +1763,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5", @@ -1876,8 +1781,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-codex", @@ -1895,8 +1799,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-image", @@ -1914,8 +1817,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-image-mini", @@ -1933,8 +1835,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-mini", @@ -1952,8 +1853,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-nano", @@ -1971,8 +1871,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-pro", @@ -1990,8 +1889,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1", @@ -1999,7 +1897,7 @@ "cost_per_1m_in": 1.25, "cost_per_1m_out": 10, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.125, + "cost_per_1m_out_cached": 0.13, "context_window": 400000, "default_max_tokens": 64000, "can_reason": true, @@ -2009,8 +1907,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-chat", @@ -2022,8 +1919,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex", @@ -2041,8 +1937,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex-max", @@ -2060,8 +1955,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex-mini", @@ -2079,8 +1973,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2", @@ -2098,8 +1991,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-chat", @@ -2109,10 +2001,9 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.175, "context_window": 128000, - "default_max_tokens": 8192, + "default_max_tokens": 16000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-pro", @@ -2130,8 +2021,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-codex", @@ -2149,8 +2039,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.3-chat", @@ -2162,8 +2051,7 @@ "context_window": 128000, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.3-codex", @@ -2181,8 +2069,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4", @@ -2200,8 +2087,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-mini", @@ -2219,8 +2105,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-nano", @@ -2238,8 +2123,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-pro", @@ -2257,8 +2141,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-oss-120b", @@ -2276,8 +2159,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-120b:free", @@ -2295,14 +2177,13 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b", "name": "OpenAI: gpt-oss-20b", - "cost_per_1m_in": 0.03, - "cost_per_1m_out": 0.14, + "cost_per_1m_in": 0.07, + "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, @@ -2314,8 +2195,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b:free", @@ -2333,8 +2213,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-safeguard-20b", @@ -2352,8 +2231,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/o1", @@ -2371,8 +2249,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3", @@ -2390,8 +2267,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3-deep-research", @@ -2409,8 +2285,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3-mini", @@ -2428,8 +2303,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/o3-mini-high", @@ -2447,8 +2321,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/o3-pro", @@ -2466,8 +2339,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o4-mini", @@ -2485,8 +2357,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o4-mini-deep-research", @@ -2504,8 +2375,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o4-mini-high", @@ -2523,8 +2393,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "prime-intellect/intellect-3", @@ -2542,8 +2411,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-2.5-72b-instruct", @@ -2555,8 +2423,7 @@ "context_window": 32768, "default_max_tokens": 8192, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwq-32b", @@ -2574,8 +2441,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-plus-2025-07-28", @@ -2587,8 +2453,7 @@ "context_window": 1000000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-plus-2025-07-28:thinking", @@ -2606,8 +2471,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-vl-max", @@ -2619,8 +2483,7 @@ "context_window": 131072, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen-max", @@ -2632,8 +2495,7 @@ "context_window": 32768, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-plus", @@ -2645,8 +2507,7 @@ "context_window": 1000000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-turbo", @@ -2658,8 +2519,7 @@ "context_window": 131072, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen-2.5-7b-instruct", @@ -2671,8 +2531,7 @@ "context_window": 32768, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-14b", @@ -2690,8 +2549,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-235b-a22b", @@ -2709,21 +2567,19 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-235b-a22b-2507", "name": "Qwen: Qwen3 235B A22B Instruct 2507", - "cost_per_1m_in": 0.22, - "cost_per_1m_out": 0.88, + "cost_per_1m_in": 0.2, + "cost_per_1m_out": 0.8, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 8192, + "default_max_tokens": 131072, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-235b-a22b-thinking-2507", @@ -2741,8 +2597,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-30b-a3b", @@ -2760,8 +2615,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-30b-a3b-instruct-2507", @@ -2771,10 +2625,9 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-30b-a3b-thinking-2507", @@ -2792,8 +2645,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-32b", @@ -2811,8 +2663,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-8b", @@ -2830,21 +2681,19 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder-30b-a3b-instruct", "name": "Qwen: Qwen3 Coder 30B A3B Instruct", - "cost_per_1m_in": 0.07, - "cost_per_1m_out": 0.27, + "cost_per_1m_in": 0.1, + "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 160000, - "default_max_tokens": 16384, + "context_window": 262144, + "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder", @@ -2856,8 +2705,7 @@ "context_window": 262144, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder:free", @@ -2869,8 +2717,7 @@ "context_window": 262000, "default_max_tokens": 131000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder-flash", @@ -2882,8 +2729,7 @@ "context_window": 1000000, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder-next", @@ -2895,8 +2741,7 @@ "context_window": 262144, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-coder-plus", @@ -2908,8 +2753,7 @@ "context_window": 1000000, "default_max_tokens": 32768, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-max", @@ -2921,8 +2765,7 @@ "context_window": 262144, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-max-thinking", @@ -2940,21 +2783,19 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-next-80b-a3b-instruct", "name": "Qwen: Qwen3 Next 80B A3B Instruct", - "cost_per_1m_in": 0.09, + "cost_per_1m_in": 0.1, "cost_per_1m_out": 1.1, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 262144, - "default_max_tokens": 26214, + "default_max_tokens": 131072, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-next-80b-a3b-instruct:free", @@ -2966,8 +2807,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-next-80b-a3b-thinking", @@ -2985,21 +2825,19 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen/qwen3-vl-235b-a22b-instruct", "name": "Qwen: Qwen3 VL 235B A22B Instruct", - "cost_per_1m_in": 0.25, - "cost_per_1m_out": 1.5, + "cost_per_1m_in": 0.26, + "cost_per_1m_out": 1.04, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 256000, - "default_max_tokens": 8192, + "context_window": 131072, + "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-235b-a22b-thinking", @@ -3017,8 +2855,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-30b-a3b-instruct", @@ -3030,8 +2867,7 @@ "context_window": 131072, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-30b-a3b-thinking", @@ -3049,8 +2885,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-32b-instruct", @@ -3062,8 +2897,7 @@ "context_window": 131072, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-8b-instruct", @@ -3075,8 +2909,7 @@ "context_window": 262144, "default_max_tokens": 131072, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3-vl-8b-thinking", @@ -3094,8 +2927,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-397b-a17b", @@ -3113,8 +2945,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-plus-02-15", @@ -3132,16 +2963,15 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-122b-a10b", "name": "Qwen: Qwen3.5-122B-A10B", - "cost_per_1m_in": 0.4, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.3, + "cost_per_1m_out": 2.4, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, + "cost_per_1m_out_cached": 0.3, "context_window": 262144, "default_max_tokens": 32768, "can_reason": true, @@ -3151,8 +2981,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-27b", @@ -3170,8 +2999,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-35b-a3b", @@ -3189,18 +3017,17 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-9b", "name": "Qwen: Qwen3.5-9B", - "cost_per_1m_in": 0.1, + "cost_per_1m_in": 0.05, "cost_per_1m_out": 0.15, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 26214, + "context_window": 256000, + "default_max_tokens": 16384, "can_reason": true, "reasoning_levels": [ "low", @@ -3208,8 +3035,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.5-flash-02-23", @@ -3227,8 +3053,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen/qwen3.6-plus-preview:free", @@ -3246,8 +3071,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "relace/relace-search", @@ -3259,8 +3083,7 @@ "context_window": 256000, "default_max_tokens": 64000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "stepfun/step-3.5-flash", @@ -3268,9 +3091,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0, - "context_window": 262144, - "default_max_tokens": 32768, + "cost_per_1m_out_cached": 0.02, + "context_window": 256000, + "default_max_tokens": 128000, "can_reason": true, "reasoning_levels": [ "low", @@ -3278,8 +3101,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "stepfun/step-3.5-flash:free", @@ -3297,8 +3119,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "tngtech/deepseek-r1t2-chimera", @@ -3316,8 +3137,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "thedrummer/rocinante-12b", @@ -3329,8 +3149,7 @@ "context_window": 32768, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "thedrummer/unslopnemo-12b", @@ -3342,8 +3161,7 @@ "context_window": 32768, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/tongyi-deepresearch-30b-a3b", @@ -3361,8 +3179,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "upstage/solar-pro-3", @@ -3380,8 +3197,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xiaomi/mimo-v2-flash", @@ -3389,9 +3205,9 @@ "cost_per_1m_in": 0.1, "cost_per_1m_out": 0.3, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.1, + "cost_per_1m_out_cached": 0.01, "context_window": 262144, - "default_max_tokens": 131072, + "default_max_tokens": 32768, "can_reason": true, "reasoning_levels": [ "low", @@ -3399,8 +3215,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xiaomi/mimo-v2-omni", @@ -3418,8 +3233,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xiaomi/mimo-v2-pro", @@ -3437,8 +3251,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4-32b", @@ -3450,8 +3263,7 @@ "context_window": 128000, "default_max_tokens": 12800, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.5", @@ -3459,9 +3271,9 @@ "cost_per_1m_in": 0.6, "cost_per_1m_out": 2.2, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.11, + "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 49152, + "default_max_tokens": 13107, "can_reason": true, "reasoning_levels": [ "low", @@ -3469,8 +3281,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.5-air", @@ -3488,8 +3299,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.5-air:free", @@ -3507,8 +3317,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.5v", @@ -3526,8 +3335,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "z-ai/glm-4.6", @@ -3545,8 +3353,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.6v", @@ -3556,7 +3363,7 @@ "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0, "context_window": 131072, - "default_max_tokens": 13107, + "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ "low", @@ -3564,8 +3371,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "z-ai/glm-4.7", @@ -3583,8 +3389,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-4.7-flash", @@ -3602,17 +3407,16 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-5", "name": "Z.ai: GLM 5", - "cost_per_1m_in": 1, - "cost_per_1m_out": 3.2, + "cost_per_1m_in": 0.95, + "cost_per_1m_out": 2.55, "cost_per_1m_in_cached": 0, "cost_per_1m_out_cached": 0.2, - "context_window": 202800, + "context_window": 204800, "default_max_tokens": 65536, "can_reason": true, "reasoning_levels": [ @@ -3621,8 +3425,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "z-ai/glm-5-turbo", @@ -3640,34 +3443,31 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "x-ai/grok-3", "name": "xAI: Grok 3", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "x-ai/grok-3-beta", "name": "xAI: Grok 3 Beta", - "cost_per_1m_in": 3, - "cost_per_1m_out": 15, + "cost_per_1m_in": 5, + "cost_per_1m_out": 25, "cost_per_1m_in_cached": 0, - "cost_per_1m_out_cached": 0.75, + "cost_per_1m_out_cached": 1.25, "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "x-ai/grok-3-mini", @@ -3685,8 +3485,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "x-ai/grok-3-mini-beta", @@ -3704,8 +3503,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "x-ai/grok-4", @@ -3723,8 +3521,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "x-ai/grok-4-fast", @@ -3742,8 +3539,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "x-ai/grok-4.1-fast", @@ -3761,8 +3557,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "x-ai/grok-4.20", @@ -3780,8 +3575,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "x-ai/grok-code-fast-1", @@ -3799,8 +3593,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false } ], "default_headers": { diff --git a/internal/providers/configs/synthetic.json b/internal/providers/configs/synthetic.json index 3e35a5c186077ef24dd59e28ea4819f4a8e4e22b..1ab1f45ce09912f1b944626dfa6262f6dedae18a 100644 --- a/internal/providers/configs/synthetic.json +++ b/internal/providers/configs/synthetic.json @@ -23,8 +23,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:deepseek-ai/DeepSeek-V3", @@ -36,8 +35,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:deepseek-ai/DeepSeek-V3.2", @@ -55,8 +53,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:zai-org/GLM-4.7", @@ -74,8 +71,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:zai-org/GLM-4.7-Flash", @@ -93,8 +89,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:zai-org/GLM-5", @@ -112,8 +107,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:moonshotai/Kimi-K2-Instruct-0905", @@ -125,8 +119,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:moonshotai/Kimi-K2-Thinking", @@ -144,8 +137,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:moonshotai/Kimi-K2.5", @@ -163,8 +155,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "hf:nvidia/Kimi-K2.5-NVFP4", @@ -182,8 +173,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "hf:meta-llama/Llama-3.3-70B-Instruct", @@ -195,8 +185,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:MiniMaxAI/MiniMax-M2.1", @@ -214,8 +203,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:MiniMaxAI/MiniMax-M2.5", @@ -233,8 +221,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:nvidia/NVIDIA-Nemotron-3-Super-120B-A12B-NVFP4", @@ -252,8 +239,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:Qwen/Qwen3-235B-A22B-Thinking-2507", @@ -271,8 +257,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:Qwen/Qwen3-Coder-480B-A35B-Instruct", @@ -284,8 +269,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "hf:Qwen/Qwen3.5-397B-A17B", @@ -297,8 +281,7 @@ "context_window": 262144, "default_max_tokens": 26214, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "hf:openai/gpt-oss-120b", @@ -316,8 +299,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] } \ No newline at end of file diff --git a/internal/providers/configs/venice.json b/internal/providers/configs/venice.json index 7767ec2565e6639c998fec88893500d7a4c7df54..333f66220bad66d10a6584816985b163068de993 100644 --- a/internal/providers/configs/venice.json +++ b/internal/providers/configs/venice.json @@ -23,8 +23,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-opus-4-6", @@ -42,8 +41,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-5", @@ -61,8 +59,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "claude-sonnet-4-6", @@ -80,8 +77,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "deepseek-v3.2", @@ -93,8 +89,7 @@ "context_window": 160000, "default_max_tokens": 32768, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org-glm-4.6", @@ -106,8 +101,7 @@ "context_window": 198000, "default_max_tokens": 16384, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org-glm-4.7", @@ -119,8 +113,7 @@ "context_window": 198000, "default_max_tokens": 16384, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org-glm-4.7-flash", @@ -138,8 +131,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "olafangensan-glm-4.7-flash-heretic", @@ -151,8 +143,7 @@ "context_window": 200000, "default_max_tokens": 24000, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai-org-glm-5", @@ -164,8 +155,7 @@ "context_window": 198000, "default_max_tokens": 32000, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai-gpt-4o-2024-11-20", @@ -177,8 +167,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-4o-mini-2024-07-18", @@ -190,8 +179,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-52", @@ -209,8 +197,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai-gpt-52-codex", @@ -228,8 +215,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-53-codex", @@ -247,8 +233,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-54", @@ -266,8 +251,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-54-mini", @@ -285,8 +269,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai-gpt-54-pro", @@ -304,8 +287,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-flash-preview", @@ -323,8 +305,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "gemini-3-1-pro-preview", @@ -342,8 +323,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google-gemma-3-27b-it", @@ -355,8 +335,7 @@ "context_window": 198000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-41-fast", @@ -374,8 +353,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-20-beta", @@ -393,8 +371,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-code-fast-1", @@ -412,8 +389,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-thinking", @@ -431,8 +407,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kimi-k2-5", @@ -450,8 +425,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "llama-3.2-3b", @@ -463,8 +437,7 @@ "context_window": 128000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "llama-3.3-70b", @@ -476,8 +449,7 @@ "context_window": 128000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m21", @@ -495,8 +467,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m25", @@ -514,8 +485,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax-m27", @@ -533,8 +503,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral-small-3-2-24b-instruct", @@ -546,8 +515,7 @@ "context_window": 256000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "nvidia-nemotron-3-nano-30b-a3b", @@ -559,8 +527,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai-gpt-oss-120b", @@ -572,8 +539,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-235b-a22b-instruct-2507", @@ -585,8 +551,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-235b-a22b-thinking-2507", @@ -598,8 +563,7 @@ "context_window": 128000, "default_max_tokens": 16384, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-480b-a35b-instruct-turbo", @@ -611,8 +575,7 @@ "context_window": 256000, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-coder-480b-a35b-instruct", @@ -624,8 +587,7 @@ "context_window": 256000, "default_max_tokens": 65536, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-next-80b", @@ -637,8 +599,7 @@ "context_window": 256000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "qwen3-5-35b-a3b", @@ -666,8 +627,7 @@ "context_window": 256000, "default_max_tokens": 32768, "can_reason": true, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "qwen3-vl-235b-a22b", @@ -679,8 +639,7 @@ "context_window": 256000, "default_max_tokens": 16384, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "venice-uncensored-role-play", @@ -692,8 +651,7 @@ "context_window": 128000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true } ] } diff --git a/internal/providers/configs/vercel.json b/internal/providers/configs/vercel.json index 2c6a6f6e56a13e9858a476e531f762808a8e0bce..39af7a4d391a680604c6e11740d4e34f84f7611c 100644 --- a/internal/providers/configs/vercel.json +++ b/internal/providers/configs/vercel.json @@ -17,8 +17,7 @@ "context_window": 200000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.5-haiku", @@ -30,8 +29,7 @@ "context_window": 200000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.5-sonnet", @@ -43,8 +41,7 @@ "context_window": 200000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.5-sonnet-20240620", @@ -56,8 +53,7 @@ "context_window": 200000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-3.7-sonnet", @@ -78,8 +74,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-haiku-4.5", @@ -100,8 +95,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4", @@ -122,8 +116,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.1", @@ -144,8 +137,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.5", @@ -166,8 +158,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-opus-4.6", @@ -188,8 +179,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4", @@ -210,8 +200,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4.5", @@ -232,8 +221,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "anthropic/claude-sonnet-4.6", @@ -254,8 +242,7 @@ "xhigh" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "cohere/command-a", @@ -267,8 +254,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3", @@ -280,8 +266,7 @@ "context_window": 163840, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.1-terminus", @@ -299,8 +284,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.2", @@ -312,8 +296,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.2-thinking", @@ -331,8 +314,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-r1", @@ -350,8 +332,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "deepseek/deepseek-v3.1", @@ -369,8 +350,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/devstral-2", @@ -382,8 +362,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/devstral-small", @@ -395,8 +374,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/devstral-small-2", @@ -408,8 +386,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.5-air", @@ -427,8 +404,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.5v", @@ -440,8 +416,7 @@ "context_window": 66000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "zai/glm-4.6", @@ -459,8 +434,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.7", @@ -478,8 +452,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.7-flash", @@ -497,8 +470,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.7-flashx", @@ -516,8 +488,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-5", @@ -535,8 +506,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-5-turbo", @@ -554,8 +524,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.5", @@ -573,8 +542,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "zai/glm-4.6v", @@ -592,8 +560,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "zai/glm-4.6v-flash", @@ -611,8 +578,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-chat", @@ -630,8 +596,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex-max", @@ -649,8 +614,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex-mini", @@ -668,8 +632,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-thinking", @@ -687,8 +650,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2", @@ -706,8 +668,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-pro", @@ -725,8 +686,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-chat", @@ -744,8 +704,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.2-codex", @@ -763,8 +722,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.3-codex", @@ -782,8 +740,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4", @@ -801,8 +758,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-mini", @@ -820,8 +776,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-nano", @@ -839,8 +794,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.4-pro", @@ -858,8 +812,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4-turbo", @@ -871,8 +824,7 @@ "context_window": 128000, "default_max_tokens": 4096, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4.1", @@ -884,8 +836,7 @@ "context_window": 1047576, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4.1-mini", @@ -897,8 +848,7 @@ "context_window": 1047576, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4.1-nano", @@ -910,8 +860,7 @@ "context_window": 1047576, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o", @@ -923,8 +872,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-4o-mini", @@ -936,8 +884,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5", @@ -955,8 +902,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-mini", @@ -974,8 +920,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-nano", @@ -993,8 +938,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-pro", @@ -1012,8 +956,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5-codex", @@ -1031,8 +974,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-5.1-instant", @@ -1050,8 +992,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.1-codex", @@ -1069,8 +1010,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/gpt-5.3-chat", @@ -1088,8 +1028,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.0-flash", @@ -1101,8 +1040,7 @@ "context_window": 1048576, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.0-flash-lite", @@ -1114,8 +1052,7 @@ "context_window": 1048576, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-flash", @@ -1133,8 +1070,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-flash-lite", @@ -1152,8 +1088,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-2.5-pro", @@ -1171,8 +1106,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3-flash", @@ -1190,8 +1124,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3-pro-preview", @@ -1209,8 +1142,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3.1-flash-lite-preview", @@ -1228,8 +1160,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "google/gemini-3.1-pro-preview", @@ -1247,8 +1178,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-2-vision", @@ -1260,8 +1190,7 @@ "context_window": 32768, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-3", @@ -1273,8 +1202,7 @@ "context_window": 131072, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-3-fast", @@ -1286,8 +1214,7 @@ "context_window": 131072, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-3-mini", @@ -1299,8 +1226,7 @@ "context_window": 131072, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-3-mini-fast", @@ -1312,8 +1238,7 @@ "context_window": 131072, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4", @@ -1331,8 +1256,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-4-fast-non-reasoning", @@ -1344,8 +1268,7 @@ "context_window": 2000000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4-fast-reasoning", @@ -1363,8 +1286,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4.1-fast-non-reasoning", @@ -1376,8 +1298,7 @@ "context_window": 2000000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4.1-fast-reasoning", @@ -1395,8 +1316,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4.20-non-reasoning-beta", @@ -1408,8 +1328,7 @@ "context_window": 2000000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-4.20-reasoning-beta", @@ -1427,8 +1346,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-4.20-multi-agent-beta", @@ -1446,8 +1364,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4.20-multi-agent", @@ -1465,8 +1382,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xai/grok-4.20-non-reasoning", @@ -1478,8 +1394,7 @@ "context_window": 2000000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-4.20-reasoning", @@ -1497,8 +1412,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "xai/grok-code-fast-1", @@ -1516,8 +1430,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "prime-intellect/intellect-3", @@ -1535,8 +1448,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "kwaipilot/kat-coder-pro-v2", @@ -1554,8 +1466,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2", @@ -1567,8 +1478,7 @@ "context_window": 131072, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-0905", @@ -1580,8 +1490,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-thinking", @@ -1599,8 +1508,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-thinking-turbo", @@ -1618,8 +1526,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2-turbo", @@ -1631,8 +1538,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "moonshotai/kimi-k2.5", @@ -1650,8 +1556,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "meta/llama-3.1-70b", @@ -1663,8 +1568,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta/llama-3.1-8b", @@ -1676,8 +1580,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta/llama-3.2-11b", @@ -1689,8 +1592,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "meta/llama-3.2-90b", @@ -1702,8 +1604,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "meta/llama-3.3-70b", @@ -1715,8 +1616,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meta/llama-4-maverick", @@ -1728,8 +1628,7 @@ "context_window": 524288, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "meta/llama-4-scout", @@ -1741,8 +1640,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "meituan/longcat-flash-chat", @@ -1754,8 +1652,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "meituan/longcat-flash-thinking", @@ -1773,8 +1670,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "inception/mercury-2", @@ -1792,8 +1688,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "inception/mercury-coder-small", @@ -1805,8 +1700,7 @@ "context_window": 32000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xiaomi/mimo-v2-flash", @@ -1824,8 +1718,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "xiaomi/mimo-v2-pro", @@ -1843,8 +1736,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2", @@ -1862,8 +1754,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.1", @@ -1881,8 +1772,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.1-lightning", @@ -1900,8 +1790,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.5", @@ -1919,8 +1808,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.5-highspeed", @@ -1938,8 +1826,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "minimax/minimax-m2.7-highspeed", @@ -1957,8 +1844,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "minimax/minimax-m2.7", @@ -1976,8 +1862,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral/ministral-3b", @@ -1989,8 +1874,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/ministral-8b", @@ -2002,8 +1886,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/codestral", @@ -2015,8 +1898,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/mistral-medium", @@ -2028,8 +1910,7 @@ "context_window": 128000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral/mistral-small", @@ -2041,8 +1922,7 @@ "context_window": 32000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nvidia/nemotron-nano-12b-v2-vl", @@ -2060,8 +1940,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "nvidia/nemotron-nano-9b-v2", @@ -2079,8 +1958,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "mistral/pixtral-12b", @@ -2092,8 +1970,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "mistral/pixtral-large", @@ -2105,8 +1982,7 @@ "context_window": 128000, "default_max_tokens": 4000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "alibaba/qwen-3-32b", @@ -2124,8 +2000,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-coder-30b-a3b", @@ -2143,8 +2018,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-max-thinking", @@ -2162,8 +2036,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3.5-flash", @@ -2181,8 +2054,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "alibaba/qwen3.5-plus", @@ -2200,8 +2072,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "alibaba/qwen3-235b-a22b-thinking", @@ -2219,8 +2090,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "alibaba/qwen3-coder", @@ -2232,8 +2102,7 @@ "context_window": 262144, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-coder-next", @@ -2245,8 +2114,7 @@ "context_window": 256000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-coder-plus", @@ -2258,8 +2126,7 @@ "context_window": 1000000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-max", @@ -2271,8 +2138,7 @@ "context_window": 262144, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-max-preview", @@ -2284,8 +2150,7 @@ "context_window": 262144, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen3-vl-thinking", @@ -2303,8 +2168,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "alibaba/qwen-3-14b", @@ -2322,8 +2186,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen-3-235b", @@ -2335,8 +2198,7 @@ "context_window": 32768, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "alibaba/qwen-3-30b", @@ -2354,8 +2216,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "bytedance/seed-1.6", @@ -2373,8 +2234,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "perplexity/sonar", @@ -2386,8 +2246,7 @@ "context_window": 127000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "perplexity/sonar-pro", @@ -2399,8 +2258,7 @@ "context_window": 200000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "arcee-ai/trinity-large-preview", @@ -2412,8 +2270,7 @@ "context_window": 131000, "default_max_tokens": 8000, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-120b", @@ -2431,8 +2288,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-20b", @@ -2450,8 +2306,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/gpt-oss-safeguard-20b", @@ -2469,8 +2324,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/o1", @@ -2488,8 +2342,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3", @@ -2507,8 +2360,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3-pro", @@ -2526,8 +2378,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3-deep-research", @@ -2545,8 +2396,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "openai/o3-mini", @@ -2564,8 +2414,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "openai/o4-mini", @@ -2583,8 +2432,7 @@ "high" ], "default_reasoning_effort": "medium", - "supports_attachments": true, - "options": {} + "supports_attachments": true } ], "default_headers": { diff --git a/internal/providers/configs/xai.json b/internal/providers/configs/xai.json index 5e4144ff8c0a6cc1f3517e909d6551366bd3877c..1ca4e56e58557614b5149707849e36aca5b9972b 100644 --- a/internal/providers/configs/xai.json +++ b/internal/providers/configs/xai.json @@ -17,8 +17,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": false, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "grok-3-mini", @@ -30,8 +29,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false }, { "id": "grok-4", @@ -43,8 +41,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": true, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-1-fast", @@ -56,8 +53,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": true, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-1-fast-non-reasoning", @@ -69,8 +65,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-fast", @@ -82,8 +77,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": true, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4-fast-non-reasoning", @@ -95,8 +89,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4.20", @@ -108,8 +101,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": true, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-4.20-non-reasoning", @@ -121,8 +113,7 @@ "context_window": 200000, "default_max_tokens": 20000, "can_reason": false, - "supports_attachments": true, - "options": {} + "supports_attachments": true }, { "id": "grok-code-fast", @@ -134,8 +125,7 @@ "context_window": 131072, "default_max_tokens": 13107, "can_reason": true, - "supports_attachments": false, - "options": {} + "supports_attachments": false } ] }