@@ -601,6 +601,25 @@
"supports_attachments": false,
"options": {}
},
+ {
+ "id": "coding-minimax-m2.5-highspeed",
+ "name": "Coding MiniMax M2.5 Highspeed",
+ "cost_per_1m_in": 0.2,
+ "cost_per_1m_out": 0.2,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 204800,
+ "default_max_tokens": 13100,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": false,
+ "options": {}
+ },
{
"id": "deepseek-math-v2",
"name": "DeepSeek Math V2",
@@ -1760,6 +1779,25 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "gpt-5.3-codex",
+ "name": "GPT-5.3-Codex",
+ "cost_per_1m_in": 1.75,
+ "cost_per_1m_out": 14,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0.175,
+ "context_window": 400000,
+ "default_max_tokens": 128000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "gpt-oss-120b",
"name": "gpt-oss-120b",
@@ -2401,6 +2439,63 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "qwen3.5-122b-a10b",
+ "name": "Qwen3.5 122B A10B",
+ "cost_per_1m_in": 0.1126,
+ "cost_per_1m_out": 0.9008,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0.1126,
+ "context_window": 991000,
+ "default_max_tokens": 64000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "qwen3.5-27b",
+ "name": "Qwen3.5 27B",
+ "cost_per_1m_in": 0.0846,
+ "cost_per_1m_out": 0.6768,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0.0846,
+ "context_window": 991000,
+ "default_max_tokens": 64000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "qwen3.5-35b-a3b",
+ "name": "Qwen3.5 35B A3B",
+ "cost_per_1m_in": 0.0564,
+ "cost_per_1m_out": 0.4512,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0.0564,
+ "context_window": 991000,
+ "default_max_tokens": 64000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "qwen3.5-397b-a17b",
"name": "Qwen3.5 397B A17B",
@@ -2420,13 +2515,32 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "qwen3.5-flash",
+ "name": "Qwen3.5 Flash",
+ "cost_per_1m_in": 0.0282,
+ "cost_per_1m_out": 0.282,
+ "cost_per_1m_in_cached": 0.03525,
+ "cost_per_1m_out_cached": 0.00282,
+ "context_window": 991000,
+ "default_max_tokens": 64000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "qwen3.5-plus",
"name": "Qwen3.5 Plus",
- "cost_per_1m_in": 0.137,
- "cost_per_1m_out": 1.37,
- "cost_per_1m_in_cached": 0.17125,
- "cost_per_1m_out_cached": 0.0137,
+ "cost_per_1m_in": 0.1096,
+ "cost_per_1m_out": 0.6576,
+ "cost_per_1m_in_cached": 0.137,
+ "cost_per_1m_out_cached": 0.01096,
"context_window": 991000,
"default_max_tokens": 64000,
"can_reason": true,
@@ -528,12 +528,12 @@
{
"id": "deepseek/deepseek-chat-v3.1",
"name": "DeepSeek: DeepSeek V3.1",
- "cost_per_1m_in": 0.19999999999999998,
- "cost_per_1m_out": 0.7999999999999999,
+ "cost_per_1m_in": 0.21,
+ "cost_per_1m_out": 0.7899999999999999,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.09999999999999999,
+ "cost_per_1m_out_cached": 0.1300000002,
"context_window": 163840,
- "default_max_tokens": 32768,
+ "default_max_tokens": 16384,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -590,7 +590,7 @@
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0.13,
"context_window": 163840,
- "default_max_tokens": 32768,
+ "default_max_tokens": 16384,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -642,12 +642,12 @@
{
"id": "deepseek/deepseek-r1-0528",
"name": "DeepSeek: R1 0528",
- "cost_per_1m_in": 0.7,
- "cost_per_1m_out": 2.5,
+ "cost_per_1m_in": 0.39999999999999997,
+ "cost_per_1m_out": 1.75,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.35,
+ "cost_per_1m_out_cached": 0.19999999999999998,
"context_window": 163840,
- "default_max_tokens": 16384,
+ "default_max_tokens": 32768,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -855,6 +855,38 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "google/gemini-3.1-pro-preview-customtools",
+ "name": "Google: Gemini 3.1 Pro Preview Custom Tools",
+ "cost_per_1m_in": 2,
+ "cost_per_1m_out": 12,
+ "cost_per_1m_in_cached": 0.375,
+ "cost_per_1m_out_cached": 0.19999999999999998,
+ "context_window": 1048576,
+ "default_max_tokens": 32768,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "google/gemma-3-27b-it:free",
+ "name": "Google: Gemma 3 27B (free)",
+ "cost_per_1m_in": 0,
+ "cost_per_1m_out": 0,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 128000,
+ "default_max_tokens": 64000,
+ "can_reason": false,
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "inception/mercury",
"name": "Inception: Mercury",
@@ -894,6 +926,19 @@
"supports_attachments": false,
"options": {}
},
+ {
+ "id": "meituan/longcat-flash-chat",
+ "name": "Meituan: LongCat Flash Chat",
+ "cost_per_1m_in": 0.19999999999999998,
+ "cost_per_1m_out": 0.7999999999999999,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0.19999999999999998,
+ "context_window": 131072,
+ "default_max_tokens": 16384,
+ "can_reason": false,
+ "supports_attachments": false,
+ "options": {}
+ },
{
"id": "meta-llama/llama-3.1-405b-instruct",
"name": "Meta: Llama 3.1 405B Instruct",
@@ -910,12 +955,12 @@
{
"id": "meta-llama/llama-3.1-70b-instruct",
"name": "Meta: Llama 3.1 70B Instruct",
- "cost_per_1m_in": 0.88,
- "cost_per_1m_out": 0.88,
+ "cost_per_1m_in": 0.39999999999999997,
+ "cost_per_1m_out": 0.39999999999999997,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
- "default_max_tokens": 13107,
+ "default_max_tokens": 8192,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -940,8 +985,8 @@
"cost_per_1m_out": 0,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 65536,
- "default_max_tokens": 6553,
+ "context_window": 128000,
+ "default_max_tokens": 64000,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -949,12 +994,12 @@
{
"id": "meta-llama/llama-4-maverick",
"name": "Meta: Llama 4 Maverick",
- "cost_per_1m_in": 0.27,
- "cost_per_1m_out": 0.85,
+ "cost_per_1m_in": 0.19999999999999998,
+ "cost_per_1m_out": 0.6,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 1048576,
- "default_max_tokens": 104857,
+ "context_window": 131072,
+ "default_max_tokens": 4096,
"can_reason": false,
"supports_attachments": true,
"options": {}
@@ -972,25 +1017,6 @@
"supports_attachments": true,
"options": {}
},
- {
- "id": "minimax/minimax-m1",
- "name": "MiniMax: MiniMax M1",
- "cost_per_1m_in": 0.44,
- "cost_per_1m_out": 1.76,
- "cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 1000000,
- "default_max_tokens": 20000,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": false,
- "options": {}
- },
{
"id": "minimax/minimax-m2",
"name": "MiniMax: MiniMax M2",
@@ -1230,6 +1256,19 @@
"supports_attachments": false,
"options": {}
},
+ {
+ "id": "mistralai/mistral-small-24b-instruct-2501",
+ "name": "Mistral: Mistral Small 3",
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.3,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 32768,
+ "default_max_tokens": 1024,
+ "can_reason": false,
+ "supports_attachments": false,
+ "options": {}
+ },
{
"id": "mistralai/mistral-small-3.1-24b-instruct:free",
"name": "Mistral: Mistral Small 3.1 24B (free)",
@@ -1350,12 +1389,12 @@
{
"id": "moonshotai/kimi-k2-0905",
"name": "MoonshotAI: Kimi K2 0905",
- "cost_per_1m_in": 1,
- "cost_per_1m_out": 3,
+ "cost_per_1m_in": 0.6,
+ "cost_per_1m_out": 2.5,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.5,
+ "cost_per_1m_out_cached": 0.3,
"context_window": 262144,
- "default_max_tokens": 8192,
+ "default_max_tokens": 26214,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -1376,8 +1415,8 @@
{
"id": "moonshotai/kimi-k2-thinking",
"name": "MoonshotAI: Kimi K2 Thinking",
- "cost_per_1m_in": 0.6,
- "cost_per_1m_out": 2.5,
+ "cost_per_1m_in": 1.15,
+ "cost_per_1m_out": 8,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0.15,
"context_window": 262144,
@@ -1395,12 +1434,12 @@
{
"id": "moonshotai/kimi-k2.5",
"name": "MoonshotAI: Kimi K2.5",
- "cost_per_1m_in": 0.6,
- "cost_per_1m_out": 3,
+ "cost_per_1m_in": 0.5700000000000001,
+ "cost_per_1m_out": 2.8499999999999996,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.09999999999999999,
+ "cost_per_1m_out_cached": 0.095,
"context_window": 262144,
- "default_max_tokens": 26214,
+ "default_max_tokens": 131072,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -1503,8 +1542,8 @@
{
"id": "nvidia/nemotron-nano-9b-v2",
"name": "NVIDIA: Nemotron Nano 9B V2",
- "cost_per_1m_in": 0.06,
- "cost_per_1m_out": 0.25,
+ "cost_per_1m_in": 0.04,
+ "cost_per_1m_out": 0.16,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
@@ -1872,7 +1911,7 @@
"cost_per_1m_in": 1.25,
"cost_per_1m_out": 10,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.125,
+ "cost_per_1m_out_cached": 0.13,
"context_window": 400000,
"default_max_tokens": 64000,
"can_reason": true,
@@ -1982,7 +2021,7 @@
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0.175,
"context_window": 128000,
- "default_max_tokens": 16000,
+ "default_max_tokens": 8192,
"can_reason": false,
"supports_attachments": true,
"options": {}
@@ -2343,25 +2382,6 @@
"supports_attachments": false,
"options": {}
},
- {
- "id": "qwen/qwq-32b",
- "name": "Qwen: QwQ 32B",
- "cost_per_1m_in": 0.15,
- "cost_per_1m_out": 0.58,
- "cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 131072,
- "default_max_tokens": 65536,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": false,
- "options": {}
- },
{
"id": "qwen/qwen-plus-2025-07-28",
"name": "Qwen: Qwen Plus 0728",
@@ -2500,12 +2520,12 @@
{
"id": "qwen/qwen3-235b-a22b-2507",
"name": "Qwen: Qwen3 235B A22B Instruct 2507",
- "cost_per_1m_in": 0.08,
- "cost_per_1m_out": 0.55,
+ "cost_per_1m_in": 0.19999999999999998,
+ "cost_per_1m_out": 0.7999999999999999,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.04,
+ "cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 32768,
+ "default_max_tokens": 131072,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -2513,12 +2533,12 @@
{
"id": "qwen/qwen3-235b-a22b-thinking-2507",
"name": "Qwen: Qwen3 235B A22B Thinking 2507",
- "cost_per_1m_in": 0.11,
- "cost_per_1m_out": 0.6,
+ "cost_per_1m_in": 0.65,
+ "cost_per_1m_out": 3,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.055,
+ "cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 131072,
+ "default_max_tokens": 26214,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -2532,12 +2552,12 @@
{
"id": "qwen/qwen3-30b-a3b",
"name": "Qwen: Qwen3 30B A3B",
- "cost_per_1m_in": 0.15,
- "cost_per_1m_out": 0.6,
+ "cost_per_1m_in": 0.13,
+ "cost_per_1m_out": 0.52,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
- "default_max_tokens": 65536,
+ "default_max_tokens": 4096,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -2564,12 +2584,12 @@
{
"id": "qwen/qwen3-30b-a3b-thinking-2507",
"name": "Qwen: Qwen3 30B A3B Thinking 2507",
- "cost_per_1m_in": 0.09,
+ "cost_per_1m_in": 0.09999999999999999,
"cost_per_1m_out": 0.3,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 65536,
+ "default_max_tokens": 26214,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -2653,12 +2673,12 @@
{
"id": "qwen/qwen3-coder",
"name": "Qwen: Qwen3 Coder 480B A35B",
- "cost_per_1m_in": 0.39999999999999997,
+ "cost_per_1m_in": 0.22,
"cost_per_1m_out": 1.7999999999999998,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 131072,
+ "default_max_tokens": 32768,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -2705,12 +2725,12 @@
{
"id": "qwen/qwen3-coder-next",
"name": "Qwen: Qwen3 Coder Next",
- "cost_per_1m_in": 0.5,
- "cost_per_1m_out": 1.2,
+ "cost_per_1m_in": 0.12,
+ "cost_per_1m_out": 0.75,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
+ "cost_per_1m_out_cached": 0.06,
"context_window": 262144,
- "default_max_tokens": 26214,
+ "default_max_tokens": 32768,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -2763,12 +2783,12 @@
{
"id": "qwen/qwen3-next-80b-a3b-instruct",
"name": "Qwen: Qwen3 Next 80B A3B Instruct",
- "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_in": 0.09,
"cost_per_1m_out": 1.1,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 131072,
+ "default_max_tokens": 26214,
"can_reason": false,
"supports_attachments": false,
"options": {}
@@ -2808,8 +2828,8 @@
{
"id": "qwen/qwen3-vl-235b-a22b-instruct",
"name": "Qwen: Qwen3 VL 235B A22B Instruct",
- "cost_per_1m_in": 0.19999999999999998,
- "cost_per_1m_out": 1.2,
+ "cost_per_1m_in": 0.26,
+ "cost_per_1m_out": 1.04,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
@@ -2840,12 +2860,12 @@
{
"id": "qwen/qwen3-vl-30b-a3b-instruct",
"name": "Qwen: Qwen3 VL 30B A3B Instruct",
- "cost_per_1m_in": 0.29,
- "cost_per_1m_out": 1,
+ "cost_per_1m_in": 0.13,
+ "cost_per_1m_out": 0.52,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 262144,
- "default_max_tokens": 131072,
+ "context_window": 131072,
+ "default_max_tokens": 16384,
"can_reason": false,
"supports_attachments": true,
"options": {}
@@ -2853,12 +2873,12 @@
{
"id": "qwen/qwen3-vl-30b-a3b-thinking",
"name": "Qwen: Qwen3 VL 30B A3B Thinking",
- "cost_per_1m_in": 0.29,
- "cost_per_1m_out": 1,
+ "cost_per_1m_in": 0,
+ "cost_per_1m_out": 0,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 262144,
- "default_max_tokens": 131072,
+ "context_window": 131072,
+ "default_max_tokens": 16384,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -2917,12 +2937,12 @@
{
"id": "qwen/qwen3.5-397b-a17b",
"name": "Qwen: Qwen3.5 397B A17B",
- "cost_per_1m_in": 0.55,
- "cost_per_1m_out": 3.5,
+ "cost_per_1m_in": 0.6,
+ "cost_per_1m_out": 3.5999999999999996,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.55,
+ "cost_per_1m_out_cached": 0,
"context_window": 262144,
- "default_max_tokens": 32768,
+ "default_max_tokens": 131072,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -2952,6 +2972,82 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "qwen/qwen3.5-122b-a10b",
+ "name": "Qwen: Qwen3.5-122B-A10B",
+ "cost_per_1m_in": 0.39999999999999997,
+ "cost_per_1m_out": 3.1999999999999997,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 262144,
+ "default_max_tokens": 32768,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "qwen/qwen3.5-27b",
+ "name": "Qwen: Qwen3.5-27B",
+ "cost_per_1m_in": 0.3,
+ "cost_per_1m_out": 2.4,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 262144,
+ "default_max_tokens": 32768,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "qwen/qwen3.5-35b-a3b",
+ "name": "Qwen: Qwen3.5-35B-A3B",
+ "cost_per_1m_in": 0.25,
+ "cost_per_1m_out": 2,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 262144,
+ "default_max_tokens": 32768,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "qwen/qwen3.5-flash-02-23",
+ "name": "Qwen: Qwen3.5-Flash",
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.39999999999999997,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 1000000,
+ "default_max_tokens": 32768,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "relace/relace-search",
"name": "Relace: Relace Search",
@@ -3022,19 +3118,6 @@
"supports_attachments": false,
"options": {}
},
- {
- "id": "thedrummer/rocinante-12b",
- "name": "TheDrummer: Rocinante 12B",
- "cost_per_1m_in": 0.16999999999999998,
- "cost_per_1m_out": 0.43,
- "cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 32768,
- "default_max_tokens": 16384,
- "can_reason": false,
- "supports_attachments": false,
- "options": {}
- },
{
"id": "thedrummer/unslopnemo-12b",
"name": "TheDrummer: UnslopNemo 12B",
@@ -3126,7 +3209,7 @@
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0.11,
"context_window": 131072,
- "default_max_tokens": 49152,
+ "default_max_tokens": 48000,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -3216,12 +3299,12 @@
{
"id": "z-ai/glm-4.6:exacto",
"name": "Z.ai: GLM 4.6 (exacto)",
- "cost_per_1m_in": 0.44,
- "cost_per_1m_out": 1.76,
+ "cost_per_1m_in": 0.6,
+ "cost_per_1m_out": 2.2,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0.11,
- "context_window": 204800,
- "default_max_tokens": 65536,
+ "context_window": 200000,
+ "default_max_tokens": 64000,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -3238,9 +3321,9 @@
"cost_per_1m_in": 0.3,
"cost_per_1m_out": 0.8999999999999999,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.15,
+ "cost_per_1m_out_cached": 0.049999999999999996,
"context_window": 131072,
- "default_max_tokens": 32768,
+ "default_max_tokens": 12000,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -3254,12 +3337,12 @@
{
"id": "z-ai/glm-4.7",
"name": "Z.ai: GLM 4.7",
- "cost_per_1m_in": 0.44999999999999996,
- "cost_per_1m_out": 2.2,
+ "cost_per_1m_in": 0.54,
+ "cost_per_1m_out": 1.9800000000000002,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.11,
+ "cost_per_1m_out_cached": 0.099,
"context_window": 204800,
- "default_max_tokens": 102400,
+ "default_max_tokens": 65536,
"can_reason": true,
"reasoning_levels": [
"low",
@@ -3273,10 +3356,10 @@
{
"id": "z-ai/glm-4.7-flash",
"name": "Z.ai: GLM 4.7 Flash",
- "cost_per_1m_in": 0.06,
- "cost_per_1m_out": 0.39999999999999997,
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.43,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.0100000002,
+ "cost_per_1m_out_cached": 0,
"context_window": 202752,
"default_max_tokens": 20275,
"can_reason": true,
@@ -582,6 +582,25 @@
"supports_attachments": false,
"options": {}
},
+ {
+ "id": "openai/gpt-5-chat",
+ "name": "GPT 5 Chat",
+ "cost_per_1m_in": 1.25,
+ "cost_per_1m_out": 10,
+ "cost_per_1m_in_cached": 0.125,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 128000,
+ "default_max_tokens": 8000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "openai/gpt-5.1-codex-max",
"name": "GPT 5.1 Codex Max",
@@ -601,6 +620,25 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "openai/gpt-5.1-codex-mini",
+ "name": "GPT 5.1 Codex Mini",
+ "cost_per_1m_in": 0.25,
+ "cost_per_1m_out": 2,
+ "cost_per_1m_in_cached": 0.024999999999999998,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 400000,
+ "default_max_tokens": 8000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "openai/gpt-5.1-thinking",
"name": "GPT 5.1 Thinking",
@@ -658,6 +696,44 @@
"supports_attachments": true,
"options": {}
},
+ {
+ "id": "openai/gpt-5.2-chat",
+ "name": "GPT 5.2 Chat",
+ "cost_per_1m_in": 1.75,
+ "cost_per_1m_out": 14,
+ "cost_per_1m_in_cached": 0.175,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 128000,
+ "default_max_tokens": 8000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
+ {
+ "id": "openai/gpt-5.2-codex",
+ "name": "GPT 5.2 Codex",
+ "cost_per_1m_in": 1.75,
+ "cost_per_1m_out": 14,
+ "cost_per_1m_in_cached": 0.175,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 400000,
+ "default_max_tokens": 8000,
+ "can_reason": true,
+ "reasoning_levels": [
+ "low",
+ "medium",
+ "high"
+ ],
+ "default_reasoning_effort": "medium",
+ "supports_attachments": true,
+ "options": {}
+ },
{
"id": "openai/gpt-5.3-codex",
"name": "GPT 5.3 Codex",
@@ -774,25 +850,6 @@
"supports_attachments": true,
"options": {}
},
- {
- "id": "openai/gpt-5-chat",
- "name": "GPT-5 Chat",
- "cost_per_1m_in": 1.25,
- "cost_per_1m_out": 10,
- "cost_per_1m_in_cached": 0.125,
- "cost_per_1m_out_cached": 0,
- "context_window": 128000,
- "default_max_tokens": 8000,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": true,
- "options": {}
- },
{
"id": "openai/gpt-5-mini",
"name": "GPT-5 mini",
@@ -869,25 +926,6 @@
"supports_attachments": true,
"options": {}
},
- {
- "id": "openai/gpt-5.1-codex-mini",
- "name": "GPT-5.1 Codex mini",
- "cost_per_1m_in": 0.25,
- "cost_per_1m_out": 2,
- "cost_per_1m_in_cached": 0.024999999999999998,
- "cost_per_1m_out_cached": 0,
- "context_window": 400000,
- "default_max_tokens": 8000,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": true,
- "options": {}
- },
{
"id": "openai/gpt-5.1-instant",
"name": "GPT-5.1 Instant",
@@ -926,44 +964,6 @@
"supports_attachments": true,
"options": {}
},
- {
- "id": "openai/gpt-5.2-chat",
- "name": "GPT-5.2 Chat",
- "cost_per_1m_in": 1.75,
- "cost_per_1m_out": 14,
- "cost_per_1m_in_cached": 0.175,
- "cost_per_1m_out_cached": 0,
- "context_window": 128000,
- "default_max_tokens": 8000,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": true,
- "options": {}
- },
- {
- "id": "openai/gpt-5.2-codex",
- "name": "GPT-5.2-Codex",
- "cost_per_1m_in": 1.75,
- "cost_per_1m_out": 14,
- "cost_per_1m_in_cached": 0.175,
- "cost_per_1m_out_cached": 0,
- "context_window": 400000,
- "default_max_tokens": 8000,
- "can_reason": true,
- "reasoning_levels": [
- "low",
- "medium",
- "high"
- ],
- "default_reasoning_effort": "medium",
- "supports_attachments": true,
- "options": {}
- },
{
"id": "google/gemini-2.5-flash",
"name": "Gemini 2.5 Flash",