@@ -29,7 +29,7 @@
"cost_per_1m_out_cached": 0.024999999999999998,
"context_window": 400000,
"default_max_tokens": 64000,
- "can_reason": true,
+ "can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": true
},
@@ -42,7 +42,7 @@
"cost_per_1m_out_cached": 0.005,
"context_window": 400000,
"default_max_tokens": 64000,
- "can_reason": true,
+ "can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": true
},
@@ -104,9 +104,9 @@
"cost_per_1m_in": 0.6,
"cost_per_1m_out": 2.2,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.11,
- "context_window": 128000,
- "default_max_tokens": 48000,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 131072,
+ "default_max_tokens": 65536,
"can_reason": true,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -117,9 +117,9 @@
"cost_per_1m_in": 0.19999999999999998,
"cost_per_1m_out": 1.1,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 131072,
- "default_max_tokens": 13107,
+ "cost_per_1m_out_cached": 0.03,
+ "context_window": 128000,
+ "default_max_tokens": 48000,
"can_reason": true,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -208,7 +208,7 @@
"cost_per_1m_in": 0.6,
"cost_per_1m_out": 2.5,
"cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0.15,
+ "cost_per_1m_out_cached": 0,
"context_window": 131072,
"default_max_tokens": 13107,
"can_reason": false,
@@ -228,19 +228,6 @@
"has_reasoning_efforts": false,
"supports_attachments": false
},
- {
- "id": "mistralai/devstral-small",
- "name": "Mistral: Devstral Small 1.1",
- "cost_per_1m_in": 0.09999999999999999,
- "cost_per_1m_out": 0.3,
- "cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 131072,
- "default_max_tokens": 13107,
- "can_reason": false,
- "has_reasoning_efforts": false,
- "supports_attachments": false
- },
{
"id": "x-ai/grok-4",
"name": "xAI: Grok 4",
@@ -569,12 +556,12 @@
{
"id": "qwen/qwen3-30b-a3b",
"name": "Qwen: Qwen3 30B A3B",
- "cost_per_1m_in": 0.08,
- "cost_per_1m_out": 0.29,
+ "cost_per_1m_in": 0.15,
+ "cost_per_1m_out": 0.6,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 40960,
- "default_max_tokens": 20480,
+ "context_window": 131072,
+ "default_max_tokens": 4000,
"can_reason": true,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -816,12 +803,12 @@
{
"id": "mistralai/mistral-small-3.1-24b-instruct",
"name": "Mistral: Mistral Small 3.1 24B",
- "cost_per_1m_in": 0.017992691999999998,
- "cost_per_1m_out": 0.07200576,
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.3,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
- "default_max_tokens": 48000,
+ "default_max_tokens": 13107,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": true
@@ -847,7 +834,7 @@
"cost_per_1m_in_cached": 3.75,
"cost_per_1m_out_cached": 0.3,
"context_window": 200000,
- "default_max_tokens": 64000,
+ "default_max_tokens": 32000,
"can_reason": true,
"has_reasoning_efforts": false,
"supports_attachments": true
@@ -972,12 +959,12 @@
{
"id": "mistralai/mistral-small-24b-instruct-2501",
"name": "Mistral: Mistral Small 3",
- "cost_per_1m_in": 0.049999999999999996,
- "cost_per_1m_out": 0.09,
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.3,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 28000,
- "default_max_tokens": 14000,
+ "context_window": 32768,
+ "default_max_tokens": 3276,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -1089,12 +1076,12 @@
{
"id": "meta-llama/llama-3.3-70b-instruct",
"name": "Meta: Llama 3.3 70B Instruct",
- "cost_per_1m_in": 0.039,
- "cost_per_1m_out": 0.12,
+ "cost_per_1m_in": 0.12,
+ "cost_per_1m_out": 0.3,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
- "default_max_tokens": 4096,
+ "default_max_tokens": 65536,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -1204,8 +1191,8 @@
"supports_attachments": false
},
{
- "id": "anthropic/claude-3.5-haiku-20241022",
- "name": "Anthropic: Claude 3.5 Haiku (2024-10-22)",
+ "id": "anthropic/claude-3.5-haiku:beta",
+ "name": "Anthropic: Claude 3.5 Haiku (self-moderated)",
"cost_per_1m_in": 0.7999999999999999,
"cost_per_1m_out": 4,
"cost_per_1m_in_cached": 1,
@@ -1217,8 +1204,8 @@
"supports_attachments": true
},
{
- "id": "anthropic/claude-3.5-haiku:beta",
- "name": "Anthropic: Claude 3.5 Haiku (self-moderated)",
+ "id": "anthropic/claude-3.5-haiku",
+ "name": "Anthropic: Claude 3.5 Haiku",
"cost_per_1m_in": 0.7999999999999999,
"cost_per_1m_out": 4,
"cost_per_1m_in_cached": 1,
@@ -1230,8 +1217,8 @@
"supports_attachments": true
},
{
- "id": "anthropic/claude-3.5-haiku",
- "name": "Anthropic: Claude 3.5 Haiku",
+ "id": "anthropic/claude-3.5-haiku-20241022",
+ "name": "Anthropic: Claude 3.5 Haiku (2024-10-22)",
"cost_per_1m_in": 0.7999999999999999,
"cost_per_1m_out": 4,
"cost_per_1m_in_cached": 1,
@@ -1399,27 +1386,27 @@
"supports_attachments": true
},
{
- "id": "meta-llama/llama-3.1-405b-instruct",
- "name": "Meta: Llama 3.1 405B Instruct",
- "cost_per_1m_in": 0.7999999999999999,
- "cost_per_1m_out": 0.7999999999999999,
+ "id": "meta-llama/llama-3.1-8b-instruct",
+ "name": "Meta: Llama 3.1 8B Instruct",
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.09999999999999999,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 32768,
- "default_max_tokens": 8192,
+ "context_window": 131072,
+ "default_max_tokens": 4000,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
},
{
- "id": "meta-llama/llama-3.1-8b-instruct",
- "name": "Meta: Llama 3.1 8B Instruct",
- "cost_per_1m_in": 0.09999999999999999,
- "cost_per_1m_out": 0.09999999999999999,
+ "id": "meta-llama/llama-3.1-405b-instruct",
+ "name": "Meta: Llama 3.1 405B Instruct",
+ "cost_per_1m_in": 0.7999999999999999,
+ "cost_per_1m_out": 0.7999999999999999,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 131072,
- "default_max_tokens": 4000,
+ "context_window": 32768,
+ "default_max_tokens": 8192,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -1427,12 +1414,12 @@
{
"id": "meta-llama/llama-3.1-70b-instruct",
"name": "Meta: Llama 3.1 70B Instruct",
- "cost_per_1m_in": 0.8999999999999999,
- "cost_per_1m_out": 0.8999999999999999,
+ "cost_per_1m_in": 0.09999999999999999,
+ "cost_per_1m_out": 0.28,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 131072,
- "default_max_tokens": 13107,
+ "default_max_tokens": 8192,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
@@ -1503,10 +1490,10 @@
"supports_attachments": true
},
{
- "id": "mistralai/mistral-7b-instruct-v0.3",
- "name": "Mistral: Mistral 7B Instruct v0.3",
- "cost_per_1m_in": 0.028,
- "cost_per_1m_out": 0.054,
+ "id": "mistralai/mistral-7b-instruct:free",
+ "name": "Mistral: Mistral 7B Instruct (free)",
+ "cost_per_1m_in": 0,
+ "cost_per_1m_out": 0,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 32768,
@@ -1516,10 +1503,10 @@
"supports_attachments": false
},
{
- "id": "mistralai/mistral-7b-instruct:free",
- "name": "Mistral: Mistral 7B Instruct (free)",
- "cost_per_1m_in": 0,
- "cost_per_1m_out": 0,
+ "id": "mistralai/mistral-7b-instruct",
+ "name": "Mistral: Mistral 7B Instruct",
+ "cost_per_1m_in": 0.028,
+ "cost_per_1m_out": 0.054,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
"context_window": 32768,
@@ -1529,8 +1516,8 @@
"supports_attachments": false
},
{
- "id": "mistralai/mistral-7b-instruct",
- "name": "Mistral: Mistral 7B Instruct",
+ "id": "mistralai/mistral-7b-instruct-v0.3",
+ "name": "Mistral: Mistral 7B Instruct v0.3",
"cost_per_1m_in": 0.028,
"cost_per_1m_out": 0.054,
"cost_per_1m_in_cached": 0,
@@ -1580,19 +1567,6 @@
"has_reasoning_efforts": false,
"supports_attachments": true
},
- {
- "id": "openai/gpt-4o-2024-05-13",
- "name": "OpenAI: GPT-4o (2024-05-13)",
- "cost_per_1m_in": 5,
- "cost_per_1m_out": 15,
- "cost_per_1m_in_cached": 0,
- "cost_per_1m_out_cached": 0,
- "context_window": 128000,
- "default_max_tokens": 2048,
- "can_reason": false,
- "has_reasoning_efforts": false,
- "supports_attachments": true
- },
{
"id": "openai/gpt-4o",
"name": "OpenAI: GPT-4o",
@@ -1620,17 +1594,17 @@
"supports_attachments": true
},
{
- "id": "meta-llama/llama-3-70b-instruct",
- "name": "Meta: Llama 3 70B Instruct",
- "cost_per_1m_in": 0.3,
- "cost_per_1m_out": 0.39999999999999997,
+ "id": "openai/gpt-4o-2024-05-13",
+ "name": "OpenAI: GPT-4o (2024-05-13)",
+ "cost_per_1m_in": 5,
+ "cost_per_1m_out": 15,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 8192,
- "default_max_tokens": 8192,
+ "context_window": 128000,
+ "default_max_tokens": 2048,
"can_reason": false,
"has_reasoning_efforts": false,
- "supports_attachments": false
+ "supports_attachments": true
},
{
"id": "meta-llama/llama-3-8b-instruct",
@@ -1646,44 +1620,57 @@
"supports_attachments": false
},
{
- "id": "mistralai/mixtral-8x22b-instruct",
- "name": "Mistral: Mixtral 8x22B Instruct",
- "cost_per_1m_in": 0.8999999999999999,
- "cost_per_1m_out": 0.8999999999999999,
+ "id": "meta-llama/llama-3-70b-instruct",
+ "name": "Meta: Llama 3 70B Instruct",
+ "cost_per_1m_in": 0.3,
+ "cost_per_1m_out": 0.39999999999999997,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 65536,
- "default_max_tokens": 6553,
+ "context_window": 8192,
+ "default_max_tokens": 8192,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": false
},
{
- "id": "openai/gpt-4-turbo",
- "name": "OpenAI: GPT-4 Turbo",
- "cost_per_1m_in": 10,
- "cost_per_1m_out": 30,
+ "id": "mistralai/mixtral-8x22b-instruct",
+ "name": "Mistral: Mixtral 8x22B Instruct",
+ "cost_per_1m_in": 2,
+ "cost_per_1m_out": 6,
"cost_per_1m_in_cached": 0,
"cost_per_1m_out_cached": 0,
- "context_window": 128000,
- "default_max_tokens": 2048,
+ "context_window": 65536,
+ "default_max_tokens": 6553,
"can_reason": false,
"has_reasoning_efforts": false,
- "supports_attachments": true
+ "supports_attachments": false
},
{
"id": "google/gemini-pro-1.5",
"name": "Google: Gemini 1.5 Pro",
"cost_per_1m_in": 1.25,
"cost_per_1m_out": 5,
- "cost_per_1m_in_cached": 2.875,
- "cost_per_1m_out_cached": 0.625,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
"context_window": 2000000,
"default_max_tokens": 4096,
"can_reason": false,
"has_reasoning_efforts": false,
"supports_attachments": true
},
+ {
+ "id": "openai/gpt-4-turbo",
+ "name": "OpenAI: GPT-4 Turbo",
+ "cost_per_1m_in": 10,
+ "cost_per_1m_out": 30,
+ "cost_per_1m_in_cached": 0,
+ "cost_per_1m_out_cached": 0,
+ "context_window": 128000,
+ "default_max_tokens": 2048,
+ "can_reason": false,
+ "has_reasoning_efforts": false,
+ "supports_attachments": true
+ },
{
"id": "cohere/command-r-plus",
"name": "Cohere: Command R+",
@@ -1906,8 +1893,8 @@
"supports_attachments": false
},
{
- "id": "openai/gpt-4-0314",
- "name": "OpenAI: GPT-4 (older v0314)",
+ "id": "openai/gpt-4",
+ "name": "OpenAI: GPT-4",
"cost_per_1m_in": 30,
"cost_per_1m_out": 60,
"cost_per_1m_in_cached": 0,
@@ -1919,8 +1906,8 @@
"supports_attachments": false
},
{
- "id": "openai/gpt-4",
- "name": "OpenAI: GPT-4",
+ "id": "openai/gpt-4-0314",
+ "name": "OpenAI: GPT-4 (older v0314)",
"cost_per_1m_in": 30,
"cost_per_1m_out": 60,
"cost_per_1m_in_cached": 0,