chore(pi): update model list

Amolith created

Change summary

dot_config/pi/models.json | 59 ++++++++++++++++++++++++----------------
1 file changed, 35 insertions(+), 24 deletions(-)

Detailed changes

dot_config/pi/models.json 🔗

@@ -6,8 +6,8 @@
 			"api": "openai-completions",
 			"models": [
 				{
-					"id": "claude-opus-4.5",
-					"name": "Claude Opus 4.5",
+					"id": "claude-opus-4.6",
+					"name": "Claude Opus 4.6",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
@@ -17,26 +17,37 @@
 					"maxTokens": 50000
 				},
 				{
-					"id": "kimi-k2.5h",
-					"name": "Kimi K2.5h",
+					"id": "claude-sonnet-4.6",
+					"name": "Claude Sonnet 4.6",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
 					"input": ["text", "image"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
-					"contextWindow": 256000,
-					"maxTokens": 64000
+					"contextWindow": 200000,
+					"maxTokens": 50000
 				},
 				{
-					"id": "minimax-m2.5",
-					"name": "MiniMax M2.5",
+					"id": "gpt-5.3-codex",
+					"name": "GPT 5.3 Codex",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
-					"input": ["text"],
+					"input": ["text", "image"],
+					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
+					"contextWindow": 400000,
+					"maxTokens": 128000
+				},
+				{
+					"id": "kimi-k2.5h",
+					"name": "Kimi K2.5h",
+					"reasoning": true,
+					"supportsReasoningEffort": true,
+					"supportsDeveloperRole": false,
+					"input": ["text", "image"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
-					"contextWindow": 205000,
-					"maxTokens": 51250
+					"contextWindow": 262144,
+					"maxTokens": 32768
 				},
 				{
 					"id": "glm-5",
@@ -47,40 +58,40 @@
 					"input": ["text"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
 					"contextWindow": 202752,
-					"maxTokens": 50688
+					"maxTokens": 20275
 				},
 				{
-					"id": "qwen3.5-397b-a17b",
-					"name": "Qwen3.5 387B A17B",
+					"id": "glm-4.7",
+					"name": "GLM 4.7",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
 					"input": ["text"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
 					"contextWindow": 202752,
-					"maxTokens": 50688
+					"maxTokens": 20275
 				},
 				{
-					"id": "glm-4.7",
-					"name": "GLM 4.7",
+					"id": "minimax-m2.5",
+					"name": "MiniMax M2.5",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
 					"input": ["text"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
-					"contextWindow": 202752,
-					"maxTokens": 50688
+					"contextWindow": 196608,
+					"maxTokens": 19660
 				},
 				{
-					"id": "deepseek-v3.2",
-					"name": "DeepSeek V3.2",
+					"id": "qwen3.5-397b-a17b",
+					"name": "Qwen3.5 397B A17B",
 					"reasoning": true,
 					"supportsReasoningEffort": true,
 					"supportsDeveloperRole": false,
 					"input": ["text"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
-					"contextWindow": 162816,
-					"maxTokens": 40704
+					"contextWindow": 262144,
+					"maxTokens": 65536
 				},
 				{
 					"id": "gpt-oss-120b",
@@ -91,7 +102,7 @@
 					"input": ["text", "image"],
 					"cost": { "input": 0, "output": 0, "cacheRead": 0, "cacheWrite": 0 },
 					"contextWindow": 131072,
-					"maxTokens": 32768
+					"maxTokens": 13107
 				},
 				{
 					"id": "diff-apply",