fix(z.ai): reduce default_max_tokens for glm4.7 (#237)

mkaaad created

Change summary

internal/providers/configs/zai.json          | 2 +-
internal/providers/configs/zhipu-coding.json | 2 +-
internal/providers/configs/zhipu.json        | 2 +-
3 files changed, 3 insertions(+), 3 deletions(-)

Detailed changes

internal/providers/configs/zai.json 🔗

@@ -47,7 +47,7 @@
       "cost_per_1m_out": 2.2,
       "cost_per_1m_in_cached": 0.11,
       "context_window": 204800,
-      "default_max_tokens": 102400,
+      "default_max_tokens": 98000,
       "can_reason": true,
       "supports_attachments": false
     },

internal/providers/configs/zhipu-coding.json 🔗

@@ -25,7 +25,7 @@
       "cost_per_1m_out": 2.2,
       "cost_per_1m_in_cached": 0.11,
       "context_window": 204800,
-      "default_max_tokens": 102400,
+      "default_max_tokens": 98000,
       "can_reason": true,
       "supports_attachments": false
     },

internal/providers/configs/zhipu.json 🔗

@@ -36,7 +36,7 @@
       "cost_per_1m_out": 2.2,
       "cost_per_1m_in_cached": 0.11,
       "context_window": 204800,
-      "default_max_tokens": 102400,
+      "default_max_tokens": 98000,
       "can_reason": true,
       "supports_attachments": false
     },