@@ -73,7 +73,7 @@ type SelectedModel struct {
Think bool `json:"think,omitempty" jsonschema:"description=Enable thinking mode for Anthropic models that support reasoning"`
// Overrides the default model configuration.
- MaxTokens int64 `json:"max_tokens,omitempty" jsonschema:"description=Maximum number of tokens for model responses,minimum=1,maximum=200000,example=4096"`
+ MaxTokens int64 `json:"max_tokens,omitempty" jsonschema:"description=Maximum number of tokens for model responses,maximum=200000,example=4096"`
Temperature *float64 `json:"temperature,omitempty" jsonschema:"description=Sampling temperature,minimum=0,maximum=1,example=0.7"`
TopP *float64 `json:"top_p,omitempty" jsonschema:"description=Top-p (nucleus) sampling parameter,minimum=0,maximum=1,example=0.9"`
TopK *int64 `json:"top_k,omitempty" jsonschema:"description=Top-k sampling parameter"`
@@ -555,7 +555,6 @@
"max_tokens": {
"type": "integer",
"maximum": 200000,
- "minimum": 1,
"description": "Maximum number of tokens for model responses",
"examples": [
4096