diff --git a/internal/llm/prompt/coder.go b/internal/llm/prompt/coder.go index df12bd17dc2320cf4f61caab2e1650a11a23a894..bf3eb5ea47d877c8adf21816900d7acb1434aa82 100644 --- a/internal/llm/prompt/coder.go +++ b/internal/llm/prompt/coder.go @@ -20,7 +20,8 @@ func CoderPrompt(p string, contextFiles ...string) string { basePrompt = string(anthropicCoderPrompt) switch p { case string(catwalk.InferenceProviderOpenAI): - basePrompt = string(openaiCoderPrompt) + // seems to behave better + basePrompt = string(coderV2Prompt) case string(catwalk.InferenceProviderGemini): basePrompt = string(geminiCoderPrompt) } diff --git a/internal/llm/provider/openai.go b/internal/llm/provider/openai.go index 9b612b64d63a55d278a686ba5b3f7bda7f973c69..8433282fc0b5e771a0a28184895a864b30b0e389 100644 --- a/internal/llm/provider/openai.go +++ b/internal/llm/provider/openai.go @@ -255,6 +255,8 @@ func (o *openaiClient) preparedParams(messages []openai.ChatCompletionMessagePar params.ReasoningEffort = shared.ReasoningEffortMedium case "high": params.ReasoningEffort = shared.ReasoningEffortHigh + case "minimal": + params.ReasoningEffort = shared.ReasoningEffort("minimal") default: params.ReasoningEffort = shared.ReasoningEffort(reasoningEffort) }