From 67525cca7154cc2eb0f16109cde0f97f13c076b1 Mon Sep 17 00:00:00 2001 From: Peter Tripp Date: Wed, 15 Jan 2025 10:44:15 -0500 Subject: [PATCH] Add ollama phi4 context size defaults (#23036) Add `phi4` maximum context length (128K). By default this clamps to `16384` but if you have enough video memory you can set it higher or connect to a non-local machine via settings: ```json "language_models": { "ollama": { "api_url": "http://localhost:11434", "available_models": [ { "name": "phi4", "display_name": "Phi4 64K", "max_tokens": 65536 } ] } } ``` Release Notes: - Improve support for Phi4 with ollama. --- crates/ollama/src/ollama.rs | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/crates/ollama/src/ollama.rs b/crates/ollama/src/ollama.rs index 5168da38be5a53c7e73931d7648e0c82db5b7750..ab092c5ecdbcfbfd05baf3e06d0ad1a00b6f3baa 100644 --- a/crates/ollama/src/ollama.rs +++ b/crates/ollama/src/ollama.rs @@ -83,8 +83,8 @@ fn get_max_tokens(name: &str) -> usize { "codellama" | "starcoder2" => 16384, "mistral" | "codestral" | "mixstral" | "llava" | "qwen2" | "qwen2.5-coder" | "dolphin-mixtral" => 32768, - "llama3.1" | "phi3" | "phi3.5" | "command-r" | "deepseek-coder-v2" | "yi-coder" - | "llama3.2" => 128000, + "llama3.1" | "phi3" | "phi3.5" | "phi4" | "command-r" | "deepseek-coder-v2" + | "yi-coder" | "llama3.2" => 128000, _ => DEFAULT_TOKENS, } .clamp(1, MAXIMUM_TOKENS)