From e5706f2349429bc8d4136e4d515602eccae62b7e Mon Sep 17 00:00:00 2001 From: Richard Feldman Date: Fri, 16 Jan 2026 17:09:08 -0500 Subject: [PATCH] Add BYOK GPT-5.2-codex support (#47025) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Screenshot 2026-01-16 at 4 52 12 PM Release Notes: - Add support for GPT-5.2-Codex via OpenAI API Key --- crates/language_models/src/provider/open_ai.rs | 5 +++-- crates/open_ai/src/open_ai.rs | 10 +++++++++- 2 files changed, 12 insertions(+), 3 deletions(-) diff --git a/crates/language_models/src/provider/open_ai.rs b/crates/language_models/src/provider/open_ai.rs index 92b47590d6bef2e2eeb21208be8ccdb5b012563c..915fd325494f523f1a1d64ef17cec63bf29cc44d 100644 --- a/crates/language_models/src/provider/open_ai.rs +++ b/crates/language_models/src/provider/open_ai.rs @@ -310,6 +310,7 @@ impl LanguageModel for OpenAiLanguageModel { | Model::FiveNano | Model::FivePointOne | Model::FivePointTwo + | Model::FivePointTwoCodex | Model::O1 | Model::O3 | Model::O4Mini => true, @@ -1166,8 +1167,8 @@ pub fn count_open_ai_tokens( | Model::FiveCodex | Model::FiveMini | Model::FiveNano => tiktoken_rs::num_tokens_from_messages(model.id(), &messages), - // GPT-5.1 and 5.2 don't have dedicated tiktoken support; use gpt-5 tokenizer - Model::FivePointOne | Model::FivePointTwo => { + // GPT-5.1, 5.2, and 5.2-codex don't have dedicated tiktoken support; use gpt-5 tokenizer + Model::FivePointOne | Model::FivePointTwo | Model::FivePointTwoCodex => { tiktoken_rs::num_tokens_from_messages("gpt-5", &messages) } } diff --git a/crates/open_ai/src/open_ai.rs b/crates/open_ai/src/open_ai.rs index 10ed9605263677a2673ca0c2a12a0094ad42b05f..0f88aaddc608d8c794d6d7e867cbac6c042d53df 100644 --- a/crates/open_ai/src/open_ai.rs +++ b/crates/open_ai/src/open_ai.rs @@ -93,6 +93,8 @@ pub enum Model { FivePointOne, #[serde(rename = "gpt-5.2")] FivePointTwo, + #[serde(rename = "gpt-5.2-codex")] + FivePointTwoCodex, #[serde(rename = "custom")] Custom { name: String, @@ -137,6 +139,7 @@ impl Model { "gpt-5-nano" => Ok(Self::FiveNano), "gpt-5.1" => Ok(Self::FivePointOne), "gpt-5.2" => Ok(Self::FivePointTwo), + "gpt-5.2-codex" => Ok(Self::FivePointTwoCodex), invalid_id => anyhow::bail!("invalid model id '{invalid_id}'"), } } @@ -161,6 +164,7 @@ impl Model { Self::FiveNano => "gpt-5-nano", Self::FivePointOne => "gpt-5.1", Self::FivePointTwo => "gpt-5.2", + Self::FivePointTwoCodex => "gpt-5.2-codex", Self::Custom { name, .. } => name, } } @@ -185,6 +189,7 @@ impl Model { Self::FiveNano => "gpt-5-nano", Self::FivePointOne => "gpt-5.1", Self::FivePointTwo => "gpt-5.2", + Self::FivePointTwoCodex => "gpt-5.2-codex", Self::Custom { name, display_name, .. } => display_name.as_ref().unwrap_or(name), @@ -211,6 +216,7 @@ impl Model { Self::FiveNano => 272_000, Self::FivePointOne => 400_000, Self::FivePointTwo => 400_000, + Self::FivePointTwoCodex => 400_000, Self::Custom { max_tokens, .. } => *max_tokens, } } @@ -238,6 +244,7 @@ impl Model { Self::FiveNano => Some(128_000), Self::FivePointOne => Some(128_000), Self::FivePointTwo => Some(128_000), + Self::FivePointTwoCodex => Some(128_000), } } @@ -256,7 +263,7 @@ impl Model { supports_chat_completions, .. } => *supports_chat_completions, - Self::FiveCodex => false, + Self::FiveCodex | Self::FivePointTwoCodex => false, _ => true, } } @@ -279,6 +286,7 @@ impl Model { | Self::FiveMini | Self::FivePointOne | Self::FivePointTwo + | Self::FivePointTwoCodex | Self::FiveNano => true, Self::O1 | Self::O3 | Self::O3Mini | Self::O4Mini | Model::Custom { .. } => false, }