1use anyhow::{Result, anyhow};
2use convert_case::{Case, Casing};
3use futures::{FutureExt, StreamExt, future, future::BoxFuture};
4use gpui::{AnyView, App, AsyncApp, Context, Entity, SharedString, Task, Window};
5use http_client::HttpClient;
6use language_model::{
7 AuthenticateError, LanguageModel, LanguageModelCompletionError, LanguageModelCompletionEvent,
8 LanguageModelId, LanguageModelName, LanguageModelProvider, LanguageModelProviderId,
9 LanguageModelProviderName, LanguageModelProviderState, LanguageModelRequest,
10 LanguageModelToolChoice, LanguageModelToolSchemaFormat, RateLimiter,
11};
12use menu;
13use open_ai::{ResponseStreamEvent, stream_completion};
14use settings::{Settings, SettingsStore};
15use std::sync::Arc;
16use ui::{ElevationIndex, Tooltip, prelude::*};
17use ui_input::SingleLineInput;
18use util::{ResultExt, truncate_and_trailoff};
19use zed_env_vars::EnvVar;
20
21use crate::api_key::ApiKeyState;
22use crate::provider::open_ai::{OpenAiEventMapper, into_open_ai};
23pub use settings::OpenAiCompatibleAvailableModel as AvailableModel;
24pub use settings::OpenAiCompatibleModelCapabilities as ModelCapabilities;
25
26#[derive(Default, Clone, Debug, PartialEq)]
27pub struct OpenAiCompatibleSettings {
28 pub api_url: String,
29 pub available_models: Vec<AvailableModel>,
30}
31
32pub struct OpenAiCompatibleLanguageModelProvider {
33 id: LanguageModelProviderId,
34 name: LanguageModelProviderName,
35 http_client: Arc<dyn HttpClient>,
36 state: Entity<State>,
37}
38
39pub struct State {
40 id: Arc<str>,
41 api_key_env_var: EnvVar,
42 api_key_state: ApiKeyState,
43 settings: OpenAiCompatibleSettings,
44}
45
46impl State {
47 const fn is_authenticated(&self) -> bool {
48 self.api_key_state.has_key()
49 }
50
51 fn set_api_key(&mut self, api_key: Option<String>, cx: &mut Context<Self>) -> Task<Result<()>> {
52 let api_url = SharedString::new(self.settings.api_url.as_str());
53 self.api_key_state
54 .store(api_url, api_key, |this| &mut this.api_key_state, cx)
55 }
56
57 fn authenticate(&mut self, cx: &mut Context<Self>) -> Task<Result<(), AuthenticateError>> {
58 let api_url = SharedString::new(self.settings.api_url.clone());
59 self.api_key_state.load_if_needed(
60 api_url,
61 &self.api_key_env_var,
62 |this| &mut this.api_key_state,
63 cx,
64 )
65 }
66}
67
68impl OpenAiCompatibleLanguageModelProvider {
69 pub fn new(id: Arc<str>, http_client: Arc<dyn HttpClient>, cx: &mut App) -> Self {
70 fn resolve_settings<'a>(id: &'a str, cx: &'a App) -> Option<&'a OpenAiCompatibleSettings> {
71 crate::AllLanguageModelSettings::get_global(cx)
72 .openai_compatible
73 .get(id)
74 }
75
76 let api_key_env_var_name = format!("{}_API_KEY", id).to_case(Case::UpperSnake).into();
77 let state = cx.new(|cx| {
78 cx.observe_global::<SettingsStore>(|this: &mut State, cx| {
79 let Some(settings) = resolve_settings(&this.id, cx).cloned() else {
80 return;
81 };
82 if &this.settings != &settings {
83 let api_url = SharedString::new(settings.api_url.as_str());
84 this.api_key_state.handle_url_change(
85 api_url,
86 &this.api_key_env_var,
87 |this| &mut this.api_key_state,
88 cx,
89 );
90 this.settings = settings;
91 cx.notify();
92 }
93 })
94 .detach();
95 let settings = resolve_settings(&id, cx).cloned().unwrap_or_default();
96 State {
97 id: id.clone(),
98 api_key_env_var: EnvVar::new(api_key_env_var_name),
99 api_key_state: ApiKeyState::new(SharedString::new(settings.api_url.as_str())),
100 settings,
101 }
102 });
103
104 Self {
105 id: id.clone().into(),
106 name: id.into(),
107 http_client,
108 state,
109 }
110 }
111
112 fn create_language_model(&self, model: AvailableModel) -> Arc<dyn LanguageModel> {
113 Arc::new(OpenAiCompatibleLanguageModel {
114 id: LanguageModelId::from(model.name.clone()),
115 provider_id: self.id.clone(),
116 provider_name: self.name.clone(),
117 model,
118 state: self.state.clone(),
119 http_client: self.http_client.clone(),
120 request_limiter: RateLimiter::new(4),
121 })
122 }
123}
124
125impl LanguageModelProviderState for OpenAiCompatibleLanguageModelProvider {
126 type ObservableEntity = State;
127
128 fn observable_entity(&self) -> Option<Entity<Self::ObservableEntity>> {
129 Some(self.state.clone())
130 }
131}
132
133impl LanguageModelProvider for OpenAiCompatibleLanguageModelProvider {
134 fn id(&self) -> LanguageModelProviderId {
135 self.id.clone()
136 }
137
138 fn name(&self) -> LanguageModelProviderName {
139 self.name.clone()
140 }
141
142 fn icon(&self) -> IconName {
143 IconName::AiOpenAiCompat
144 }
145
146 fn default_model(&self, cx: &App) -> Option<Arc<dyn LanguageModel>> {
147 self.state
148 .read(cx)
149 .settings
150 .available_models
151 .first()
152 .map(|model| self.create_language_model(model.clone()))
153 }
154
155 fn default_fast_model(&self, _cx: &App) -> Option<Arc<dyn LanguageModel>> {
156 None
157 }
158
159 fn provided_models(&self, cx: &App) -> Vec<Arc<dyn LanguageModel>> {
160 self.state
161 .read(cx)
162 .settings
163 .available_models
164 .iter()
165 .map(|model| self.create_language_model(model.clone()))
166 .collect()
167 }
168
169 fn is_authenticated(&self, cx: &App) -> bool {
170 self.state.read(cx).is_authenticated()
171 }
172
173 fn authenticate(&self, cx: &mut App) -> Task<Result<(), AuthenticateError>> {
174 self.state.update(cx, |state, cx| state.authenticate(cx))
175 }
176
177 fn configuration_view(
178 &self,
179 _target_agent: language_model::ConfigurationViewTargetAgent,
180 window: &mut Window,
181 cx: &mut App,
182 ) -> AnyView {
183 cx.new(|cx| ConfigurationView::new(self.state.clone(), window, cx))
184 .into()
185 }
186
187 fn reset_credentials(&self, cx: &mut App) -> Task<Result<()>> {
188 self.state
189 .update(cx, |state, cx| state.set_api_key(None, cx))
190 }
191}
192
193pub struct OpenAiCompatibleLanguageModel {
194 id: LanguageModelId,
195 provider_id: LanguageModelProviderId,
196 provider_name: LanguageModelProviderName,
197 model: AvailableModel,
198 state: Entity<State>,
199 http_client: Arc<dyn HttpClient>,
200 request_limiter: RateLimiter,
201}
202
203impl OpenAiCompatibleLanguageModel {
204 fn stream_completion(
205 &self,
206 request: open_ai::Request,
207 cx: &AsyncApp,
208 ) -> BoxFuture<'static, Result<futures::stream::BoxStream<'static, Result<ResponseStreamEvent>>>>
209 {
210 let http_client = self.http_client.clone();
211
212 let Ok((api_key, api_url)) = self.state.read_with(cx, |state, _cx| {
213 let api_url = &state.settings.api_url;
214 (
215 state.api_key_state.key(api_url),
216 state.settings.api_url.clone(),
217 )
218 }) else {
219 return future::ready(Err(anyhow!("App state dropped"))).boxed();
220 };
221
222 let provider = self.provider_name.clone();
223 let future = self.request_limiter.stream(async move {
224 let Some(api_key) = api_key else {
225 return Err(LanguageModelCompletionError::NoApiKey { provider });
226 };
227 let request = stream_completion(http_client.as_ref(), &api_url, &api_key, request);
228 let response = request.await?;
229 Ok(response)
230 });
231
232 async move { Ok(future.await?.boxed()) }.boxed()
233 }
234}
235
236impl LanguageModel for OpenAiCompatibleLanguageModel {
237 fn id(&self) -> LanguageModelId {
238 self.id.clone()
239 }
240
241 fn name(&self) -> LanguageModelName {
242 LanguageModelName::from(
243 self.model
244 .display_name
245 .clone()
246 .unwrap_or_else(|| self.model.name.clone()),
247 )
248 }
249
250 fn provider_id(&self) -> LanguageModelProviderId {
251 self.provider_id.clone()
252 }
253
254 fn provider_name(&self) -> LanguageModelProviderName {
255 self.provider_name.clone()
256 }
257
258 fn supports_tools(&self) -> bool {
259 self.model.capabilities.tools
260 }
261
262 fn tool_input_format(&self) -> LanguageModelToolSchemaFormat {
263 LanguageModelToolSchemaFormat::JsonSchemaSubset
264 }
265
266 fn supports_images(&self) -> bool {
267 self.model.capabilities.images
268 }
269
270 fn supports_tool_choice(&self, choice: LanguageModelToolChoice) -> bool {
271 match choice {
272 LanguageModelToolChoice::Auto => self.model.capabilities.tools,
273 LanguageModelToolChoice::Any => self.model.capabilities.tools,
274 LanguageModelToolChoice::None => true,
275 }
276 }
277
278 fn telemetry_id(&self) -> String {
279 format!("openai/{}", self.model.name)
280 }
281
282 fn max_token_count(&self) -> u64 {
283 self.model.max_tokens
284 }
285
286 fn max_output_tokens(&self) -> Option<u64> {
287 self.model.max_output_tokens
288 }
289
290 fn count_tokens(
291 &self,
292 request: LanguageModelRequest,
293 cx: &App,
294 ) -> BoxFuture<'static, Result<u64>> {
295 let max_token_count = self.max_token_count();
296 cx.background_spawn(async move {
297 let messages = super::open_ai::collect_tiktoken_messages(request);
298 let model = if max_token_count >= 100_000 {
299 // If the max tokens is 100k or more, it is likely the o200k_base tokenizer from gpt4o
300 "gpt-4o"
301 } else {
302 // Otherwise fallback to gpt-4, since only cl100k_base and o200k_base are
303 // supported with this tiktoken method
304 "gpt-4"
305 };
306 tiktoken_rs::num_tokens_from_messages(model, &messages).map(|tokens| tokens as u64)
307 })
308 .boxed()
309 }
310
311 fn stream_completion(
312 &self,
313 request: LanguageModelRequest,
314 cx: &AsyncApp,
315 ) -> BoxFuture<
316 'static,
317 Result<
318 futures::stream::BoxStream<
319 'static,
320 Result<LanguageModelCompletionEvent, LanguageModelCompletionError>,
321 >,
322 LanguageModelCompletionError,
323 >,
324 > {
325 let request = into_open_ai(
326 request,
327 &self.model.name,
328 self.model.capabilities.parallel_tool_calls,
329 self.model.capabilities.prompt_cache_key,
330 self.max_output_tokens(),
331 None,
332 );
333 let completions = self.stream_completion(request, cx);
334 async move {
335 let mapper = OpenAiEventMapper::new();
336 Ok(mapper.map_stream(completions.await?).boxed())
337 }
338 .boxed()
339 }
340}
341
342struct ConfigurationView {
343 api_key_editor: Entity<SingleLineInput>,
344 state: Entity<State>,
345 load_credentials_task: Option<Task<()>>,
346}
347
348impl ConfigurationView {
349 fn new(state: Entity<State>, window: &mut Window, cx: &mut Context<Self>) -> Self {
350 let api_key_editor = cx.new(|cx| {
351 SingleLineInput::new(
352 window,
353 cx,
354 "000000000000000000000000000000000000000000000000000",
355 )
356 });
357
358 cx.observe(&state, |_, _, cx| {
359 cx.notify();
360 })
361 .detach();
362
363 let load_credentials_task = Some(cx.spawn_in(window, {
364 let state = state.clone();
365 async move |this, cx| {
366 if let Some(task) = state
367 .update(cx, |state, cx| state.authenticate(cx))
368 .log_err()
369 {
370 // We don't log an error, because "not signed in" is also an error.
371 let _ = task.await;
372 }
373 this.update(cx, |this, cx| {
374 this.load_credentials_task = None;
375 cx.notify();
376 })
377 .log_err();
378 }
379 }));
380
381 Self {
382 api_key_editor,
383 state,
384 load_credentials_task,
385 }
386 }
387
388 fn save_api_key(&mut self, _: &menu::Confirm, window: &mut Window, cx: &mut Context<Self>) {
389 let api_key = self.api_key_editor.read(cx).text(cx).trim().to_string();
390 if api_key.is_empty() {
391 return;
392 }
393
394 // url changes can cause the editor to be displayed again
395 self.api_key_editor
396 .update(cx, |input, cx| input.set_text("", window, cx));
397
398 let state = self.state.clone();
399 cx.spawn_in(window, async move |_, cx| {
400 state
401 .update(cx, |state, cx| state.set_api_key(Some(api_key), cx))?
402 .await
403 })
404 .detach_and_log_err(cx);
405 }
406
407 fn reset_api_key(&mut self, window: &mut Window, cx: &mut Context<Self>) {
408 self.api_key_editor
409 .update(cx, |input, cx| input.set_text("", window, cx));
410
411 let state = self.state.clone();
412 cx.spawn_in(window, async move |_, cx| {
413 state
414 .update(cx, |state, cx| state.set_api_key(None, cx))?
415 .await
416 })
417 .detach_and_log_err(cx);
418 }
419
420 fn should_render_editor(&self, cx: &Context<Self>) -> bool {
421 !self.state.read(cx).is_authenticated()
422 }
423}
424
425impl Render for ConfigurationView {
426 fn render(&mut self, _: &mut Window, cx: &mut Context<Self>) -> impl IntoElement {
427 let state = self.state.read(cx);
428 let env_var_set = state.api_key_state.is_from_env_var();
429 let env_var_name = &state.api_key_env_var.name;
430
431 let api_key_section = if self.should_render_editor(cx) {
432 v_flex()
433 .on_action(cx.listener(Self::save_api_key))
434 .child(Label::new("To use Zed's agent with an OpenAI-compatible provider, you need to add an API key."))
435 .child(
436 div()
437 .pt(DynamicSpacing::Base04.rems(cx))
438 .child(self.api_key_editor.clone())
439 )
440 .child(
441 Label::new(
442 format!("You can also assign the {env_var_name} environment variable and restart Zed."),
443 )
444 .size(LabelSize::Small).color(Color::Muted),
445 )
446 .into_any()
447 } else {
448 h_flex()
449 .mt_1()
450 .p_1()
451 .justify_between()
452 .rounded_md()
453 .border_1()
454 .border_color(cx.theme().colors().border)
455 .bg(cx.theme().colors().background)
456 .child(
457 h_flex()
458 .gap_1()
459 .child(Icon::new(IconName::Check).color(Color::Success))
460 .child(Label::new(if env_var_set {
461 format!("API key set in {env_var_name} environment variable")
462 } else {
463 format!("API key configured for {}", truncate_and_trailoff(&state.settings.api_url, 32))
464 })),
465 )
466 .child(
467 Button::new("reset-api-key", "Reset API Key")
468 .label_size(LabelSize::Small)
469 .icon(IconName::Undo)
470 .icon_size(IconSize::Small)
471 .icon_position(IconPosition::Start)
472 .layer(ElevationIndex::ModalSurface)
473 .when(env_var_set, |this| {
474 this.tooltip(Tooltip::text(format!("To reset your API key, unset the {env_var_name} environment variable.")))
475 })
476 .on_click(cx.listener(|this, _, window, cx| this.reset_api_key(window, cx))),
477 )
478 .into_any()
479 };
480
481 if self.load_credentials_task.is_some() {
482 div().child(Label::new("Loading credentials…")).into_any()
483 } else {
484 v_flex().size_full().child(api_key_section).into_any()
485 }
486 }
487}