deepseek.rs

  1use anyhow::{Result, anyhow};
  2use collections::{BTreeMap, HashMap};
  3use deepseek::DEEPSEEK_API_URL;
  4
  5use futures::Stream;
  6use futures::{FutureExt, StreamExt, future::BoxFuture, stream::BoxStream};
  7use gpui::{AnyView, App, AsyncApp, Context, Entity, SharedString, Task, Window};
  8use http_client::HttpClient;
  9use language_model::{
 10    ApiKeyState, AuthenticateError, EnvVar, IconOrSvg, LanguageModel, LanguageModelCompletionError,
 11    LanguageModelCompletionEvent, LanguageModelId, LanguageModelName, LanguageModelProvider,
 12    LanguageModelProviderId, LanguageModelProviderName, LanguageModelProviderState,
 13    LanguageModelRequest, LanguageModelToolChoice, LanguageModelToolResultContent,
 14    LanguageModelToolUse, MessageContent, RateLimiter, Role, StopReason, TokenUsage, env_var,
 15};
 16pub use settings::DeepseekAvailableModel as AvailableModel;
 17use settings::{Settings, SettingsStore};
 18use std::pin::Pin;
 19use std::sync::{Arc, LazyLock};
 20
 21use ui::{ButtonLink, ConfiguredApiCard, List, ListBulletItem, prelude::*};
 22use ui_input::InputField;
 23use util::ResultExt;
 24
 25use crate::provider::util::parse_tool_arguments;
 26
 27const PROVIDER_ID: LanguageModelProviderId = LanguageModelProviderId::new("deepseek");
 28const PROVIDER_NAME: LanguageModelProviderName = LanguageModelProviderName::new("DeepSeek");
 29
 30const API_KEY_ENV_VAR_NAME: &str = "DEEPSEEK_API_KEY";
 31static API_KEY_ENV_VAR: LazyLock<EnvVar> = env_var!(API_KEY_ENV_VAR_NAME);
 32
 33#[derive(Default)]
 34struct RawToolCall {
 35    id: String,
 36    name: String,
 37    arguments: String,
 38}
 39
 40#[derive(Default, Clone, Debug, PartialEq)]
 41pub struct DeepSeekSettings {
 42    pub api_url: String,
 43    pub available_models: Vec<AvailableModel>,
 44}
 45pub struct DeepSeekLanguageModelProvider {
 46    http_client: Arc<dyn HttpClient>,
 47    state: Entity<State>,
 48}
 49
 50pub struct State {
 51    api_key_state: ApiKeyState,
 52}
 53
 54impl State {
 55    fn is_authenticated(&self) -> bool {
 56        self.api_key_state.has_key()
 57    }
 58
 59    fn set_api_key(&mut self, api_key: Option<String>, cx: &mut Context<Self>) -> Task<Result<()>> {
 60        let api_url = DeepSeekLanguageModelProvider::api_url(cx);
 61        self.api_key_state
 62            .store(api_url, api_key, |this| &mut this.api_key_state, cx)
 63    }
 64
 65    fn authenticate(&mut self, cx: &mut Context<Self>) -> Task<Result<(), AuthenticateError>> {
 66        let api_url = DeepSeekLanguageModelProvider::api_url(cx);
 67        self.api_key_state
 68            .load_if_needed(api_url, |this| &mut this.api_key_state, cx)
 69    }
 70}
 71
 72impl DeepSeekLanguageModelProvider {
 73    pub fn new(http_client: Arc<dyn HttpClient>, cx: &mut App) -> Self {
 74        let state = cx.new(|cx| {
 75            cx.observe_global::<SettingsStore>(|this: &mut State, cx| {
 76                let api_url = Self::api_url(cx);
 77                this.api_key_state
 78                    .handle_url_change(api_url, |this| &mut this.api_key_state, cx);
 79                cx.notify();
 80            })
 81            .detach();
 82            State {
 83                api_key_state: ApiKeyState::new(Self::api_url(cx), (*API_KEY_ENV_VAR).clone()),
 84            }
 85        });
 86
 87        Self { http_client, state }
 88    }
 89
 90    fn create_language_model(&self, model: deepseek::Model) -> Arc<dyn LanguageModel> {
 91        Arc::new(DeepSeekLanguageModel {
 92            id: LanguageModelId::from(model.id().to_string()),
 93            model,
 94            state: self.state.clone(),
 95            http_client: self.http_client.clone(),
 96            request_limiter: RateLimiter::new(4),
 97        })
 98    }
 99
100    fn settings(cx: &App) -> &DeepSeekSettings {
101        &crate::AllLanguageModelSettings::get_global(cx).deepseek
102    }
103
104    fn api_url(cx: &App) -> SharedString {
105        let api_url = &Self::settings(cx).api_url;
106        if api_url.is_empty() {
107            DEEPSEEK_API_URL.into()
108        } else {
109            SharedString::new(api_url.as_str())
110        }
111    }
112}
113
114impl LanguageModelProviderState for DeepSeekLanguageModelProvider {
115    type ObservableEntity = State;
116
117    fn observable_entity(&self) -> Option<Entity<Self::ObservableEntity>> {
118        Some(self.state.clone())
119    }
120}
121
122impl LanguageModelProvider for DeepSeekLanguageModelProvider {
123    fn id(&self) -> LanguageModelProviderId {
124        PROVIDER_ID
125    }
126
127    fn name(&self) -> LanguageModelProviderName {
128        PROVIDER_NAME
129    }
130
131    fn icon(&self) -> IconOrSvg {
132        IconOrSvg::Icon(IconName::AiDeepSeek)
133    }
134
135    fn default_model(&self, _cx: &App) -> Option<Arc<dyn LanguageModel>> {
136        Some(self.create_language_model(deepseek::Model::default()))
137    }
138
139    fn default_fast_model(&self, _cx: &App) -> Option<Arc<dyn LanguageModel>> {
140        Some(self.create_language_model(deepseek::Model::default_fast()))
141    }
142
143    fn provided_models(&self, cx: &App) -> Vec<Arc<dyn LanguageModel>> {
144        let mut models = BTreeMap::default();
145
146        models.insert("deepseek-chat", deepseek::Model::Chat);
147        models.insert("deepseek-reasoner", deepseek::Model::Reasoner);
148
149        for available_model in &Self::settings(cx).available_models {
150            models.insert(
151                &available_model.name,
152                deepseek::Model::Custom {
153                    name: available_model.name.clone(),
154                    display_name: available_model.display_name.clone(),
155                    max_tokens: available_model.max_tokens,
156                    max_output_tokens: available_model.max_output_tokens,
157                },
158            );
159        }
160
161        models
162            .into_values()
163            .map(|model| self.create_language_model(model))
164            .collect()
165    }
166
167    fn is_authenticated(&self, cx: &App) -> bool {
168        self.state.read(cx).is_authenticated()
169    }
170
171    fn authenticate(&self, cx: &mut App) -> Task<Result<(), AuthenticateError>> {
172        self.state.update(cx, |state, cx| state.authenticate(cx))
173    }
174
175    fn configuration_view(
176        &self,
177        _target_agent: language_model::ConfigurationViewTargetAgent,
178        window: &mut Window,
179        cx: &mut App,
180    ) -> AnyView {
181        cx.new(|cx| ConfigurationView::new(self.state.clone(), window, cx))
182            .into()
183    }
184
185    fn reset_credentials(&self, cx: &mut App) -> Task<Result<()>> {
186        self.state
187            .update(cx, |state, cx| state.set_api_key(None, cx))
188    }
189}
190
191pub struct DeepSeekLanguageModel {
192    id: LanguageModelId,
193    model: deepseek::Model,
194    state: Entity<State>,
195    http_client: Arc<dyn HttpClient>,
196    request_limiter: RateLimiter,
197}
198
199impl DeepSeekLanguageModel {
200    fn stream_completion(
201        &self,
202        request: deepseek::Request,
203        cx: &AsyncApp,
204    ) -> BoxFuture<'static, Result<BoxStream<'static, Result<deepseek::StreamResponse>>>> {
205        let http_client = self.http_client.clone();
206
207        let (api_key, api_url) = self.state.read_with(cx, |state, cx| {
208            let api_url = DeepSeekLanguageModelProvider::api_url(cx);
209            (state.api_key_state.key(&api_url), api_url)
210        });
211
212        let future = self.request_limiter.stream(async move {
213            let Some(api_key) = api_key else {
214                return Err(LanguageModelCompletionError::NoApiKey {
215                    provider: PROVIDER_NAME,
216                });
217            };
218            let request =
219                deepseek::stream_completion(http_client.as_ref(), &api_url, &api_key, request);
220            let response = request.await?;
221            Ok(response)
222        });
223
224        async move { Ok(future.await?.boxed()) }.boxed()
225    }
226}
227
228impl LanguageModel for DeepSeekLanguageModel {
229    fn id(&self) -> LanguageModelId {
230        self.id.clone()
231    }
232
233    fn name(&self) -> LanguageModelName {
234        LanguageModelName::from(self.model.display_name().to_string())
235    }
236
237    fn provider_id(&self) -> LanguageModelProviderId {
238        PROVIDER_ID
239    }
240
241    fn provider_name(&self) -> LanguageModelProviderName {
242        PROVIDER_NAME
243    }
244
245    fn supports_tools(&self) -> bool {
246        true
247    }
248
249    fn supports_tool_choice(&self, _choice: LanguageModelToolChoice) -> bool {
250        true
251    }
252
253    fn supports_images(&self) -> bool {
254        false
255    }
256
257    fn telemetry_id(&self) -> String {
258        format!("deepseek/{}", self.model.id())
259    }
260
261    fn max_token_count(&self) -> u64 {
262        self.model.max_token_count()
263    }
264
265    fn max_output_tokens(&self) -> Option<u64> {
266        self.model.max_output_tokens()
267    }
268
269    fn count_tokens(
270        &self,
271        request: LanguageModelRequest,
272        cx: &App,
273    ) -> BoxFuture<'static, Result<u64>> {
274        cx.background_spawn(async move {
275            let messages = request
276                .messages
277                .into_iter()
278                .map(|message| tiktoken_rs::ChatCompletionRequestMessage {
279                    role: match message.role {
280                        Role::User => "user".into(),
281                        Role::Assistant => "assistant".into(),
282                        Role::System => "system".into(),
283                    },
284                    content: Some(message.string_contents()),
285                    name: None,
286                    function_call: None,
287                })
288                .collect::<Vec<_>>();
289
290            tiktoken_rs::num_tokens_from_messages("gpt-4", &messages).map(|tokens| tokens as u64)
291        })
292        .boxed()
293    }
294
295    fn stream_completion(
296        &self,
297        request: LanguageModelRequest,
298        cx: &AsyncApp,
299    ) -> BoxFuture<
300        'static,
301        Result<
302            BoxStream<'static, Result<LanguageModelCompletionEvent, LanguageModelCompletionError>>,
303            LanguageModelCompletionError,
304        >,
305    > {
306        let request = into_deepseek(request, &self.model, self.max_output_tokens());
307        let stream = self.stream_completion(request, cx);
308
309        async move {
310            let mapper = DeepSeekEventMapper::new();
311            Ok(mapper.map_stream(stream.await?).boxed())
312        }
313        .boxed()
314    }
315}
316
317pub fn into_deepseek(
318    request: LanguageModelRequest,
319    model: &deepseek::Model,
320    max_output_tokens: Option<u64>,
321) -> deepseek::Request {
322    let is_reasoner = model == &deepseek::Model::Reasoner;
323
324    let mut messages = Vec::new();
325    let mut current_reasoning: Option<String> = None;
326
327    for message in request.messages {
328        for content in message.content {
329            match content {
330                MessageContent::Text(text) => messages.push(match message.role {
331                    Role::User => deepseek::RequestMessage::User { content: text },
332                    Role::Assistant => deepseek::RequestMessage::Assistant {
333                        content: Some(text),
334                        tool_calls: Vec::new(),
335                        reasoning_content: current_reasoning.take(),
336                    },
337                    Role::System => deepseek::RequestMessage::System { content: text },
338                }),
339                MessageContent::Thinking { text, .. } => {
340                    // Accumulate reasoning content for next assistant message
341                    current_reasoning.get_or_insert_default().push_str(&text);
342                }
343                MessageContent::RedactedThinking(_) => {}
344                MessageContent::Image(_) => {}
345                MessageContent::ToolUse(tool_use) => {
346                    let tool_call = deepseek::ToolCall {
347                        id: tool_use.id.to_string(),
348                        content: deepseek::ToolCallContent::Function {
349                            function: deepseek::FunctionContent {
350                                name: tool_use.name.to_string(),
351                                arguments: serde_json::to_string(&tool_use.input)
352                                    .unwrap_or_default(),
353                            },
354                        },
355                    };
356
357                    if let Some(deepseek::RequestMessage::Assistant { tool_calls, .. }) =
358                        messages.last_mut()
359                    {
360                        tool_calls.push(tool_call);
361                    } else {
362                        messages.push(deepseek::RequestMessage::Assistant {
363                            content: None,
364                            tool_calls: vec![tool_call],
365                            reasoning_content: current_reasoning.take(),
366                        });
367                    }
368                }
369                MessageContent::ToolResult(tool_result) => {
370                    match &tool_result.content {
371                        LanguageModelToolResultContent::Text(text) => {
372                            messages.push(deepseek::RequestMessage::Tool {
373                                content: text.to_string(),
374                                tool_call_id: tool_result.tool_use_id.to_string(),
375                            });
376                        }
377                        LanguageModelToolResultContent::Image(_) => {}
378                    };
379                }
380            }
381        }
382    }
383
384    deepseek::Request {
385        model: model.id().to_string(),
386        messages,
387        stream: true,
388        max_tokens: max_output_tokens,
389        temperature: if is_reasoner {
390            None
391        } else {
392            request.temperature
393        },
394        response_format: None,
395        tools: request
396            .tools
397            .into_iter()
398            .map(|tool| deepseek::ToolDefinition::Function {
399                function: deepseek::FunctionDefinition {
400                    name: tool.name,
401                    description: Some(tool.description),
402                    parameters: Some(tool.input_schema),
403                },
404            })
405            .collect(),
406    }
407}
408
409pub struct DeepSeekEventMapper {
410    tool_calls_by_index: HashMap<usize, RawToolCall>,
411}
412
413impl DeepSeekEventMapper {
414    pub fn new() -> Self {
415        Self {
416            tool_calls_by_index: HashMap::default(),
417        }
418    }
419
420    pub fn map_stream(
421        mut self,
422        events: Pin<Box<dyn Send + Stream<Item = Result<deepseek::StreamResponse>>>>,
423    ) -> impl Stream<Item = Result<LanguageModelCompletionEvent, LanguageModelCompletionError>>
424    {
425        events.flat_map(move |event| {
426            futures::stream::iter(match event {
427                Ok(event) => self.map_event(event),
428                Err(error) => vec![Err(LanguageModelCompletionError::from(error))],
429            })
430        })
431    }
432
433    pub fn map_event(
434        &mut self,
435        event: deepseek::StreamResponse,
436    ) -> Vec<Result<LanguageModelCompletionEvent, LanguageModelCompletionError>> {
437        let Some(choice) = event.choices.first() else {
438            return vec![Err(LanguageModelCompletionError::from(anyhow!(
439                "Response contained no choices"
440            )))];
441        };
442
443        let mut events = Vec::new();
444        if let Some(content) = choice.delta.content.clone() {
445            events.push(Ok(LanguageModelCompletionEvent::Text(content)));
446        }
447
448        if let Some(reasoning_content) = choice.delta.reasoning_content.clone() {
449            events.push(Ok(LanguageModelCompletionEvent::Thinking {
450                text: reasoning_content,
451                signature: None,
452            }));
453        }
454
455        if let Some(tool_calls) = choice.delta.tool_calls.as_ref() {
456            for tool_call in tool_calls {
457                let entry = self.tool_calls_by_index.entry(tool_call.index).or_default();
458
459                if let Some(tool_id) = tool_call.id.clone() {
460                    entry.id = tool_id;
461                }
462
463                if let Some(function) = tool_call.function.as_ref() {
464                    if let Some(name) = function.name.clone() {
465                        entry.name = name;
466                    }
467
468                    if let Some(arguments) = function.arguments.clone() {
469                        entry.arguments.push_str(&arguments);
470                    }
471                }
472            }
473        }
474
475        if let Some(usage) = event.usage {
476            events.push(Ok(LanguageModelCompletionEvent::UsageUpdate(TokenUsage {
477                input_tokens: usage.prompt_tokens,
478                output_tokens: usage.completion_tokens,
479                cache_creation_input_tokens: 0,
480                cache_read_input_tokens: 0,
481            })));
482        }
483
484        match choice.finish_reason.as_deref() {
485            Some("stop") => {
486                events.push(Ok(LanguageModelCompletionEvent::Stop(StopReason::EndTurn)));
487            }
488            Some("tool_calls") => {
489                events.extend(self.tool_calls_by_index.drain().map(|(_, tool_call)| {
490                    match parse_tool_arguments(&tool_call.arguments) {
491                        Ok(input) => Ok(LanguageModelCompletionEvent::ToolUse(
492                            LanguageModelToolUse {
493                                id: tool_call.id.clone().into(),
494                                name: tool_call.name.as_str().into(),
495                                is_input_complete: true,
496                                input,
497                                raw_input: tool_call.arguments.clone(),
498                                thought_signature: None,
499                            },
500                        )),
501                        Err(error) => Ok(LanguageModelCompletionEvent::ToolUseJsonParseError {
502                            id: tool_call.id.clone().into(),
503                            tool_name: tool_call.name.as_str().into(),
504                            raw_input: tool_call.arguments.into(),
505                            json_parse_error: error.to_string(),
506                        }),
507                    }
508                }));
509
510                events.push(Ok(LanguageModelCompletionEvent::Stop(StopReason::ToolUse)));
511            }
512            Some(stop_reason) => {
513                log::error!("Unexpected DeepSeek stop_reason: {stop_reason:?}",);
514                events.push(Ok(LanguageModelCompletionEvent::Stop(StopReason::EndTurn)));
515            }
516            None => {}
517        }
518
519        events
520    }
521}
522
523struct ConfigurationView {
524    api_key_editor: Entity<InputField>,
525    state: Entity<State>,
526    load_credentials_task: Option<Task<()>>,
527}
528
529impl ConfigurationView {
530    fn new(state: Entity<State>, window: &mut Window, cx: &mut Context<Self>) -> Self {
531        let api_key_editor =
532            cx.new(|cx| InputField::new(window, cx, "sk-00000000000000000000000000000000"));
533
534        cx.observe(&state, |_, _, cx| {
535            cx.notify();
536        })
537        .detach();
538
539        let load_credentials_task = Some(cx.spawn({
540            let state = state.clone();
541            async move |this, cx| {
542                if let Some(task) = Some(state.update(cx, |state, cx| state.authenticate(cx))) {
543                    let _ = task.await;
544                }
545
546                this.update(cx, |this, cx| {
547                    this.load_credentials_task = None;
548                    cx.notify();
549                })
550                .log_err();
551            }
552        }));
553
554        Self {
555            api_key_editor,
556            state,
557            load_credentials_task,
558        }
559    }
560
561    fn save_api_key(&mut self, _: &menu::Confirm, _window: &mut Window, cx: &mut Context<Self>) {
562        let api_key = self.api_key_editor.read(cx).text(cx).trim().to_string();
563        if api_key.is_empty() {
564            return;
565        }
566
567        let state = self.state.clone();
568        cx.spawn(async move |_, cx| {
569            state
570                .update(cx, |state, cx| state.set_api_key(Some(api_key), cx))
571                .await
572        })
573        .detach_and_log_err(cx);
574    }
575
576    fn reset_api_key(&mut self, window: &mut Window, cx: &mut Context<Self>) {
577        self.api_key_editor
578            .update(cx, |editor, cx| editor.set_text("", window, cx));
579
580        let state = self.state.clone();
581        cx.spawn(async move |_, cx| {
582            state
583                .update(cx, |state, cx| state.set_api_key(None, cx))
584                .await
585        })
586        .detach_and_log_err(cx);
587    }
588
589    fn should_render_editor(&self, cx: &mut Context<Self>) -> bool {
590        !self.state.read(cx).is_authenticated()
591    }
592}
593
594impl Render for ConfigurationView {
595    fn render(&mut self, _window: &mut Window, cx: &mut Context<Self>) -> impl IntoElement {
596        let env_var_set = self.state.read(cx).api_key_state.is_from_env_var();
597        let configured_card_label = if env_var_set {
598            format!("API key set in {API_KEY_ENV_VAR_NAME} environment variable")
599        } else {
600            let api_url = DeepSeekLanguageModelProvider::api_url(cx);
601            if api_url == DEEPSEEK_API_URL {
602                "API key configured".to_string()
603            } else {
604                format!("API key configured for {}", api_url)
605            }
606        };
607
608        if self.load_credentials_task.is_some() {
609            div()
610                .child(Label::new("Loading credentials..."))
611                .into_any_element()
612        } else if self.should_render_editor(cx) {
613            v_flex()
614                .size_full()
615                .on_action(cx.listener(Self::save_api_key))
616                .child(Label::new("To use DeepSeek in Zed, you need an API key:"))
617                .child(
618                    List::new()
619                        .child(
620                            ListBulletItem::new("")
621                                .child(Label::new("Get your API key from the"))
622                                .child(ButtonLink::new(
623                                    "DeepSeek console",
624                                    "https://platform.deepseek.com/api_keys",
625                                )),
626                        )
627                        .child(ListBulletItem::new(
628                            "Paste your API key below and hit enter to start using the assistant",
629                        )),
630                )
631                .child(self.api_key_editor.clone())
632                .child(
633                    Label::new(format!(
634                        "You can also set the {API_KEY_ENV_VAR_NAME} environment variable and restart Zed."
635                    ))
636                    .size(LabelSize::Small)
637                    .color(Color::Muted),
638                )
639                .into_any_element()
640        } else {
641            ConfiguredApiCard::new(configured_card_label)
642                .disabled(env_var_set)
643                .on_click(cx.listener(|this, _, window, cx| this.reset_api_key(window, cx)))
644                .into_any_element()
645        }
646    }
647}