@@ -65,11 +65,9 @@ Crush can use LSPs for additional context to help inform its decisions, just lik
{
"lsp": {
"go": {
- "disabled": false,
"command": "gopls"
},
"typescript": {
- "disabled": false,
"command": "typescript-language-server",
"args": ["--stdio"]
},
@@ -80,83 +78,30 @@ Crush can use LSPs for additional context to help inform its decisions, just lik
}
```
-### Amazon Bedrock
-
-To use AWS Bedrock with Claude models, configure your AWS credentials and region:
-
-```json
-{
- "providers": [
- {
- "id": "bedrock",
- "provider_type": "bedrock",
- "extra_params": {
- "region": "us-east-1"
- }
- }
- ]
-}
-```
-
-Bedrock uses your AWS credentials from environment variables (`AWS_ACCESS_KEY_ID`, `AWS_SECRET_ACCESS_KEY`) or AWS credential profiles. The region can be specified in the configuration or via the `AWS_REGION` environment variable.
-
-### Google Vertex AI
-
-For Google Cloud Vertex AI with Gemini models, configure your project and location:
-
-```json
-{
- "providers": [
- {
- "id": "vertexai",
- "provider_type": "vertexai",
- "extra_headers": {
- "project": "your-gcp-project-id",
- "location": "us-central1"
- }
- }
- ]
-}
-```
-
-Vertex AI uses Google Cloud authentication. Ensure you have the `GOOGLE_APPLICATION_CREDENTIALS` environment variable set or are authenticated via `gcloud auth application-default login`.
-
### OpenAI-Compatible APIs
-Crush supports all OpenAI-compatible APIs, including local models via Ollama:
+Crush supports all OpenAI-compatible APIs. Here's an example configuration for Deepseek, which uses an OpenAI-compatible API. Don't forget to set `DEEPSEEK_API_KEY` in your environment.
```json
{
- "providers": [
- {
- "id": "ollama",
+ "providers": {
+ "deepseek": {
"provider_type": "openai",
- "base_url": "http://localhost:11434/v1",
+ "base_url": "https://api.deepseek.com/v1",
"models": [
{
- "id": "llama3.2:3b",
- "name": "Llama 3.2 3B",
- "context_window": 8192,
- "default_max_tokens": 4096
+ "id": "deepseek-chat",
+ "model": "Deepseek V3",
+ "cost_per_1m_in": 0.27,
+ "cost_per_1m_out": 1.1,
+ "cost_per_1m_in_cached": 0.07,
+ "cost_per_1m_out_cached": 1.1,
+ "context_window": 64000,
+ "default_max_tokens": 5000
}
]
}
- ]
-}
-```
-
-For other OpenAI-compatible providers, adjust the `base_url` and provide an `api_key` if required:
-
-```json
-{
- "providers": [
- {
- "id": "custom-openai",
- "provider_type": "openai",
- "base_url": "https://api.example.com/v1",
- "api_key": "your-api-key"
- }
- ]
+ }
}
```