Deepseek
Integrates Deepseek models with any MCP-compatible client, such as Claude Desktop.
MCP Server for Deepseek Integration
This repository contains a Model Control Protocol (MCP) server implementation that allows Claude Desktop to use Deepseek models running in Docker.
Prerequisites
- Docker
- Python 3.11 or later
- A Deepseek API key
- Claude Desktop
Installation
- Clone the repository:
git clone https://github.com/vincentf305/mcp-server-deepseek.git
cd mcp-server-deepseek
- Install dependencies:
pip install -r requirements.txt
Setup Environment Variables
Create a .env file in the root directory of the project and add the following environment variable:
DEEPSEEK_API_KEY=your_api_key_here
Make sure to replace your_api_key_here with your actual Deepseek API key.
Running the Server
Using Docker
- Build the Docker image:
docker build -t mcp_server_deepseek .
- Run the container:
docker run -d \
--name mcp-server-deepseek \
-p 8765:8765 \
-e DEEPSEEK_API_KEY=your_api_key_here \
mcp-server-deepseek
Running Locally
python -m mcp_server_deepseek.server
Usage with Claude Desktop
-
Ensure you have a Deepseek API key
-
Add the following to your Claude Desktop configuration (claude_desktop_config.json):
{
"mcpServers": {
"deepseek-server": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"DEEPSEEK_API_KEY",
"mcp_server_deepseek"
],
"env": {
"DEEPSEEK_API_KEY": "your_api_key_here"
}
}
}
}
- Restart Claude Desktop to load the new configuration
Contributing
- Fork the repository
- Create your feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add some amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Create a Pull Request
License
MIT License - see the LICENSE file for details
Servidores relacionados
GPT-SoVITS MCP Server
An MCP server for GPT-SoVITS, providing text-to-speech synthesis, voice cloning, and multi-language support.
Remote MCP Server on Cloudflare
A remote MCP server deployable on Cloudflare Workers with OAuth login support, using Cloudflare KV for data storage.
Joe Sandbox
Analyze files and extract Indicators of Compromise (IOCs) by interacting with the Joe Sandbox Cloud service.
Fleet-MCP
Enables AI assistants to interact with Fleet Device Management for device management, security monitoring, and compliance enforcement.
Lumenore-MCP
A Model Context Protocol (MCP) server that provides AI assistants with access to Lumenore's analytics and natural language query capabilities. Built with FastMCP and Python 3.13 for scalable data analytics integration.
DYPAI
Deploy production backends, APIs, cron jobs and automations from any AI assistant. Database, auth, storage and 24+ integrations included.
Aiven
Navigate your Aiven projects and interact with the PostgreSQL®, Apache Kafka®, ClickHouse® and OpenSearch® services
GCP MCP Server
Manage and interact with Google Cloud Platform (GCP) resources through an AI assistant.
Azure Pricing MCP Server
Query Azure retail pricing information using the Azure Retail Prices API.
Whissle AI
Provides speech-to-text, diarization, translation, and text summarization via the Whissle AI API.