DeepSeek MCP Server
A server for code generation and completion using the DeepSeek API.
DeepSeek MCP Server
An MCP server implementation that provides code generation and completion capabilities using the DeepSeek API, with support for tool chaining and cost optimization.
Features
- Code generation with language-specific support
- Code completion with context awareness
- Code optimization with multiple targets
- Tool chaining for complex operations
- Built-in caching for cost optimization
- TypeScript implementation with full type safety
Tools
1. generate_code
Generate code using DeepSeek API with language-specific support.
{
"name": "generate_code",
"params": {
"prompt": "Write a function that sorts an array",
"language": "typescript",
"temperature": 0.7
}
}
2. complete_code
Get intelligent code completions based on existing context.
{
"name": "complete_code",
"params": {
"code": "function processData(data) {",
"prompt": "Add input validation and error handling",
"temperature": 0.7
}
}
3. optimize_code
Optimize existing code for performance, memory usage, or readability.
{
"name": "optimize_code",
"params": {
"code": "your code here",
"target": "performance"
}
}
4. execute_chain
Execute a chain of tools in sequence, with context passing between steps.
{
"name": "execute_chain",
"params": {
"steps": [
{
"toolName": "generate_code",
"params": {
"prompt": "Create a REST API endpoint",
"language": "typescript"
}
},
{
"toolName": "optimize_code",
"params": {
"target": "performance"
}
}
]
}
}
Installation
- Clone the repository
- Install dependencies:
npm install
- Build the project:
npm run build
- Configure your DeepSeek API key in the MCP settings file:
{
"mcpServers": {
"deepseek": {
"command": "node",
"args": ["/path/to/deepseek-mcp/build/index.js"],
"env": {
"DEEPSEEK_API_KEY": "your-api-key"
}
}
}
}
Usage
The server can be used with any MCP-compatible client. Here's an example using the MCP CLI:
mcp use deepseek generate_code --params '{"prompt": "Write a hello world program", "language": "python"}'
Tool Chaining
Tool chaining allows you to combine multiple operations into a single workflow. Each tool in the chain can access the results of previous tools through the chain context.
Example chain:
- Generate initial code
- Complete the code with error handling
- Optimize the final result
{
"steps": [
{
"toolName": "generate_code",
"params": {
"prompt": "Create a user authentication function",
"language": "typescript"
}
},
{
"toolName": "complete_code",
"params": {
"prompt": "Add input validation and error handling"
}
},
{
"toolName": "optimize_code",
"params": {
"target": "security"
}
}
]
}
Cost Optimization
The server implements several strategies to optimize API costs:
- Request caching with TTL
- Chain result caching
- Smart prompt construction
- Metadata tracking for usage analysis
Development
To start development:
npm run dev
To clean and rebuild:
npm run rebuild
Requirements
- Node.js >= 18.0.0
- DeepSeek API key
- MCP-compatible client
License
ISC
เซิร์ฟเวอร์ที่เกี่ยวข้อง
Scout Monitoring MCP
ผู้สนับสนุนPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
ผู้สนับสนุนAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
MCP-Inscription Server
Interact with Ordinals Inscriptions and display content from transactions.
Unity MCP
Perform actions in the Unity Editor for game development using AI clients.
MCP LaTeX Server
Create, edit, and manage LaTeX files. Requires an external LaTeX distribution like MiKTeX, TeX Live, or MacTeX.
Gateway MCP Server
A gateway server that intelligently routes MCP requests to multiple backend servers based on external configuration.
Remote MCP Server (Authless)
An example of a remote MCP server deployable on Cloudflare Workers, without authentication.
MCP Python REPL Server
An interactive Python REPL server with built-in support for the uv package manager and virtual environments.
Kibana MCP Server
Access and interact with your Kibana instance using natural language or programmatic requests.
ZIN MCP Client
A lightweight CLI client that bridges local LLMs running on Ollama with STDIO MCP Servers.
Raysurfer Code Caching
MCP server for LLM output caching and reuse. Caches and retrieves code from prior AI agent executions, delivering cached outputs up to 30x faster.
Shipyard
The Shipyard CLI provides an MCP server for agents to manage Shipyard environments directly: by pulling logs, comparing branches, running tests, and stopping/starting environments..