Deepseek Thinker
Provides Deepseek's reasoning capabilities to AI clients, supporting both the Deepseek API and local Ollama server modes.
Deepseek Thinker MCP Server
A MCP (Model Context Protocol) provider Deepseek reasoning content to MCP-enabled AI Clients, like Claude Desktop. Supports access to Deepseek's thought processes from the Deepseek API service or from a local Ollama server.
Core Features
-
🤖 Dual Mode Support
- OpenAI API mode support
- Ollama local mode support
-
🎯 Focused Reasoning
- Captures Deepseek's thinking process
- Provides reasoning output
Available Tools
get-deepseek-thinker
- Description: Perform reasoning using the Deepseek model
- Input Parameters:
originPrompt(string): User's original prompt
- Returns: Structured text response containing the reasoning process
Environment Configuration
OpenAI API Mode
Set the following environment variables:
API_KEY=<Your OpenAI API Key>
BASE_URL=<API Base URL>
Ollama Mode
Set the following environment variable:
USE_OLLAMA=true
Usage
Integration with AI Client, like Claude Desktop
Add the following configuration to your claude_desktop_config.json:
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Using Ollama Mode
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"USE_OLLAMA": "true"
}
}
}
}
Local Server Configuration
{
"mcpServers": {
"deepseek-thinker": {
"command": "node",
"args": [
"/your-path/deepseek-thinker-mcp/build/index.js"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Development Setup
# Install dependencies
npm install
# Build project
npm run build
# Run service
node build/index.js
FAQ
Response like this: "MCP error -32001: Request timed out"
This error occurs when the Deepseek API response is too slow or when the reasoning content output is too long, causing the MCP server to timeout.
Tech Stack
- TypeScript
- @modelcontextprotocol/sdk
- OpenAI API
- Ollama
- Zod (parameter validation)
License
This project is licensed under the MIT License. See the LICENSE file for details.
相关服务器
Alpha Vantage MCP Server
赞助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
PyMOL-MCP
Enables conversational structural biology, molecular visualization, and analysis in PyMOL through natural language.
CIE - Code Intelligence Engine
Local code analysis MCP server with 25+ tools: semantic search, call graph tracing, dependency analysis, and symbol navigation. Built with Tree-sitter and CozoDB. Supports Go, Python, JS, TS.
gluestack-ui MCP Server
An MCP server tailored for React Native–first development using Gluestack UI
Airflow MCP Server
MCP server for Airflow
MCP Invoice Express
An MCP server for integrating with the InvoiceExpress API.
TIA-Portal MCP-Server
A VS Code extension to connect and interact with Siemens TIA Portal projects directly from the editor.
Eterna MCP
Managed MCP server for Bybit perpetual futures trading. Isolated sub-accounts, built-in risk management, 12 trading tools.
Enkrypt AI
Integrate red-teaming, prompt auditing, and AI safety analysis into any MCP-compatible client.
Lucide Icons
Provides access to the Lucide icon library for use in LLM and agentic applications.
MapleStory MCP Server
Access NEXON MapleStory open API data for character info, union details, guild data, rankings, and game mechanics.