Deepseek Thinker
Provides Deepseek's reasoning capabilities to AI clients, supporting both the Deepseek API and local Ollama server modes.
Deepseek Thinker MCP Server
A MCP (Model Context Protocol) provider Deepseek reasoning content to MCP-enabled AI Clients, like Claude Desktop. Supports access to Deepseek's thought processes from the Deepseek API service or from a local Ollama server.
Core Features
-
🤖 Dual Mode Support
- OpenAI API mode support
- Ollama local mode support
-
🎯 Focused Reasoning
- Captures Deepseek's thinking process
- Provides reasoning output
Available Tools
get-deepseek-thinker
- Description: Perform reasoning using the Deepseek model
- Input Parameters:
originPrompt(string): User's original prompt
- Returns: Structured text response containing the reasoning process
Environment Configuration
OpenAI API Mode
Set the following environment variables:
API_KEY=<Your OpenAI API Key>
BASE_URL=<API Base URL>
Ollama Mode
Set the following environment variable:
USE_OLLAMA=true
Usage
Integration with AI Client, like Claude Desktop
Add the following configuration to your claude_desktop_config.json:
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Using Ollama Mode
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"USE_OLLAMA": "true"
}
}
}
}
Local Server Configuration
{
"mcpServers": {
"deepseek-thinker": {
"command": "node",
"args": [
"/your-path/deepseek-thinker-mcp/build/index.js"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Development Setup
# Install dependencies
npm install
# Build project
npm run build
# Run service
node build/index.js
FAQ
Response like this: "MCP error -32001: Request timed out"
This error occurs when the Deepseek API response is too slow or when the reasoning content output is too long, causing the MCP server to timeout.
Tech Stack
- TypeScript
- @modelcontextprotocol/sdk
- OpenAI API
- Ollama
- Zod (parameter validation)
License
This project is licensed under the MIT License. See the LICENSE file for details.
İlgili Sunucular
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Mentor MCP
Provides AI-powered mentorship to LLM agents for tasks like code review, design critique, and brainstorming, using the Deepseek API.
MCP Jupiter
Access Jupiter's swap API on the Solana blockchain.
LLMling
An MCP server with an LLMling backend that uses YAML files to configure LLM applications.
NEI MCP Server
A server for interacting with the NEI platform to query project resources.
Solana Dev MCP
An MCP server for Solana development providing basic RPC methods and helpful prompts.
Code Understanding
Analyzes local and remote GitHub repositories to provide code understanding and context generation, including structure analysis, file identification, and semantic mapping.
MediaWiki
Interact with MediaWiki installations through the MediaWiki API as a bot user.
Azure DevOps MCP
Integrates with Azure DevOps, allowing interaction with its services. Requires a Personal Access Token (PAT) for authentication.
Pistachio MobileDev MCP
Android + iOS development for non-technical users
Model Context Protocol servers
A collection of reference implementations for the Model Context Protocol (MCP), showcasing servers implemented with TypeScript and Python SDKs.