MCP Documentation Server
Integrates LLM applications with documentation sources using the Model Context Protocol.
MCP Documentation Server
A customized version of the MCP documentation server that enables integration between LLM applications (like Cursor, Claude Desktop, Windsurf) and documentation sources via the Model Context Protocol.
Overview
This server provides MCP host applications with:
- Access to specific documentation files (langgraph.txt and mcp.txt)
- Tools to fetch documentation from URLs within those files
Supported Documentation
Currently set up for:
- LangGraph Documentation (from https://raw.githubusercontent.com/esakrissa/mcp-doc/main/docs/langgraph.txt)
- MCP Documentation (from https://raw.githubusercontent.com/esakrissa/mcp-doc/main/docs/mcp.txt)
Quick Start
Setup and Run
# Clone the repository
git clone https://github.com/esakrissa/mcp-doc.git
cd mcp-doc
# Create and activate a virtual environment
python -m venv venv
source venv/bin/activate # On Windows: venv\Scripts\activate
# Install the package in development mode
pip install -e .
Running the Server
You can run the server using the installed command:
# Run the server with the config file
mcpdoc \
--json config.json \
--transport sse \
--port 8082 \
--host localhost
Or if you prefer using UV:
# Install uv (if not already installed)
curl -LsSf https://astral.sh/uv/install.sh | sh
# Run the server with UV
uvx --from mcpdoc mcpdoc \
--json config.json \
--transport sse \
--port 8082 \
--host localhost
IDE Integration
Cursor
Add to ~/.cursor/mcp.json
{
"mcpServers": {
"mcp-doc": {
"command": "uvx",
"args": [
"--from",
"mcpdoc",
"mcpdoc",
"--urls",
"LangGraph:https://raw.githubusercontent.com/esakrissa/mcp-doc/main/docs/langgraph.txt",
"ModelContextProtocol:https://raw.githubusercontent.com/esakrissa/mcp-doc/main/docs/mcp.txt",
"--allowed-domains",
"*",
"--transport",
"stdio"
]
}
}
}
Then add these instructions to Cursor's Custom Instructions:
for ANY question about LangGraph and Model Context Protocol (MCP), use the mcp-doc server to help answer --
+ call list_doc_sources tool to get the available documentation files
+ call fetch_docs tool to read the langgraph.txt or mcp.txt file
+ reflect on the urls in langgraph.txt or mcp.txt
+ reflect on the input question
+ call fetch_docs on any urls relevant to the question
+ use this to answer the question
To test if the integration is working, ask Cursor a question about LangGraph or MCP, and check if it uses the documentation server tools to fetch information.
Security Note
For security reasons, strict domain access controls are implemented:
- Remote documentation files: Only the specific domain is automatically allowed
- Local documentation files: No domains are automatically allowed
- Use
--allowed-domainsto explicitly add domains or--allowed-domains '*'to allow all (use with caution)
References
This project is based on the original mcpdoc by LangChain AI, modified to provide focused documentation access for LangGraph and MCP.
相关服务器
Alpha Vantage MCP Server
赞助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
MCP Prompt Optimizer
Optimize prompts with research-backed strategies for 15-74% performance improvements.
SACL MCP Server
A framework for bias-aware code retrieval using semantic-augmented reranking and localization.
AppleScript MCP
Execute AppleScript on macOS
MCP Create Server
A service for dynamically creating, running, and managing Model Context Protocol (MCP) servers.
MCP-Inscription Server
Interact with Ordinals Inscriptions and display content from transactions.
Jupyter Notebook MCP Server
Interact with Jupyter notebooks, allowing for code execution, cell manipulation, and notebook management.
Authless Remote MCP Server
A remote MCP server deployable on Cloudflare Workers that does not require authentication.
memtrace
Memtrace gives AI coding agents structural memory — your codebase as a live knowledge graph so agents stop re-deriving code structure from scratch and start reasoning from fact.
Pinelabs MCP Server
The Pine Labs Online MCP Server implements the Model Context Protocol (MCP) to enable seamless integration between Pine Labs’ online payment APIs and AI tools. It allows AI assistants to perform Pine Labs Online API operations, empowering developers to build intelligent, AI-driven payment applications with ease.
McpDocServer
An MCP-based server for searching and retrieving development framework documentation, supporting crawling and local file loading.