Everything
Reference / test server with prompts, resources, and tools
Everything MCP Server
Architecture | Project Structure | Startup Process | Server Features | Extension Points | How It Works
This MCP server attempts to exercise all the features of the MCP protocol. It is not intended to be a useful server, but rather a test server for builders of MCP clients. It implements prompts, tools, resources, sampling, and more to showcase MCP capabilities.
Tools, Resources, Prompts, and Other Features
A complete list of the registered MCP primitives and other protocol features demonstrated can be found in the Server Features document.
Usage with Claude Desktop (uses stdio Transport)
Add to your claude_desktop_config.json:
{
"mcpServers": {
"everything": {
"command": "npx",
"args": [
"-y",
"@modelcontextprotocol/server-everything"
]
}
}
}
On Windows, use cmd /c to launch npx:
{
"mcpServers": {
"everything": {
"command": "cmd",
"args": [
"/c",
"npx",
"-y",
"@modelcontextprotocol/server-everything"
]
}
}
}
Usage with VS Code
For quick installation, use one of the one-click install buttons below...
For manual installation, you can configure the MCP server using one of these methods:
Method 1: User Configuration (Recommended)
Add the configuration to your user-level MCP configuration file. Open the Command Palette (Ctrl + Shift + P) and run MCP: Open User Configuration. This will open your user mcp.json file where you can add the server configuration.
Method 2: Workspace Configuration
Alternatively, you can add the configuration to a file called .vscode/mcp.json in your workspace. This will allow you to share the configuration with others.
For more details about MCP configuration in VS Code, see the official VS Code MCP documentation.
NPX
{
"servers": {
"everything": {
"command": "npx",
"args": ["-y", "@modelcontextprotocol/server-everything"]
}
}
}
On Windows, use:
{
"servers": {
"everything": {
"command": "cmd",
"args": ["/c", "npx", "-y", "@modelcontextprotocol/server-everything"]
}
}
}
Running from source with HTTP+SSE Transport (deprecated as of 2025-03-26)
cd src/everything
npm install
npm run start:sse
Run from source with Streamable HTTP Transport
cd src/everything
npm install
npm run start:streamableHttp
Running as an installed package
Install
npm install -g @modelcontextprotocol/server-everything@latest
Run the default (stdio) server
npx @modelcontextprotocol/server-everything
Or specify stdio explicitly
npx @modelcontextprotocol/server-everything stdio
Run the SSE server
npx @modelcontextprotocol/server-everything sse
Run the streamable HTTP server
npx @modelcontextprotocol/server-everything streamableHttp
相關伺服器
Alpha Vantage MCP Server
贊助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Authenticator App MCP Server
A secure MCP server for AI agents to interact with the Authenticator App for 2FA codes and passwords.
Slowtime MCP Server
A server for secure time-based operations, featuring timing attack protection and timelock encryption.
Code Editor
Enables AI assistants to write, edit, and manage code files directly in a specified directory, respecting .gitignore patterns.
Gemini CLI MCP Server
An MCP server and CLI wrapper for Google's Gemini CLI, featuring OAuth authentication support.
OpenAPI Schema
Exposes OpenAPI schema information to Large Language Models (LLMs). The server loads OpenAPI schema files specified via command line.
Skills-ContextManager
Don’t pollute your AI agent’s context with 1,000 skills. Use Skills-ContextManager, a self-hosted web UI for managing AI skills and workflows by providing skills to an AI agent via MCP only when needed. Simply add skills to your library and enable or disable them with a toggle. Choose whether a skill is always loaded into context or dynamically activated when the AI agent determines it’s needed.
LLMS.TXT Documentation Server
Access and read llms.txt documentation files for various Large Language Models.
POX MCP Server
An MCP server for the POX SDN controller, enabling network control, management, and analysis using Python and OpenFlow.
Code Assistant
A server for code modification and generation using Large Language Models.
Engram
Prevents regression by providing Blast Radius data to AI based on your git history