Everything MCP Server
A test server that demonstrates all features of the MCP protocol, including prompts, tools, resources, and sampling.
Everything MCP Server
This MCP server attempts to exercise all the features of the MCP protocol. It is not intended to be a useful server, but rather a test server for builders of MCP clients. It implements prompts, tools, resources, sampling, and more to showcase MCP capabilities.
Components
Tools
-
echo- Simple tool to echo back input messages
- Input:
message(string): Message to echo back
- Returns: Text content with echoed message
-
add- Adds two numbers together
- Inputs:
a(number): First numberb(number): Second number
- Returns: Text result of the addition
-
longRunningOperation- Demonstrates progress notifications for long operations
- Inputs:
duration(number, default: 10): Duration in secondssteps(number, default: 5): Number of progress steps
- Returns: Completion message with duration and steps
- Sends progress notifications during execution
-
sampleLLM- Demonstrates LLM sampling capability using MCP sampling feature
- Inputs:
prompt(string): The prompt to send to the LLMmaxTokens(number, default: 100): Maximum tokens to generate
- Returns: Generated LLM response
-
getTinyImage- Returns a small test image
- No inputs required
- Returns: Base64 encoded PNG image data
-
printEnv- Prints all environment variables
- Useful for debugging MCP server configuration
- No inputs required
- Returns: JSON string of all environment variables
-
annotatedMessage- Demonstrates how annotations can be used to provide metadata about content
- Inputs:
messageType(enum: "error" | "success" | "debug"): Type of message to demonstrate different annotation patternsincludeImage(boolean, default: false): Whether to include an example image
- Returns: Content with varying annotations:
- Error messages: High priority (1.0), visible to both user and assistant
- Success messages: Medium priority (0.7), user-focused
- Debug messages: Low priority (0.3), assistant-focused
- Optional image: Medium priority (0.5), user-focused
- Example annotations:
{ "priority": 1.0, "audience": ["user", "assistant"] }
-
getResourceReference- Returns a resource reference that can be used by MCP clients
- Inputs:
resourceId(number, 1-100): ID of the resource to reference
- Returns: A resource reference with:
- Text introduction
- Embedded resource with
type: "resource" - Text instruction for using the resource URI
Resources
The server provides 100 test resources in two formats:
-
Even numbered resources:
- Plaintext format
- URI pattern:
test://static/resource/{even_number} - Content: Simple text description
-
Odd numbered resources:
- Binary blob format
- URI pattern:
test://static/resource/{odd_number} - Content: Base64 encoded binary data
Resource features:
- Supports pagination (10 items per page)
- Allows subscribing to resource updates
- Demonstrates resource templates
- Auto-updates subscribed resources every 5 seconds
Prompts
-
simple_prompt- Basic prompt without arguments
- Returns: Single message exchange
-
complex_prompt- Advanced prompt demonstrating argument handling
- Required arguments:
temperature(number): Temperature setting
- Optional arguments:
style(string): Output style preference
- Returns: Multi-turn conversation with images
-
resource_prompt- Demonstrates embedding resource references in prompts
- Required arguments:
resourceId(number): ID of the resource to embed (1-100)
- Returns: Multi-turn conversation with an embedded resource reference
- Shows how to include resources directly in prompt messages
Logging
The server sends random-leveled log messages every 15 seconds, e.g.:
{
"method": "notifications/message",
"params": {
"level": "info",
"data": "Info-level message"
}
}
Usage with Claude Desktop (uses stdio Transport)
Add to your claude_desktop_config.json:
{
"mcpServers": {
"everything": {
"command": "npx",
"args": [
"-y",
"@modelcontextprotocol/server-everything"
]
}
}
}
Usage with VS Code
For quick installation, use of of the one-click install buttons below...
For manual installation, add the following JSON block to your User Settings (JSON) file in VS Code. You can do this by pressing Ctrl + Shift + P and typing Preferences: Open User Settings (JSON).
Optionally, you can add it to a file called .vscode/mcp.json in your workspace. This will allow you to share the configuration with others.
Note that the
mcpkey is not needed in the.vscode/mcp.jsonfile.
NPX
{
"mcp": {
"servers": {
"everything": {
"command": "npx",
"args": ["-y", "@modelcontextprotocol/server-everything"]
}
}
}
}
Running from source with HTTP+SSE Transport (deprecated as of 2025-03-26)
cd src/everything
npm install
npm run start:sse
Run from source with Streamable HTTP Transport
cd src/everything
npm install
npm run start:streamableHttp
Running as an installed package
Install
npm install -g @modelcontextprotocol/server-everything@latest
Run the default (stdio) server
npx @modelcontextprotocol/server-everything
Or specify stdio explicitly
npx @modelcontextprotocol/server-everything stdio
Run the SSE server
npx @modelcontextprotocol/server-everything sse
Run the streamable HTTP server
npx @modelcontextprotocol/server-everything streamableHttp
Verwandte Server
Scout Monitoring MCP
SponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
SponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
AgentChatBus
AgentChatBus is a persistent AI communication bus that lets multiple independent AI Agents chat, collaborate, and delegate tasks — across terminals, across IDEs, and across frameworks.
Criage MCP Server
An MCP server for the Criage package manager, providing full client functionality via the MCP protocol.
Mermaid
Generate mermaid diagram and chart with AI MCP dynamically.
CrowdCent MCP Server
Integrates with the CrowdCent Challenge API, allowing AI assistants to manage prediction challenges, datasets, and submissions.
Blend MCP
An AI gateway for the Blend Protocol on Stellar, enabling DeFi actions like lending, borrowing, and pool creation through AI assistants or apps.
Advanced Gemini MCP Server
An open-source MCP server that integrates with Google's Gemini AI. Requires a Google Gemini API key.
SpecLock
AI constraint engine — persistent memory + active enforcement. Stops AI from breaking locked code. Semantic conflict detection, file-level guards, session continuity. 19 MCP tools.
MetaTrader 4
Integrate with the MetaTrader 4 trading platform to access trading functions and data via an HTTP bridge and Expert Advisor.
MCP Image Extractor
Extracts images from files, URLs, or base64 strings and converts them to base64 for LLM analysis.
ndlovu-code-reviewer
Manual code reviews are time-consuming and often miss the opportunity to combine static analysis with contextual, human-friendly feedback. This project was created to experiment with MCP tooling that gives AI assistants access to a purpose-built reviewer. Uses the Gemini cli application to process the reviews at this time and linting only for typescript/javascript apps at the moment. Will add API based calls to LLM's in the future and expand linting abilities. It's also cheaper than using coderabbit ;)