MCP Starter Server
A minimal template for building AI assistant tools using the ModelContextProtocol.
MCP Starter Server
A minimal ModelContextProtocol server template for building AI assistant tools. This starter provides a basic structure for creating MCP tools that can be used with AI assistants like Claude.
Features
- Simple "hello world" tool example
- TypeScript + esbuild setup
- Development tools preconfigured
Setup to build and run with Claude
-
Download and install Claude desktop app from claude.ai/download
-
Clone the repo, install dependencies and build:
npm install
npm run build
- Configure Claude to use this MCP server. If this is your first MCP server, in the root of this project run:
echo '{
"mcpServers": {
"mcp-starter": {
"command": "node",
"args": ["'$PWD'/dist/index.cjs"]
}
}
}' > ~/Library/Application\ Support/Claude/claude_desktop_config.json
This should result in an entry in your claude_desktop_config.json like:
"mcpServers": {
"mcp-starter": {
"command": "node",
"args": ["/Users/matt/code/mcp-starter/dist/index.cjs"]
}
}
If you have existing MCP servers, add the mcp-starter block to your existing config. It's an important detail that the args is the path to <path_to_repo_on_your_machine>/mcp-starter/dist/index.cjs.
-
Restart Claude Desktop.
-
Look for the hammer icon with the number of available tools in Claude's interface to confirm the server is running.
-
If this is all working, you should be able to develop your MCP server using
npm run devand test it in Claude. You'll need to restart Claude each time to restart the MCP server.
Developing with Inspector
For development and debugging purposes, you can use the MCP Inspector tool. The Inspector provides a visual interface for testing and monitoring MCP server interactions.
Visit the Inspector documentation for detailed setup instructions.
To test locally with Inspector:
npm run inspect
To build on file changes run:
npm run watch
Or run both the watcher and inspector:
npm run dev
Publishing
Once you're ready to distribute your server, it's simple!
-
Set up an NPM account.
-
Run
npm publish. This will publish a package using the project name inpackage.json -
Once published, others can install the server with a config entry like:
"mcpServers": {
"<your-package-name>": {
"command": "npx",
"args": ["<your-package-name>"]
}
}
Available Tools
The server provides:
hello_tool: A simple example tool that takes a name parameter and returns a greeting
Creating New Tools
To add new tools:
- Define the tool schema in
index.ts - Add it to the tools array in the
ListToolsRequestSchemahandler - Add the implementation in the
CallToolRequestSchemahandler
See the hello_tool implementation as an example.
İlgili Sunucular
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Python REPL
A Python REPL with persistent sessions and automatic dependency management using uv.
E2B
Run code in secure sandboxes hosted by E2B
swarm.at
Settlement protocol that validates agent actions before execution. Guards destructive operations with hash-chained ledger receipts, trust tiers, and 31 workflow blueprints.
Ghost MCP
An MCP server for the Ghost blogging platform with Server-Sent Events (SSE) transport support.
iFlytek Workflow MCP Server
An MCP server for executing iFlytek workflows through MCP tools.
MCP-S Gateway
A secure, open-source OAuth gateway for MCP authentication.
Autoconsent MCP
A server for browser automation to create and test Autoconsent rules for web page consent management.
ActionKit MCP Starter
A demonstration server for ActionKit, providing access to Slack actions via Claude Desktop.
MCP Server
Automate data science stages using your own CSV data files.
Unified.to MCP Server
Unified MCP Server is a remote MCP connector for AI agents and vertical AI products that provides access to 22,000+ authorized SaaS tools across 400+ integrations and 24 categories directly inside LLMs.