MCP Server Collector
Discovers and collects MCP servers from the internet.
mcp-server-collector MCP server
A MCP Server used to collect MCP Servers over the internet.
Components
Resources
No resources yet.
Prompts
No prompts yet.
Tools
The server implements 3 tools:
- extract-mcp-servers-from-url: Extracts MCP Servers from given URL.
- Takes "url" as required string argument
- extract-mcp-servers-from-content: Extracts MCP Servers from given content.
- Takes "content" as required string argument
- submit-mcp-server: Submits a MCP Server to the MCP Server Directory like mcp.so.
- Takes "url" as required string argument and "avatar_url" as optional string argument
Configuration
.env file is required to be set up.
OPENAI_API_KEY="sk-xxx"
OPENAI_BASE_URL="https://api.openai.com/v1"
OPENAI_MODEL="gpt-4o-mini"
MCP_SERVER_SUBMIT_URL="https://mcp.so/api/submit-project"
Quickstart
Install
Claude Desktop
On MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
Development/Unpublished Servers Configuration
``` "mcpServers": { "fetch": { "command": "uvx", "args": ["mcp-server-fetch"] }, "mcp-server-collector": { "command": "uv", "args": [ "--directory", "path-to/mcp-server-collector", "run", "mcp-server-collector" ], "env": { "OPENAI_API_KEY": "sk-xxx", "OPENAI_BASE_URL": "https://api.openai.com/v1", "OPENAI_MODEL": "gpt-4o-mini", "MCP_SERVER_SUBMIT_URL": "https://mcp.so/api/submit-project" } } } ```Published Servers Configuration
``` "mcpServers": { "fetch": { "command": "uvx", "args": ["mcp-server-fetch"] }, "mcp-server-collector": { "command": "uvx", "args": [ "mcp-server-collector" ], "env": { "OPENAI_API_KEY": "sk-xxx", "OPENAI_BASE_URL": "https://api.openai.com/v1", "OPENAI_MODEL": "gpt-4o-mini", "MCP_SERVER_SUBMIT_URL": "https://mcp.so/api/submit-project" } } } ```Development
Building and Publishing
To prepare the package for distribution:
- Sync dependencies and update lockfile:
uv sync
- Build package distributions:
uv build
This will create source and wheel distributions in the dist/ directory.
- Publish to PyPI:
uv publish
Note: You'll need to set PyPI credentials via environment variables or command flags:
- Token:
--tokenorUV_PUBLISH_TOKEN - Or username/password:
--username/UV_PUBLISH_USERNAMEand--password/UV_PUBLISH_PASSWORD
Debugging
Since MCP servers run over stdio, debugging can be challenging. For the best debugging experience, we strongly recommend using the MCP Inspector.
You can launch the MCP Inspector via npm with this command:
npx @modelcontextprotocol/inspector uv --directory path-to/mcp-server-collector run mcp-server-collector
Upon launching, the Inspector will display a URL that you can access in your browser to begin debugging.
Community
About the author
Verwandte Server
Bright Data
SponsorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
ScraperCity
B2B lead generation MCP server - Apollo, Google Maps, email finder, skip trace, and 15+ more tools.
Fetch
Fetch web content as HTML, JSON, plain text, or Markdown.
Puppeteer Real Browser
Enables powerful, detection-resistant browser automation for AI assistants using puppeteer-real-browser.
Website Snapshot
A MCP server that provides comprehensive website snapshot capabilities using Playwright. This server enables LLMs to capture and analyze web pages through structured accessibility snapshots, network monitoring, and console message collection.
AgentQL
Enable AI agents to get structured data from unstructured web with AgentQL.
SnapSender
Capture any website as PNG, JPEG, WebP, or PDF with a single tool call.
Driflyte
The Driflyte MCP Server exposes tools that allow AI assistants to query and retrieve topic-specific knowledge from recursively crawled and indexed web pages.
Conduit
Headless browser with SHA-256 hash-chained audit trails and Ed25519-signed proof bundles. MCP server for AI agents.
Scrapezy
Extract structured data from websites using the Scrapezy API.
B2Proxy
1GB Free Trial, World's Leading Proxy Service Platform, Efficient Data Collection