MCP Hangar
Kubernetes-native registry for managing multiple MCP servers with lazy loading, health monitoring, and RBAC
MCP Hangar
Parallel MCP tool execution. One interface. 50x faster.
The Problem
Your AI agent calls 5 tools sequentially. Each takes 200ms. That's 1 second of waiting.
Hangar runs them in parallel. 200ms total. Same results, 50x faster.
Quick Start
30 seconds to working MCP providers:
# Install, configure, and start - zero interaction
curl -sSL https://mcp-hangar.io/install.sh | bash && mcp-hangar init -y && mcp-hangar serve
That's it. Filesystem, fetch, and memory providers are now available to Claude.
What just happened?
- Install - Downloaded and installed
mcp-hangarvia pip/uv - Init - Created
~/.config/mcp-hangar/config.yamlwith starter providers - Serve - Started the MCP server (stdio mode for Claude Desktop)
The init -y flag uses sensible defaults:
- Detects available runtimes (uvx preferred, npx fallback)
- Configures starter bundle: filesystem, fetch, memory
- Updates Claude Desktop config automatically
Manual Setup
If you prefer step-by-step:
# 1. Install
pip install mcp-hangar
# or: uv pip install mcp-hangar
# 2. Initialize with wizard
mcp-hangar init
# 3. Start server
mcp-hangar serve
Custom Configuration
Create ~/.config/mcp-hangar/config.yaml:
providers:
github:
mode: subprocess
command: [uvx, mcp-server-github]
env:
GITHUB_TOKEN: ${GITHUB_TOKEN}
slack:
mode: subprocess
command: [uvx, mcp-server-slack]
internal-api:
mode: remote
endpoint: "http://localhost:8080"
Claude Desktop is auto-configured by mcp-hangar init. Manual setup:
Add to Claude Desktop (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"hangar": {
"command": "mcp-hangar",
"args": ["serve", "--config", "~/.config/mcp-hangar/config.yaml"]
}
}
}
Restart Claude Desktop. Done.
One Interface
hangar_call([
{"provider": "github", "tool": "search_repos", "arguments": {"query": "mcp"}},
{"provider": "slack", "tool": "post_message", "arguments": {"channel": "#dev"}},
{"provider": "internal-api", "tool": "get_status", "arguments": {}}
])
Single call. Parallel execution. All results returned together.
Benchmarks
| Scenario | Sequential | Hangar | Speedup |
|---|---|---|---|
| 15 tools, 2 providers | ~20s | 380ms | 50x |
| 50 concurrent requests | ~15s | 1.3s | 10x |
| Cold start + batch | ~5s | <500ms | 10x |
100% success rate. <10ms framework overhead.
Why It's Fast
Single-flight cold starts. When 10 parallel calls hit a cold provider, it initializes once — not 10 times.
Automatic concurrency. Configurable parallelism with backpressure. No thundering herd.
Provider pooling. Hot providers stay warm. Cold providers spin up on demand, shut down after idle TTL.
Production Ready
Lifecycle management. Lazy loading, health checks, automatic restart, graceful shutdown.
Circuit breaker. One failing provider doesn't kill your batch. Automatic isolation and recovery.
Observability. Correlation IDs across parallel calls. OpenTelemetry traces, Prometheus metrics.
Multi-provider. Subprocess, Docker, remote HTTP — mix them in a single batch call.
Configuration
providers:
- id: fast-provider
command: ["python", "fast.py"]
idle_ttl_s: 300 # Shutdown after 5min idle
health_check_interval_s: 60 # Check health every minute
max_consecutive_failures: 3 # Circuit breaker threshold
- id: docker-provider
image: my-registry/mcp-server:latest
network: bridge
- id: remote-provider
url: "https://api.example.com/mcp"
Works Everywhere
- Home lab: 2 providers, zero config complexity
- Team setup: Shared providers, Docker containers
- Enterprise: 50+ providers, observability stack, Kubernetes
Same API. Same reliability. Different scale.
Documentation
License
MIT — use it, fork it, ship it.
Servidores relacionados
Scout Monitoring MCP
patrocinadorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
patrocinadorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Pickapicon
Quickly retrieve SVGs using the Iconify API, with no external data files required.
AvaloniaUI
Tools, resources, and guidance for building cross-platform applications with AvaloniaUI.
Together AI Image Server
A TypeScript-based server for generating images using the Together AI API.
VSCode MCP
Enables AI agents and assistants to interact with Visual Studio Code through the Model Context Protocol.
RubyGems Package Info
Fetches comprehensive information about Ruby gems from RubyGems.org, including READMEs, metadata, and search functionality.
Model Context Protocol servers
A collection of reference MCP server implementations in TypeScript and Python, demonstrating MCP features and SDKs.
Scorecard
Access Scorecard's AI model evaluation and testing tools via a Cloudflare Workers deployment.
AST2LLM for Go
An AST-powered tool that enhances LLM context by automatically injecting relevant Go code structures into prompts.
iOS Simulator MCP Server
A Model Context Protocol (MCP) server for interacting with iOS simulators. This server allows you to interact with iOS simulators by getting information about them, controlling UI interactions, and inspecting UI elements.
Valar API MCP
Integrate with the Valar API, assisting frontend developers with API documentation and endpoint discovery.