MCP Hangar
Kubernetes-native registry for managing multiple MCP servers with lazy loading, health monitoring, and RBAC
MCP Hangar
Parallel MCP tool execution. One interface. 50x faster.
The Problem
Your AI agent calls 5 tools sequentially. Each takes 200ms. That's 1 second of waiting.
Hangar runs them in parallel. 200ms total. Same results, 50x faster.
Quick Start
30 seconds to working MCP providers:
# Install, configure, and start - zero interaction
curl -sSL https://mcp-hangar.io/install.sh | bash && mcp-hangar init -y && mcp-hangar serve
That's it. Filesystem, fetch, and memory providers are now available to Claude.
What just happened?
- Install - Downloaded and installed
mcp-hangarvia pip/uv - Init - Created
~/.config/mcp-hangar/config.yamlwith starter providers - Serve - Started the MCP server (stdio mode for Claude Desktop)
The init -y flag uses sensible defaults:
- Detects available runtimes (uvx preferred, npx fallback)
- Configures starter bundle: filesystem, fetch, memory
- Updates Claude Desktop config automatically
Manual Setup
If you prefer step-by-step:
# 1. Install
pip install mcp-hangar
# or: uv pip install mcp-hangar
# 2. Initialize with wizard
mcp-hangar init
# 3. Start server
mcp-hangar serve
Custom Configuration
Create ~/.config/mcp-hangar/config.yaml:
providers:
github:
mode: subprocess
command: [uvx, mcp-server-github]
env:
GITHUB_TOKEN: ${GITHUB_TOKEN}
slack:
mode: subprocess
command: [uvx, mcp-server-slack]
internal-api:
mode: remote
endpoint: "http://localhost:8080"
Claude Desktop is auto-configured by mcp-hangar init. Manual setup:
Add to Claude Desktop (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"hangar": {
"command": "mcp-hangar",
"args": ["serve", "--config", "~/.config/mcp-hangar/config.yaml"]
}
}
}
Restart Claude Desktop. Done.
One Interface
hangar_call([
{"provider": "github", "tool": "search_repos", "arguments": {"query": "mcp"}},
{"provider": "slack", "tool": "post_message", "arguments": {"channel": "#dev"}},
{"provider": "internal-api", "tool": "get_status", "arguments": {}}
])
Single call. Parallel execution. All results returned together.
Benchmarks
| Scenario | Sequential | Hangar | Speedup |
|---|---|---|---|
| 15 tools, 2 providers | ~20s | 380ms | 50x |
| 50 concurrent requests | ~15s | 1.3s | 10x |
| Cold start + batch | ~5s | <500ms | 10x |
100% success rate. <10ms framework overhead.
Why It's Fast
Single-flight cold starts. When 10 parallel calls hit a cold provider, it initializes once — not 10 times.
Automatic concurrency. Configurable parallelism with backpressure. No thundering herd.
Provider pooling. Hot providers stay warm. Cold providers spin up on demand, shut down after idle TTL.
Production Ready
Lifecycle management. Lazy loading, health checks, automatic restart, graceful shutdown.
Circuit breaker. One failing provider doesn't kill your batch. Automatic isolation and recovery.
Observability. Correlation IDs across parallel calls. OpenTelemetry traces, Prometheus metrics.
Multi-provider. Subprocess, Docker, remote HTTP — mix them in a single batch call.
Configuration
providers:
- id: fast-provider
command: ["python", "fast.py"]
idle_ttl_s: 300 # Shutdown after 5min idle
health_check_interval_s: 60 # Check health every minute
max_consecutive_failures: 3 # Circuit breaker threshold
- id: docker-provider
image: my-registry/mcp-server:latest
network: bridge
- id: remote-provider
url: "https://api.example.com/mcp"
Works Everywhere
- Home lab: 2 providers, zero config complexity
- Team setup: Shared providers, Docker containers
- Enterprise: 50+ providers, observability stack, Kubernetes
Same API. Same reliability. Different scale.
Documentation
License
MIT — use it, fork it, ship it.
관련 서버
Scout Monitoring MCP
스폰서Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
스폰서Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Petstore MCP Server & Client
An MCP server and client implementation for the Swagger Petstore API.
Metal MCP Server
Search Metal Framework documentation and generate code.
Tauri Development MCP Server
Build, test, and debug mobile and desktop apps with the Tauri framework faster with automated UI interaction, screenshots, DOM state, and console logs from your app under development.
Detrix
Agentic debugger
mcp-graphql
A GraphQL server that supports the Model Context Protocol (MCP), enabling Large Language Models (LLMs) to interact with GraphQL APIs through schema introspection and query execution.
Cisco SSH MCP Server
Connect to, configure, and monitor Cisco network devices like routers and switches via SSH.
mcprouter
A proxy for routing requests to remote MCP servers.
FastAPI with MCP
A FastAPI application demonstrating MCP integration for mathematical operations and tool registration.
Video Agent MCP Server
An MCP server for AI-powered video creation using the FAL AI API.
iOS Device Control
An MCP server to control iOS simulators and real devices, enabling AI assistant integration on macOS.