TransformerBee.MCP
An MCP server for the transformer.bee service, configurable via environment variables.
TransformerBee.MCP
This is a simple PoC of a Model Context Protocol (MCP) server for transformer.bee, written in Python.
Under the hood it uses python-mcp and transformerbeeclient.py.
Installation
You can install the MCP server as Python package or pull the Docker image.
Install as Python Package
uv install transformerbeemcp
or if you are using pip:
pip install transformerbeemcp
Install as Docker Image
docker pull ghcr.io/hochfrequenz/transformerbee.mcp:latest
Start the Server via CLI
Python
_The package ships a simple CLI argument to start the server.
In a terminal inside the virtual environment in which you installed the package (here myvenv), call:
(myvenv) run-transformerbee-mcp-server
Docker
docker run --network host -i --rm -e TRANSFORMERBEE_HOST=http://localhost:5021 ghcr.io/hochfrequenz/transformerbee.mcp:latest
(For the environment variables -e ..., see below or the transformerbeeclient.py docs.)
Register MCP Server in Claude Desktop
If you checked out this repository
cd path/to/reporoot/src/transformerbeemcp
mcp install server.py
If you installed the package via pip/uv
Modify your claude_desktop_config.json (that can be found in Claude Desktop menu via "Datei > Einstellungen > Entwickler > Konfiguration bearbeiten"):
{
"mcpServers": {
"TransformerBee.mcp": {
"command": "C:\\github\\MyProject\\.myvenv\\Scripts\\run-transformerbee-mcp-server.exe",
"args": [],
"env": {
"TRANSFORMERBEE_HOST": "http://localhost:5021",
"TRANSFORMERBEE_CLIENT_ID": "",
"TRANSFORMERBEE_CLIENT_SECRET": ""
}
}
}
}
where C:\github\MyProject\.myvenv is the path to your virtual environment where you installed the package and localhost:5021 exposes transformer.bee running in a docker container.
Alternatively, if you haven't configured this handy CLI command
https://github.com/Hochfrequenz/TransformerBee.mcp/blob/c0898769670469df13f23b57a55fe4b71ed9795b/pyproject.toml#L101-L102
you can just call python with non-empty args.
Note that this package marks uv as a dev-dependency, so you might need to install it pip install transformerbeempc[dev] in your virtual environment as well as a lot of MCP tooling assumes you have uv installed.
For details about the environment variables and/or starting transformer.bee locally, check transformerbeeclient.py docs.
If you installed the package via Docker
{
"mcpServers": {
"TransformerBee.mcp": {
"command": "docker",
"args": [
"run",
"--network",
"host",
"-i",
"--rm",
"-e",
"TRANSFORMERBEE_HOST=http://localhost:5021",
"ghcr.io/hochfrequenz/transformerbee.mcp:latest"
],
"env": {
"TRANSFORMERBEE_HOST": "http://localhost:5021",
"TRANSFORMERBEE_CLIENT_ID": "",
"TRANSFORMERBEE_CLIENT_SECRET": ""
}
}
}
}
I'm aware, that using the --network host option is a bit hacky and not best practice.
관련 서버
Scout Monitoring MCP
스폰서Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
스폰서Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Kai
Kai provides a bridge between large language models (LLMs) and your Kubernetes clusters, enabling natural language interaction with Kubernetes resources. The server exposes a comprehensive set of tools for managing clusters, namespaces, pods, deployments, services, and other Kubernetes resources
Remote MCP Server Authless
An example of a remote MCP server deployed on Cloudflare Workers without authentication.
RenderLens
Visual verification MCP server — render code to screenshots, run WCAG accessibility audits, and pixel-diff UI changes. Free, no API key.
FlowZap
FlowZap's MCP generates Workflow, Sequence and Architecture Diagrams from your App in seconds. Pretty ones.
Creatify
MCP Server that exposes Creatify AI API capabilities for AI video generation, including avatar videos, URL-to-video conversion, text-to-speech, and AI-powered editing tools.
Remote MCP Server (Authless)
An example of a remote MCP server deployable on Cloudflare Workers, featuring customizable tools and no authentication.
AI pair programming
Orchestrates a dual-AI engineering loop where a Primary AI plans and implements, while a Review AI validates and reviews, with continuous feedback for optimal code quality. Supports custom AI pairing (Claude, Codex, Gemini, etc.)
SACL MCP Server
A framework for bias-aware code retrieval using semantic-augmented reranking and localization.
Just Prompt
A unified interface for various Large Language Model (LLM) providers, including OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama.
CodebaseIQ Pro
Provides AI assistants with a comprehensive, one-time analysis for complete codebase context and understanding.