Integrates ComfyUI with MCP, allowing the use of custom workflows. Requires a running ComfyUI server.
python src/test_comfyui.py
mcp dev src/server.py
Edit src/.env
to set ComfyUI host and port:
COMFYUI_HOST=localhost
COMFYUI_PORT=8188
workflows
directory and declare them as new tools in the system.text_to_image
download_image
tool, ordownload_image
text_to_image
) using the image URL.run_workflow_with_file
Run a workflow by providing the path to a workflow JSON file.
# You should ask to agent like this.
Run comfyui workflow with text_to_image.json
example image of CursorAI
run_workflow_with_json
Run a workflow by providing the workflow JSON data directly.
# You should ask to agent like this.
Run comfyui workflow with this
{
"3": {
"inputs": {
"seed": 156680208700286,
"steps": 20,
... (workflow JSON example)
}
Example mcp.json
:
{
"mcpServers": {
"comfyui": {
"command": "uv",
"args": [
"--directory",
"PATH/MCP/comfyui",
"run",
"--with",
"mcp",
"--with",
"websocket-client",
"--with",
"python-dotenv",
"mcp",
"run",
"src/server.py:mcp"
]
}
}
}
download_image
may be difficult since the Docker container does not share the host filesystem.RETURN_URL=false
in .env
to receive image data as bytes.COMFYUI_HOST
in .env
to the appropriate address (e.g., host.docker.internal
or your server's IP).# First build image
docker image build -t mcp/comfyui .
{
"mcpServers": {
"comfyui": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-p",
"3001:3000",
"mcp/comfyui"
]
}
}
}
Also you can use prebuilt image.
{
"mcpServers": {
"comfyui": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-p",
"3001:3000",
"overseer66/mcp-comfyui"
]
}
}
}
Run the SSE server with Docker:
docker run -i --rm -p 8001:8000 overseer66/mcp-comfyui-sse
Configure mcp.json
(change localhost to your IP or domain if needed):
{
"mcpServers": {
"comfyui": {
"url": "http://localhost:8001/sse"
}
}
}
NOTE: When adding new workflows as tools, you need to rebuild and redeploy the Docker images to make them available.
Receive system notifications in your code editor when an AI response is complete.
Take screenshots and analyze mobile apps with AI assistance directly from your IDE.
A template for building Model Context Protocol (MCP) servers using the mcp-framework for Node.js.
Explore and analyze OpenAPI specifications from local files or remote URLs.
Retrieve image dimensions and compress images from URLs or local files using Tinify and Figma APIs.
A Tabby plugin implementing an MCP server for AI-powered terminal control and automation.
Interact with the JFrog Platform API for repository management, build tracking, and release lifecycle management.
Generate images using Bytedance's SeedDream 3.0 model via the FAL AI platform. Requires a FAL AI API key.
A remote, authentication-free MCP server for weather data, deployable on Cloudflare Workers or run locally via npm.
A unified interface for various Large Language Model (LLM) providers, including OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama.