ECharts MCP Server
A server for generating various types of charts using the ECharts library.
ECharts MCP Server

A Model Context Protocol server for generating charts using ECharts inspired by AntV.
This is a TypeScript-based MCP server that provides chart generation capabilities. It allows you to create various types of charts through MCP tools. You can also use it in Dify.
✨ Features
Nearly all types of chart supported. Each tool will return a chart preview URL which is the echarts official chart preview page.
For example, this pie chart
generate_bar_chart- Generate abarchart.generate_line_chart- Generate alinechart.generate_pie_chart- Generate apiechart.generate_scatter_chart- Generate ascatterchart.generate_radar_chart- Generate aradarchart.generate_heatmap_chart- Generate aheatmapchart.generate_general_chart- Generate aeffectScatter, tree, treemap, sunburst, boxplot, candlestick, graph, sankey, funnel, gauge or themeRiverchart.
🤖 Usage
To use with Desktop APP, such as Claude, VSCode, Cline, Cherry Studio, Cursor, and so on, add the MCP server config below.
{
"mcpServers": {
"echarts-server": {
"command": "node",
"args": [
"your-path/echarts-mcp-server/build/index.js"
]
}
}
}
Or
{
"mcpServers": {
"echarts-server": {
"command": "npx",
"args": [
"-y",
"echarts-mcp-server"
]
}
}
}
Also, you can use it on aliyun, modelscope, glama.ai, smithery.ai or others with HTTP, SSE Protocol.
🚰 Run with SSE or Streamable transport
Install the package globally.
npm install -g echarts-mcp-server
Run the server with your preferred transport option:
# For SSE transport (default endpoint: /sse)
echarts-mcp-server --transport sse
# For Streamable transport with custom endpoint
echarts-mcp-server --transport streamable
Then you can access the server at:
- SSE transport:
http://localhost:1122/sse - Streamable transport:
http://localhost:1122/mcp
🎮 CLI Options
You can also use the following CLI options when running the MCP server. Command options by run cli with -h.
MCP Server Chart CLI
Options:
--transport, -t Specify the transport protocol: "stdio", "sse", or "streamable" (default: "stdio")
--port, -p Specify the port for SSE or streamable transport (default: 1122)
--endpoint, -e Specify the endpoint for the transport:
- For SSE: default is "/sse"
- For streamable: default is "/mcp"
--help, -h Show this help message
🔨 Development
Install dependencies:
npm install
Build the server:
npm run build
Start the MCP server:
npm run start
관련 서버
Scout Monitoring MCP
스폰서Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
스폰서Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Zero-Vector v3
A server for Zero-Vector's hybrid vector-graph persona and memory management system, featuring advanced LangGraph workflow capabilities.
Metal MCP Server
Search Metal Framework documentation and generate code.
NestJS MCP Server Module
A NestJS module for building MCP servers to expose tools and resources for AI, with support for multiple transport types.
MLflow MCP Server
Integrates with MLflow, enabling AI assistants to interact with experiments, runs, and registered models.
Agentic Control Framework (ACF)
A toolkit for autonomous agent development with tools for task management, filesystem operations, browser automation, and terminal control.
Hoofy
Your AI development companion. An MCP server that gives your AI persistent memory, structured specifications, and adaptive change management — so it builds what you actually want.
OPNsense MCP Server
A comprehensive MCP server for managing OPNsense firewalls, offering over 300 tools for configuration and monitoring.
Interactive Feedback MCP
An MCP server for AI-assisted development tools like Cursor and Claude, supporting interactive feedback workflows with AI.
OpenAPI Invoker
Invokes any OpenAPI specification through a Model Context Protocol (MCP) server.
MCP Server
A backend service providing tools, resources, and prompts for AI models using the Model Context Protocol (MCP).