ECharts MCP Server
A server for generating various types of charts using the ECharts library.
ECharts MCP Server

A Model Context Protocol server for generating charts using ECharts inspired by AntV.
This is a TypeScript-based MCP server that provides chart generation capabilities. It allows you to create various types of charts through MCP tools. You can also use it in Dify.
✨ Features
Nearly all types of chart supported. Each tool will return a chart preview URL which is the echarts official chart preview page.
For example, this pie chart
generate_bar_chart- Generate abarchart.generate_line_chart- Generate alinechart.generate_pie_chart- Generate apiechart.generate_scatter_chart- Generate ascatterchart.generate_radar_chart- Generate aradarchart.generate_heatmap_chart- Generate aheatmapchart.generate_general_chart- Generate aeffectScatter, tree, treemap, sunburst, boxplot, candlestick, graph, sankey, funnel, gauge or themeRiverchart.
🤖 Usage
To use with Desktop APP, such as Claude, VSCode, Cline, Cherry Studio, Cursor, and so on, add the MCP server config below.
{
"mcpServers": {
"echarts-server": {
"command": "node",
"args": [
"your-path/echarts-mcp-server/build/index.js"
]
}
}
}
Or
{
"mcpServers": {
"echarts-server": {
"command": "npx",
"args": [
"-y",
"echarts-mcp-server"
]
}
}
}
Also, you can use it on aliyun, modelscope, glama.ai, smithery.ai or others with HTTP, SSE Protocol.
🚰 Run with SSE or Streamable transport
Install the package globally.
npm install -g echarts-mcp-server
Run the server with your preferred transport option:
# For SSE transport (default endpoint: /sse)
echarts-mcp-server --transport sse
# For Streamable transport with custom endpoint
echarts-mcp-server --transport streamable
Then you can access the server at:
- SSE transport:
http://localhost:1122/sse - Streamable transport:
http://localhost:1122/mcp
🎮 CLI Options
You can also use the following CLI options when running the MCP server. Command options by run cli with -h.
MCP Server Chart CLI
Options:
--transport, -t Specify the transport protocol: "stdio", "sse", or "streamable" (default: "stdio")
--port, -p Specify the port for SSE or streamable transport (default: 1122)
--endpoint, -e Specify the endpoint for the transport:
- For SSE: default is "/sse"
- For streamable: default is "/mcp"
--help, -h Show this help message
🔨 Development
Install dependencies:
npm install
Build the server:
npm run build
Start the MCP server:
npm run start
Servidores relacionados
Scout Monitoring MCP
patrocinadorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
patrocinadorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
MCPR
Expose R functions through the Model Context Protocol (MCP) for seamless integration with AI assistants.
gget-mcp
An MCP server for the gget bioinformatics library, enabling standardized access to genomics tools and databases.
Dive AI Agent
An open-source desktop application for hosting MCP servers that integrates with function-calling LLMs.
Roo Activity Logger
Automatically logs AI coding assistant activities, such as command executions and code generation, into searchable JSON files.
Lucide Icons
Provides access to the Lucide icon library for use in LLM and agentic applications.
Obsidian Claude Code
An Obsidian plugin that integrates Claude Code into your vaults via an MCP server.
Subotiz MCP
Connect AI assistants to Subotiz - Using Subotiz's external capabilities through natural language
Super Secret MCP Server
A fun tool to generate random combinations of US States and signature soups.
MCP Mermaid Server
Generate and analyze Mermaid diagrams.
Deliberate Reasoning Engine (DRE)
Transforms linear AI reasoning into structured, auditable thought graphs, enabling language models to externalize their reasoning process as a directed acyclic graph (DAG).