mcp-graphql
A GraphQL server that supports the Model Context Protocol (MCP), enabling Large Language Models (LLMs) to interact with GraphQL APIs through schema introspection and query execution.
mcp-graphql
A Model Context Protocol server that enables LLMs to interact with GraphQL APIs. This implementation provides schema introspection and query execution capabilities, allowing models to discover and use GraphQL APIs dynamically.
Usage
Run mcp-graphql with the correct endpoint, it will automatically try to introspect your queries.
Command Line Arguments
| Argument | Description | Default |
|---|---|---|
--endpoint | GraphQL endpoint URL | http://localhost:4000/graphql |
--headers | JSON string containing headers for requests | {} |
--enable-mutations | Enable mutation operations (disabled by default) | false |
--name | Name of the MCP server | mcp-graphql |
--schema | Path to a local GraphQL schema file (optional) | - |
Examples
# Basic usage with a local GraphQL server
npx mcp-graphql --endpoint http://localhost:3000/graphql
# Using with custom headers
npx mcp-graphql --endpoint https://api.example.com/graphql --headers '{"Authorization":"Bearer token123"}'
# Enable mutation operations
npx mcp-graphql --endpoint http://localhost:3000/graphql --enable-mutations
# Using a local schema file instead of introspection
npx mcp-graphql --endpoint http://localhost:3000/graphql --schema ./schema.graphql
Available Tools
The server provides two main tools:
-
introspect-schema: This tool retrieves the GraphQL schema. Use this first if you don't have access to the schema as a resource. This uses either the local schema file or an introspection query.
-
query-graphql: Execute GraphQL queries against the endpoint. By default, mutations are disabled unless
--enable-mutationsis specified.
Resources
- graphql-schema: The server exposes the GraphQL schema as a resource that clients can access. This is either the local schema file or based on an introspection query.
Installation
Installing via Smithery
To install GraphQL MCP Toolkit for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install mcp-graphql --client claude
Installing Manually
It can be manually installed to Claude:
{
"mcpServers": {
"mcp-graphql": {
"command": "npx",
"args": ["mcp-graphql", "--endpoint", "http://localhost:3000/graphql"]
}
}
}
Security Considerations
Mutations are disabled by default as a security measure to prevent an LLM from modifying your database or service data. Consider carefully before enabling mutations in production environments.
Customize for your own server
This is a very generic implementation where it allows for complete introspection and for your users to do whatever (including mutations). If you need a more specific implementation I'd suggest to just create your own MCP and lock down tool calling for clients to only input specific query fields and/or variables. You can use this as a reference.
相關伺服器
Scout Monitoring MCP
贊助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
贊助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Persona MCP Server
Dynamically manage AI personas from markdown files for AI assistants like Claude.
Deno Gemini Grounding MCP Server
A Model Context Protocol server that provides access to the connpass users API v2, utilizing Gemini for grounding.
Windows API
An MCP server for interacting with the native Windows API, enabling control over system functions and resources.
OSSInsight
Analyze GitHub repositories, developers, and organizations with data from OSSInsight.io.
QuantConnect
A server for local interactions with the QuantConnect API.
Skills-ContextManager
Don’t pollute your AI agent’s context with 1,000 skills. Use Skills-ContextManager, a self-hosted web UI for managing AI skills and workflows by providing skills to an AI agent via MCP only when needed. Simply add skills to your library and enable or disable them with a toggle. Choose whether a skill is always loaded into context or dynamically activated when the AI agent determines it’s needed.
Rongda MCP Server
An MCP server for Rongda that supports authentication through environment variables.
MCP Code Executor
Allows LLMs to execute Python code within a specified and configurable Python environment.
MCP Base Server
A base template for creating new MCP servers, designed for easy containerized deployment with Docker.
MCP Arduino Server
An MCP server for the Arduino CLI, offering tools to manage sketches, boards, libraries, and files.