Nuanced MCP Server
Provides call graph analysis for LLMs using the nuanced library.
Nuanced MCP Server
A Model Context Protocol (MCP) server that provides call graph analysis capabilities to LLMs through the nuanced library.
Overview
This MCP server enables LLMs to understand code structure by accessing function call graphs through standardized tools and resources. It allows AI assistants to:
- Initialize call graphs for Python repos
- Explore function call relationships
- Analyze dependencies between functions
- Provide more contextually aware code assistance
API
Tools
-
initialize_graph
- Initialize a code graph for the given repository path
- Input:
repo_path(string)
-
switch_repository
- Switch to a different initialized repository
- Input:
repo_path(string)
-
list_repositories
- List all initialized repositories
- No inputs required
-
get_function_call_graph
- Get the call graph for a specific function
- Inputs:
file_path(string)function_name(string)repo_path(string, optional) - uses active repository if not specified
-
analyze_dependencies
- Find all module or file dependencies in the codebase
- Inputs (at least one required):
file_path(string, optional)module_name(string, optional)
-
analyze_change_impact
- Analyze the impact of changing a specific function
- Inputs:
file_path(string)function_name(string)
Resources
-
graph://summary
- Get a summary of the currently loaded code graph
- No parameters required
-
graph://repo/{repo_path}/summary
- Get a summary of a specific repository's code graph
- Parameters:
repo_path(string) - Path to the repository
-
graph://function/{file_path}/{function_name}
- Get detailed information about a specific function
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
Prompts
-
analyze_function
- Create a prompt to analyze a function with its call graph
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
-
impact_analysis
- Create a prompt to analyze the impact of changing a function
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
-
analyze_dependencies_prompt
- Create a prompt to analyze dependencies of a file or module
- Parameters (at least one required):
file_path(string, optional) - Path to the file to analyzemodule_name(string, optional) - Name of the module to analyze
Usage with Claude Desktop
Add this to your claude_desktop_config.json
UV
{
"mcpServers": {
"nuanced": {
"command": "uv",
"args": [
"--directory",
"/path/to/nuanced-mcp",
"run",
"nuanced_mcp_server.py"
]
}
}
}
相关服务器
Scout Monitoring MCP
赞助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
赞助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
DeepSeek MCP Server
A server for code generation and completion using the DeepSeek API.
Osquery MCP Server
An MCP server for Osquery that allows AI assistants to answer system diagnostic questions using natural language.
Jinni
A tool to provide Large Language Models with project context by intelligently filtering and concatenating relevant files.
Lingo.dev
Make your AI agent speak every language on the planet, using Lingo.dev Localization Engine.
Auto API - YApi
A tool to retrieve API interface information from YApi, with authentication configurable via environment variables.
Figma MCP Server
Connects to the Figma API, allowing AI tools to access and interact with your Figma designs.
my-mcp-server
A template for building Model Context Protocol (MCP) servers using the mcp-framework for Node.js.
Authless Remote MCP Server
A remote MCP server without authentication, deployable on Cloudflare Workers or locally with npm.
oclif MCP Server Plugin
An oclif CLI plugin that automatically discovers and serves commands via the Model Context Protocol (MCP).
MCP Github OAuth
An MCP server with built-in GitHub OAuth support, deployable on Cloudflare Workers.