Nuanced MCP Server
Provides call graph analysis for LLMs using the nuanced library.
Nuanced MCP Server
A Model Context Protocol (MCP) server that provides call graph analysis capabilities to LLMs through the nuanced library.
Overview
This MCP server enables LLMs to understand code structure by accessing function call graphs through standardized tools and resources. It allows AI assistants to:
- Initialize call graphs for Python repos
- Explore function call relationships
- Analyze dependencies between functions
- Provide more contextually aware code assistance
API
Tools
-
initialize_graph
- Initialize a code graph for the given repository path
- Input:
repo_path(string)
-
switch_repository
- Switch to a different initialized repository
- Input:
repo_path(string)
-
list_repositories
- List all initialized repositories
- No inputs required
-
get_function_call_graph
- Get the call graph for a specific function
- Inputs:
file_path(string)function_name(string)repo_path(string, optional) - uses active repository if not specified
-
analyze_dependencies
- Find all module or file dependencies in the codebase
- Inputs (at least one required):
file_path(string, optional)module_name(string, optional)
-
analyze_change_impact
- Analyze the impact of changing a specific function
- Inputs:
file_path(string)function_name(string)
Resources
-
graph://summary
- Get a summary of the currently loaded code graph
- No parameters required
-
graph://repo/{repo_path}/summary
- Get a summary of a specific repository's code graph
- Parameters:
repo_path(string) - Path to the repository
-
graph://function/{file_path}/{function_name}
- Get detailed information about a specific function
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
Prompts
-
analyze_function
- Create a prompt to analyze a function with its call graph
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
-
impact_analysis
- Create a prompt to analyze the impact of changing a function
- Parameters:
file_path(string) - Path to the file containing the functionfunction_name(string) - Name of the function to analyze
-
analyze_dependencies_prompt
- Create a prompt to analyze dependencies of a file or module
- Parameters (at least one required):
file_path(string, optional) - Path to the file to analyzemodule_name(string, optional) - Name of the module to analyze
Usage with Claude Desktop
Add this to your claude_desktop_config.json
UV
{
"mcpServers": {
"nuanced": {
"command": "uv",
"args": [
"--directory",
"/path/to/nuanced-mcp",
"run",
"nuanced_mcp_server.py"
]
}
}
}
संबंधित सर्वर
Scout Monitoring MCP
प्रायोजकPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
प्रायोजकAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Agentic Control Framework (ACF)
A toolkit for autonomous agent development with tools for task management, filesystem operations, browser automation, and terminal control.
Harness
Access and interact with Harness platform data, including pipelines, repositories, logs, and artifact registries.
go-mcp実験場
A Go-based MCP server example demonstrating correct usage of go.mod and build/run commands.
Enrichr MCP Server
Performs gene set enrichment analysis using the Enrichr API, supporting all available gene set libraries.
ExMCP Test Server
An Elixir-based MCP server for testing and experimenting with the Model Context Protocol.
HiGHS MCP Server
Provides linear programming (LP) and mixed-integer programming (MIP) optimization capabilities using the HiGHS solver.
ComfyUI
An MCP server for ComfyUI integration.
Dify MCP Server
A TypeScript-based server that integrates the Dify AI application platform with the MCP Client.
Python Notebook MCP
Enables AI assistants to interact with local Jupyter notebooks (.ipynb).
Bifrost VSCode Dev Tools
Exposes VSCode dev tools features to MCP clients, with support for project-specific configurations.