Codebase Context Dumper
Easily provide codebase context to Large Language Models (LLMs).
codebase-context-dumper MCP Server
A Model Context Protocol (MCP) server designed to easily dump your codebase context into Large Language Models (LLMs).
Why Use This?
Large context windows in LLMs are powerful, but manually selecting and formatting files from a large codebase is tedious. This tool automates the process by:
- Recursively scanning your project directory.
- Including text files from the specified directory tree that are not excluded by
.gitignorerules. - Automatically skipping binary files.
- Concatenating the content with clear file path markers.
- Supporting chunking to handle codebases larger than the LLM's context window.
- Integrating seamlessly with MCP-compatible clients.
Usage (Recommended: npx)
The easiest way to use this tool is via npx, which runs the latest version without needing a local installation.
Configure your MCP client (e.g., Claude Desktop, VS Code extensions) to use the following command:
{
"mcpServers": {
"codebase-context-dumper": {
"command": "npx",
"args": [
"-y",
"@lex-tools/codebase-context-dumper"
]
}
}
}
The MCP client will then be able to invoke the dump_codebase_context tool provided by this server.
Features & Tool Details
Tool: dump_codebase_context
Recursively reads text files from a specified directory, respecting .gitignore rules and skipping binary files. Concatenates content with file path headers/footers. Supports chunking the output for large codebases.
Functionality:
- Scans the directory provided in
base_path. - Respects
.gitignorefiles at all levels (including nested ones and.gitby default). - Detects and skips binary files.
- Reads the content of each valid text file.
- Prepends a header (
--- START: relative/path/to/file ---) and appends a footer (--- END: relative/path/to/file ---) to each file's content. - Concatenates all processed file contents into a single string.
Input Parameters:
base_path(string, required): The absolute path to the project directory to scan.num_chunks(integer, optional, default: 1): The total number of chunks to divide the output into. Must be >= 1.chunk_index(integer, optional, default: 1): The 1-based index of the chunk to return. Requiresnum_chunks > 1andchunk_index <= num_chunks.
Output: Returns the concatenated (and potentially chunked) text content.
Local Installation & Usage (Advanced)
If you prefer to run a local version (e.g., for development):
- Clone the repository:
git clone [email protected]:lex-tools/codebase-context-dumper.git cd codebase-context-dumper - Install dependencies:
npm install - Build the server:
npm run build - Configure your MCP client to point to the local build output:
{ "mcpServers": { "codebase-context-dumper": { "command": "/path/to/your/local/codebase-context-dumper/build/index.js" // Adjust path } } }
Contributing
Contributions are welcome! Please see CONTRIBUTING.md for details on development, debugging, and releasing new versions.
License
This project is licensed under the Apache License 2.0. See the LICENSE file for details.
Servidores relacionados
Alpha Vantage MCP Server
patrocinadorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
ask-gemini-mcp
MCP server that enables AI assistants to interact with Google Gemini CLI
Dev.to MCP Server
An MCP server for the Dev.to API to search, browse, read, and create content on the platform.
AI Counsel
True deliberative consensus MCP server where AI models debate and refine positions across multiple rounds
MCP Sandbox
Execute Python code and install packages safely within isolated Docker containers.
ThoughtSpot SpotterCode MCP Server
AI-powered MCP server from ThoughtSpot that helps developers integrate ThoughtSpot content, Visual Embed SDK, and REST APIs in AI-native IDEs.
XLUXX Trust Layer
Runtime trust scoring for MCP servers. Monitors 15,000+ servers with reliability metrics, drift detection, and fallback recommendations.
SeedDream 3.0
Generate images using Bytedance's SeedDream 3.0 model via the FAL AI platform.
MCP Client for Ollama
A Python client that connects local LLMs via Ollama to Model Context Protocol servers, enabling them to use tools.
MCP Neurolora
An intelligent server for code analysis, collection, and documentation generation using the OpenAI API.
MCP SGF Server
Process SGF (Smart Game Format) files to extract game information and generate visual board diagrams.
