stdout-mcp-server
Captures and manages stdout logs from multiple processes via a named pipe system for real-time debugging and analysis.
stdout-mcp-server
A Model Context Protocol (MCP) server that captures and manages stdout logs through a named pipe system. This server is particularly useful for:
- Capturing logs from multiple processes or applications and making them available for debugging in Cursor IDE.
- Monitoring application output in real-time and providing a MCP interface to query, filter, and analyze logs
How It Works
-
The server creates a named pipe at a specific location (
/tmp/stdout_pipeon Unix/MacOS or\\.\pipe\stdout_pipeon Windows) -
Any application can write logs to this pipe using standard output redirection. For example:
your_application | tee /tmp/stdout_pipe # or
your_application > /tmp/stdout_pipe
-
The server monitors the pipe, captures all incoming logs, and maintains a history of the last 100 entries
-
Through MCP tools, you can query, filter, and analyze these logs
System Requirements
Before installing, please ensure you have:
- Node.js v18 or newer
Installation Options
Option 1: Installation in Cursor
- Open Cursor and navigate to
Cursor > Settings > MCP Servers - Click on "Add new MCP Server"
- Update your MCP settings file with the following configuration:
name: stdout-mcp-server
type: command
command: npx stdout-mcp-server
Option 2: Installation in other MCP clients
Installation in other MCP clients
For macOS/Linux:
{
"mcpServers": {
"stdio-mcp-server": {
"command": "npx",
"args": [
"stdio-mcp-server"
]
}
}
}
For Windows:
{
"mcpServers": {
"mcp-installer": {
"command": "cmd.exe",
"args": ["/c", "npx", "stdio-mcp-server"]
}
}
}
Usage Examples
Redirecting Application Logs
To send your application's output to the pipe:
# Unix/MacOS
your_application > /tmp/stdout_pipe
# Windows (PowerShell)
your_application > \\.\pipe\stdout_pipe
Monitoring Multiple Applications
You can redirect logs from multiple sources:
# Application 1
app1 > /tmp/stdout_pipe &
# Application 2
app2 > /tmp/stdout_pipe &
Querying Logs
Your AI will use the get-logs tool in your MCP client to retrieve and filter logs:
// Get last 50 logs
get-logs()
// Get last 100 logs containing "error"
get-logs({ lines: 100, filter: "error" })
// Get logs since a specific timestamp
get-logs({ since: 1648675200000 }) // Unix timestamp in milliseconds
Features
- Named pipe creation and monitoring
- Real-time log capture and storage
- Log filtering and retrieval through MCP tools
- Configurable log history (default: 100 entries)
- Cross-platform support (Windows and Unix-based systems)
Named Pipe Locations
- Windows:
\\.\pipe\stdout_pipe - Unix/MacOS:
/tmp/stdout_pipe
Available Tools
get-logs
Retrieve logs from the named pipe with optional filtering:
Parameters:
lines(optional, default: 50): Number of log lines to returnfilter(optional): Text to filter logs bysince(optional): Timestamp to get logs after
Example responses:
// Response format
{
content: [{
type: "text",
text: "[2024-03-20T10:15:30.123Z] Application started\n[2024-03-20T10:15:31.456Z] Connected to database"
}]
}
License
MIT License
Serveurs connexes
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Storybook MCP
A universal MCP server that connects to any Storybook site and extracts documentation in real-time using Playwright. Use it with any AI or client that supports MCP (Model Context Protocol)—Cursor, Claude Desktop, Windsurf, or other MCP hosts.
Metasploit MCP Server
An MCP server for integrating with the Metasploit Framework, enabling payload generation and management.
OpenGrok
OpenGrok MCP Server is a native Model Context Protocol (MCP) VS Code extension that seamlessly bridges the gap between your organization's OpenGrok indices and GitHub Copilot Chat. It arms your AI assistant with the deep, instantaneous repository context required to traverse, understand, and search massive codebases using only natural language.
GrowthBook
Create and read feature flags, review experiments, generate flag types, search docs, and interact with GrowthBook's feature flagging and experimentation platform.
DeepSeek-Claude MCP Server
Enhance Claude's reasoning capabilities by integrating DeepSeek's advanced engine.
Enrichr MCP Server
Performs gene set enrichment analysis using the Enrichr API, supporting all available gene set libraries.
Windsor
Windsor MCP enables your LLM to query, explore, and analyze your full-stack business data integrated into Windsor.ai with zero SQL writing or custom scripting.
BlueMouse
The "Prefrontal Cortex" for LLMs. A local, data-driven logic gate that interviews AI to prevent hallucinations.
MCP ZepAi Server
A server for Zep, a long-term memory store for AI applications, requiring a ZEP_API_KEY for access.
MCP Code Executor
Allows LLMs to execute Python code within a specified and configurable Python environment.