WinTerm MCP
Provides programmatic access to the Windows terminal, enabling AI models to interact with the command line interface.
WinTerm MCP
A Model Context Protocol server that provides programmatic access to the Windows terminal. This server enables AI models to interact with the Windows command line interface through a set of standardized tools.
Features
- Write to Terminal: Execute commands or write text to the Windows terminal
- Read Terminal Output: Retrieve output from previously executed commands
- Send Control Characters: Send control signals (e.g., Ctrl+C) to the terminal
- Windows-Native: Built specifically for Windows command line interaction
Installation
-
Clone the Repository:
git clone https://github.com/capecoma/winterm-mcp.git cd winterm-mcp -
Install Dependencies:
npm install -
Build the Project:
npm run build -
Configure Claude Desktop:
Add the server config to %APPDATA%/Claude/claude_desktop_config.json:
{
"mcpServers": {
"github.com/capecoma/winterm-mcp": {
"command": "node",
"args": ["path/to/build/index.js"],
"disabled": false,
"autoApprove": []
}
}
}
Note: Replace "path/to/build/index.js" with the actual path to your built index.js file.
Available Tools
write_to_terminal
Writes text or commands to the terminal.
{
"command": "echo Hello, World!"
}
read_terminal_output
Reads the specified number of lines from terminal output.
{
"linesOfOutput": 5
}
send_control_character
Sends a control character to the terminal (e.g., Ctrl+C).
{
"letter": "C"
}
Development
For development with auto-rebuild:
npm run dev
License
MIT License - see LICENSE file.
Related Servers
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Apifox
A TypeScript MCP server to access Apifox API data via Stdio.
Stata-MCP
Perform regression analysis using Stata with the help of an LLM. Requires a local Stata installation and an external LLM API key.
Stability AI
Integrates with the Stability AI API for image generation, editing, and upscaling.
Chainlink Feeds
Provides real-time access to Chainlink's decentralized on-chain price feeds.
Ollama MCP Bridge
A bridge API service connecting Ollama with Model Context Protocol (MCP) servers.
Windows CLI
Interact with Windows command-line interfaces like PowerShell, CMD, Git Bash, and WSL.
codegraph-rust
A blazingly fast codebase graphRAG implementation in 100% Rust
BlenderMCP
Connects Blender to Claude AI via the Model Context Protocol (MCP), enabling direct interaction and control for prompt-assisted 3D modeling, scene creation, and manipulation.
Ghidra MCP Server
Exposes binary analysis data from Ghidra, including functions and pseudocode, to LLMs.
MCP-guide
A guide for setting up an MCP server using a Python virtual environment and integrating it with the Cline VS Code extension.