iTerm
Access and control local iTerm2 terminal sessions.
iterm-mcp
A Model Context Protocol server that provides access to your iTerm session.

Features
Efficient Token Use: iterm-mcp gives the model the ability to inspect only the output that the model is interested in. The model typically only wants to see the last few lines of output even for long running commands.
Natural Integration: You share iTerm with the model. You can ask questions about what's on the screen, or delegate a task to the model and watch as it performs each step.
Full Terminal Control and REPL support: The model can start and interact with REPL's as well as send control characters like ctrl-c, ctrl-z, etc.
Easy on the Dependencies: iterm-mcp is built with minimal dependencies and is runnable via npx. It's designed to be easy to add to Claude Desktop and other MCP clients. It should just work.
Safety Considerations
- The user is responsible for using the tool safely.
- No built-in restrictions: iterm-mcp makes no attempt to evaluate the safety of commands that are executed.
- Models can behave in unexpected ways. The user is expected to monitor activity and abort when appropriate.
- For multi-step tasks, you may need to interrupt the model if it goes off track. Start with smaller, focused tasks until you're familiar with how the model behaves.
Tools
write_to_terminal- Writes to the active iTerm terminal, often used to run a command. Returns the number of lines of output produced by the command.read_terminal_output- Reads the requested number of lines from the active iTerm terminal.send_control_character- Sends a control character to the active iTerm terminal.
Requirements
- iTerm2 must be running
- Node version 18 or greater
Installation
To use with Claude Desktop, add the server config:
On macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"iterm-mcp": {
"command": "npx",
"args": [
"-y",
"iterm-mcp"
]
}
}
}
Installing via Smithery
To install iTerm for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install iterm-mcp --client claude
Development
Install dependencies:
yarn install
Build the server:
yarn run build
For development with auto-rebuild:
yarn run watch
Debugging
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
yarn run inspector
yarn debug <command>
The Inspector will provide a URL to access debugging tools in your browser.
相關伺服器
Scout Monitoring MCP
贊助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
贊助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
llm-mcp
A Ruby gem for integrating Large Language Models (LLMs) via the Model Context Protocol (MCP) into development workflows.
Tripwire
Context injection for AI agents via MCP. Define path-based policies in YAML — when an agent reads a matching file, relevant knowledge is auto-injected. Prevents mistakes before they happen. Works with Claude Code, Cursor, and any MCP client.
Enhanced QR Code MCP Server
Generates QR codes with metadata, batch processing, and file management tools.
MCP Playground
A demonstration MCP server implementation in Go featuring real-time bidirectional file communication.
Windows CLI
MCP server for secure command-line interactions on Windows systems, enabling controlled access to PowerShell, CMD, and Git Bash shells.
USolver
A server for solving combinatorial, convex, integer, and non-linear optimization problems.
Crates MCP Server
Query Rust crates from crates.io and docs.rs. Search for crates, get info, versions, dependencies, and documentation.
Figma Copilot
Enables AI assistants to interact with and automate Figma designs programmatically.
Tauri MCP Server
A server for testing and interacting with Tauri v2 applications, providing tools for process management, window manipulation, and debugging.
MCP-Mem0
Integrate long-term memory into AI agents using Mem0.