Codex MCP Wrapper
An MCP server that wraps the OpenAI Codex CLI, exposing its functionality through the MCP API.
Agentic Developer MCP
This project wraps OpenAI's Codex CLI as an MCP (Model Context Protocol) server, making it accessible through the TeaBranch/open-responses-server middleware.
This engine may be replaced with OpenCode or Amazon Strands
Requirements
- Node 22 (
nvm install 22.15.1 | nvm use 22.15.1) required for Codex
Overview
The setup consists of three main components:
- Codex CLI: OpenAI's command-line interface for interacting with Codex.
- MCP Wrapper Server: A Node.js Express server that forwards MCP requests to Codex CLI and formats responses as MCP.
- open-responses-server: A middleware service that provides Responses API compatibility and MCP support.
Installation
Using Docker (Recommended)
# Clone this repository
git clone https://github.com/yourusername/codex-mcp-wrapper.git
cd codex-mcp-wrapper
# Start the services
./start.sh
This will start:
- Codex MCP wrapper on port 8080
- open-responses-server on port 3000
Manual Installation
# Install dependencies
npm install
# Install Codex CLI globally
npm install -g @openai/codex
# Start the MCP server
node mcp-server.js
# Install the package in development mode
pip install -e .
Usage
You can run the MCP server using either stdio or SSE transport:
# Using stdio (default)
python -m mcp_server
# Using SSE on a specific port
python -m mcp_server --transport sse --port 8000
Tool Documentation
run_codex
Clones a repository, checks out a specific branch (optional), navigates to a specific folder (optional), and runs Codex with the given request.
Parameters
repository(required): Git repository URLbranch(optional): Git branch to checkoutfolder(optional): Folder within the repository to focus onrequest(required): Codex request/prompt to run
Example
{
"repository": "https://github.com/username/repo.git",
"branch": "main",
"folder": "src",
"request": "Analyze this code and suggest improvements"
}
clone_and_write_prompt
Clones a repository, reads the system prompt from .agent/system.md, parses modelId from .agent/agent.json, writes the request to a .prompt file, and invokes the Codex CLI with the extracted model.
Parameters
repository(required): Git repository URLrequest(required): Prompt text to run through Codexfolder(optional, default/): Subfolder within the repository to operate in
Example
{
"repository": "https://github.com/username/repo.git",
"folder": "src",
"request": "Analyze this code and suggest improvements"
}
MCPS Configuration
Place a mcps.json file under the .agent/ directory to register available MCP tools. Codex will load this configuration automatically.
Example .agent/mcps.json:
{
"mcpServers": {
"agentic-developer-mcp": {
"url": "..."
}
}
}
Development
This project uses the MCP Python SDK to implement an MCP server. The primary implementation is in mcp_server/server.py.
License
MIT
Máy chủ liên quan
Alpha Vantage MCP Server
nhà tài trợAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
XcodeProj MCP Server
An MCP server for manipulating Xcode project files (.xcodeproj) using Swift. Requires Docker and macOS.
HAL (HTTP API Layer)
An MCP server that enables Large Language Models to make HTTP requests and interact with web APIs. It supports automatic tool generation from OpenAPI/Swagger specifications.
Azure DevOps MCP
Integrates with Azure DevOps, allowing interaction with its services. Requires a Personal Access Token (PAT) for authentication.
Gemini Image MCP Server
Image generation using Google's Gemini API.
TLS MCP Server
Analyze TLS certificates using OpenSSL and zlint.
Remote MCP Server on Cloudflare
A template for deploying a remote MCP server on Cloudflare Workers, allowing for custom tool integration.
WinCC Unified MCP Server
An MCP server for interfacing with SIEMENS WinCC Unified SCADA systems via their GraphQL API.
Anubis
Embeds intelligent guidance into AI workflows to organize development and ensure quality.
Sugar
Autonomous AI development system for Claude Code with task queue management and workflow automation.
Loggles
Loggles is a local-first log sink with an MCP interface that lets coding agents (Claude Code, Cursor) query application logs directly