Sequa MCP
A proxy that connects local STDIO with remote MCP servers, enabling IDEs to use MCP without extra infrastructure.
Sequa MCP
Self Maintaining Docs for Devs & AI — Turn your codebase into always-current documentation, so engineers and AI never read outdated docs again.
This repository is the entry point for using Sequa via the Model Context Protocol (MCP). Add Sequa as an MCP server to Cursor, Claude, Windsurf, VSCode, Cline, Highlight, Augment, or any other MCP‑capable client.
🤔 What is Sequa?
70% of docs are never updated as underlying facts change, and 30% of dev time is lost searching, handling interruptions, and reworking. Documentation drifts out of sync, knowledge gets scattered across tools, and experts repeat themselves. The result: onboarding slows, context is lost, and teams waste time chasing answers instead of building.
Sequa fixes this. It's a Contextual Knowledge Engine that:
- Generates docs that stay accurate — architecture overviews, API docs, runbooks, onboarding guides
- Continuously updates from your code — eliminating stale pages and manual upkeep
- Unifies knowledge across repositories — no more hunting across Confluence, Slack, and GitHub
- Powers your team and AI agents — ask in plain language, get current answers with source links
Why use Sequa MCP?
Stop stitching context for Copilot and Cursor. With Sequa MCP, your AI tools know your entire codebase and docs out of the box.
Sequa gives your AI assistants deep, real-time knowledge of your codebase, docs, tickets, and internal standards. Instead of writing code you have to fix, they write code that fits — first try.
🧱 Prerequisites
Before you configure any AI agent:
- Create / sign in to your Sequa account at https://app.sequa.ai/login
- Setup a Project inside the Sequa app
- Inside that project, locate the MCP Setup URLs and select the transport your AI agent supports
- Copy the URL or configuration and install it in your client
❗ If you skip project creation the MCP server will refuse connections — the proxy can launch but you will receive auth / project errors.
🚀 Quick Start (Proxy Launch)
NPX (most common)
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Replace the URL if you use an endpoint from the specific project
🔌 IDE / Tool Configuration
Cursor (~/.cursor/mcp.json)
{
"mcpServers": {
"sequa": {
"url": "https://mcp.sequa.ai/v1/setup-code-assistant"
}
}
}
Claude Desktop (Settings → Developer → Edit Config)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
Windsurf (~/.codeium/windsurf/mcp_config.json)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
VS Code (.vscode/mcp.json)
{
"servers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
Cline / Claude Dev Tools (cline_mcp_settings.json)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
],
"disabled": false,
"autoApprove": []
}
}
}
Highlight AI (GUI → Plugins → Custom Plugin → Add using a command)
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Augment Code
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Or augment_config.json:
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
相关服务器
Scout Monitoring MCP
赞助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
赞助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Local MCP Test
A test server for local MCP development and setup.
Moralis
Interact with the Moralis Web3 API to access blockchain data and services.
chuk-mcp
A Python client for the Model Context Protocol (MCP), an open standard for connecting AI assistants to external data and tools.
Postman Tool Generation
Generates AI agent tools from Postman collections and requests using the Postman API.
Baby-SkyNet
An autonomous memory management system for Claude AI, featuring multi-provider LLM integration and a persistent memory database.
Cloudflare Remote MCP Server
A remote MCP server example deployable on Cloudflare Workers without authentication.
MCP-Logic
Provides automated reasoning for AI systems using the Prover9 and Mace4 theorem provers.
Stata-MCP
Perform regression analysis using Stata with the help of an LLM. Requires a local Stata installation and an external LLM API key.
Adios MCP
A remote MCP server deployable on Cloudflare Workers without authentication.
Shaka Packager MCP Server
Video transcoding, packaging, and analysis using the Shaka Packager tool, integrated with Claude AI.