Sequa MCP
A proxy that connects local STDIO with remote MCP servers, enabling IDEs to use MCP without extra infrastructure.
Sequa MCP
Self Maintaining Docs for Devs & AI — Turn your codebase into always-current documentation, so engineers and AI never read outdated docs again.
This repository is the entry point for using Sequa via the Model Context Protocol (MCP). Add Sequa as an MCP server to Cursor, Claude, Windsurf, VSCode, Cline, Highlight, Augment, or any other MCP‑capable client.
🤔 What is Sequa?
70% of docs are never updated as underlying facts change, and 30% of dev time is lost searching, handling interruptions, and reworking. Documentation drifts out of sync, knowledge gets scattered across tools, and experts repeat themselves. The result: onboarding slows, context is lost, and teams waste time chasing answers instead of building.
Sequa fixes this. It's a Contextual Knowledge Engine that:
- Generates docs that stay accurate — architecture overviews, API docs, runbooks, onboarding guides
- Continuously updates from your code — eliminating stale pages and manual upkeep
- Unifies knowledge across repositories — no more hunting across Confluence, Slack, and GitHub
- Powers your team and AI agents — ask in plain language, get current answers with source links
Why use Sequa MCP?
Stop stitching context for Copilot and Cursor. With Sequa MCP, your AI tools know your entire codebase and docs out of the box.
Sequa gives your AI assistants deep, real-time knowledge of your codebase, docs, tickets, and internal standards. Instead of writing code you have to fix, they write code that fits — first try.
🧱 Prerequisites
Before you configure any AI agent:
- Create / sign in to your Sequa account at https://app.sequa.ai/login
- Setup a Project inside the Sequa app
- Inside that project, locate the MCP Setup URLs and select the transport your AI agent supports
- Copy the URL or configuration and install it in your client
❗ If you skip project creation the MCP server will refuse connections — the proxy can launch but you will receive auth / project errors.
🚀 Quick Start (Proxy Launch)
NPX (most common)
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Replace the URL if you use an endpoint from the specific project
🔌 IDE / Tool Configuration
Cursor (~/.cursor/mcp.json)
{
"mcpServers": {
"sequa": {
"url": "https://mcp.sequa.ai/v1/setup-code-assistant"
}
}
}
Claude Desktop (Settings → Developer → Edit Config)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
Windsurf (~/.codeium/windsurf/mcp_config.json)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
VS Code (.vscode/mcp.json)
{
"servers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
Cline / Claude Dev Tools (cline_mcp_settings.json)
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
],
"disabled": false,
"autoApprove": []
}
}
}
Highlight AI (GUI → Plugins → Custom Plugin → Add using a command)
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Augment Code
npx -y @sequa-ai/sequa-mcp@latest https://mcp.sequa.ai/v1/setup-code-assistant
Or augment_config.json:
{
"mcpServers": {
"sequa": {
"command": "npx",
"args": [
"-y",
"@sequa-ai/sequa-mcp@latest",
"https://mcp.sequa.ai/v1/setup-code-assistant"
]
}
}
}
Servidores relacionados
Alpha Vantage MCP Server
patrocinadorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Explorium API
Interact with the Explorium API to access external data for machine learning.
PAMPA
An MCP server for intelligent semantic search and automatic learning within codebases, allowing AI agents to efficiently query and index project artifacts.
Logfire
Provides access to OpenTelemetry traces and metrics through Logfire.
Signet
Cryptographic action receipts for AI agents. Signs every MCP tool call with Ed25519, hash-chained audit log. 3 lines of code to integrate.
POX MCP Server
An MCP server for the POX SDN controller, enabling network control, management, and analysis using Python and OpenFlow.
MCP Server
A backend service providing tools, resources, and prompts for AI models using the Model Context Protocol (MCP).
Azure DevOps
Interact with Azure DevOps for managing projects, pipelines, and repositories.
MCP Stdio-HTTP Proxy
A TypeScript proxy that connects stdio MCP clients to HTTP SSE MCP servers, handling OAuth authentication.
WhichModel
Cost-optimised LLM model routing for autonomous agents
即梦AI多模态MCP
A multimodal generation service using Volcengine Jimeng AI for image generation, video generation, and image-to-video conversion.