Comet Opik
resmiQuery and analyze your Opik logs, traces, prompts and all other telemtry data from your LLMs in natural language.
Model Context Protocol (MCP) server for Opik, with both local stdio and remote streamable-http transports.
Website • Slack community • Twitter • Documentation
[!IMPORTANT] This repository ships the MCP server implementation only. We do not currently provide a hosted remote MCP service for Opik. If you run
streamable-httpremotely, authentication is fail-closed by default.
Why this server
Opik MCP Server gives MCP-compatible clients one interface for:
- Prompt lifecycle management
- Workspace, project, and trace exploration
- Metrics and dataset operations
- MCP resources and resource templates for metadata-aware flows
Quickstart
1. Run with npx
# Opik Cloud
npx -y opik-mcp --apiKey YOUR_API_KEY
For self-hosted Opik, pass --apiUrl (for example http://localhost:5173/api) and use your local auth strategy.
2. Add to your MCP client
Cursor (.cursor/mcp.json):
{
"mcpServers": {
"opik": {
"command": "npx",
"args": ["-y", "opik-mcp", "--apiKey", "YOUR_API_KEY"]
}
}
}
VS Code / GitHub Copilot (.vscode/mcp.json):
{
"inputs": [
{
"type": "promptString",
"id": "opik-api-key",
"description": "Opik API Key",
"password": true
}
],
"servers": {
"opik-mcp": {
"type": "stdio",
"command": "npx",
"args": ["-y", "opik-mcp", "--apiKey", "${input:opik-api-key}"]
}
}
}
Windsurf (raw config):
{
"mcpServers": {
"opik": {
"command": "npx",
"args": ["-y", "opik-mcp", "--apiKey", "YOUR_API_KEY"]
}
}
}
More client-specific examples: docs/ide-integration.md
Run from source
git clone https://github.com/comet-ml/opik-mcp.git
cd opik-mcp
npm install
npm run build
Optional local config:
cp .env.example .env
Start the server:
npm run start:stdio
npm run start:http
Transport modes
| Transport | Use case | Command |
|---|---|---|
stdio | Local MCP integration (same machine as client) | npm run start:stdio |
streamable-http | Remote/self-hosted MCP endpoint (/mcp) | npm run start:http |
Remote auth defaults (streamable-http)
Authorization: Bearer <OPIK_API_KEY>orx-api-keyis required by default.- Workspace is resolved server-side (token map recommended); workspace headers are not trusted by default.
- In remote mode, request-context workspace takes precedence over tool
workspaceName. - Missing or invalid auth returns HTTP
401.
Key environment flags:
STREAMABLE_HTTP_REQUIRE_AUTH(defaulttrue)STREAMABLE_HTTP_VALIDATE_REMOTE_AUTH(defaulttrue, except test env)REMOTE_TOKEN_WORKSPACE_MAP(JSON token-to-workspace map)STREAMABLE_HTTP_TRUST_WORKSPACE_HEADERS(defaultfalse)
Deep dive: docs/streamable-http-transport.md
Toolsets
Toolsets let you narrow which capabilities are enabled:
coreintegrationexpert-promptsexpert-datasetsexpert-trace-actionsexpert-project-actionsmetricsall(enables all modern toolsets)
Configure via:
- CLI:
--toolsets all - Env:
OPIK_TOOLSETS=core,expert-prompts,metrics
Details: docs/configuration.md
MCP resources and prompts
resources/listexposes static URIs (for exampleopik://workspace-info)resources/templates/listexposes dynamic URI templates (for exampleopik://projects/{page}/{size})resources/readsupports static and templated URIsprompts/listandprompts/getexpose workflow prompts
Development
# Lint
npm run lint
# Test
npm test
# Build
npm run build
# Run precommit checks
make precommit
Documentation
Contributing
Please read CONTRIBUTING.md before opening a PR.
Citation
If you use this project in research, cite:
Comet ML, Inc, Koc, V., & Boiko, Y. (2025). Opik MCP Server. Github. https://doi.org/10.5281/zenodo.15411156
BibTeX:
@software{CometML_Opik_MCP_Server_2025,
author = {{Comet ML, Inc} and Koc, V. and Boiko, Y.},
title = {{Opik MCP Server}},
year = {2025},
publisher = {GitHub},
url = {https://doi.org/10.5281/zenodo.15411156},
doi = {10.5281/zenodo.15411156}
}
Citation metadata is also available in CITATION.cff.
License
Apache 2.0
Server Terkait
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
AgentChatBus
AgentChatBus is a persistent AI communication bus that lets multiple independent AI Agents chat, collaborate, and delegate tasks — across terminals, across IDEs, and across frameworks.
cxpak
Spends CPU cycles so you don't spend tokens. The LLM gets a briefing packet instead of a flashlight in a dark room.
AI Sessions
Searching and access your AI coding sessions from Claude Code, Gemini CLI, opencode, and OpenAI Codex.
JetBrains
Work on your code with JetBrains IDEs
FMP MCP Server
Provides tools, resources, and prompts for financial analysis using the Financial Modelling Prep API.
pabal-store-api-mcp
MCP server that provides App Store / Play Store ASO workflows as tools.
Kai
Kai provides a bridge between large language models (LLMs) and your Kubernetes clusters, enabling natural language interaction with Kubernetes resources. The server exposes a comprehensive set of tools for managing clusters, namespaces, pods, deployments, services, and other Kubernetes resources
MCP Yeoman Server
Search for and run Yeoman generator templates programmatically.
Smithery Reference Servers
A collection of reference implementations for Model Context Protocol (MCP) servers in Typescript and Python, demonstrating MCP features and SDK usage.
Recraft AI
Generate and edit raster/vector images, vectorize, remove/replace backgrounds, and upscale using the Recraft AI API.