Omilia MCP Tools
A set of tools for managing miniapps, orchestrator apps, and dialog logs on the Omilia Cloud Platform (OCP).
Omilia MCP Tools
This repository contains a set of tools for working with the Omilia Cloud Platform (OCP). These utilities help manage miniapps, orchestrator apps, and dialog logs.
Tools Overview
- search_miniapps: Search for miniapps by name or keyword.
- get_miniapp: Retrieve details for a specific miniapp using its ID.
- set_miniapp_prompt: Update prompts (welcome, error, reaction messages) for a miniapp.
- get_dialog_logs: Fetch logs for a specific dialog session.
- search_orchestrator_apps: Search for Orchestrator apps by keyword.
- get_orchestrator_app: Retrieve the canvas (nodes and edges) for an Orchestrator app by ID.
- search_dialog_logs: Search dialog logs with various filters (date, app, region, etc.).
- search_numbers: Search for phone numbers with optional search term.
- search_variable_collections: Search variable collections with optional search term.
- get_collection_variables: Get a list of all variables in a collection by ID.
Installation
- Make sure you have Python 3.10 or newer installed.
- Install uv.
- Clone this repository and navigate to the project directory.
- Copy the file
.env.exampleto.envand set the appropriate values. - Test if the istallation is correct by running
uv run mcp dev src/main.py. This should open the mcp development server. Click on connect and try it out.
Usage
You can use these tools in two main ways:
1. Self-hosting (MCP Python SDK)
You can run your own MCP server using the official Python MCP SDK. This is the most flexible option and is recommended for advanced users. For full instructions, see the MCP Python SDK README.
2. Local usage with Gemini CLI, Cursor, or Claude Desktop
You can also use this project locally with any MCP-compatible client, such as:
Each of these clients allows you to connect to local MCP servers. For more information, see their respective documentation:
- Gemini CLI: Configuring custom MCP servers
- Claude Desktop: MCP servers
- Cursor: Configuring custom MCP servers
Configuring MCP Servers
To use this project with any of the above clients, you need to configure your MCP servers. For example, you can use the following mcp.json configuration (place it in the appropriate config directory for your client):
{
"mcpServers": {
"Omilia MCP": {
"command": "uv",
"args": [
"run",
"--with",
"mcp",
"mcp",
"run",
"<path_to_cloned_repository>>/omilia-mcp/src/main.py"
],
"env": {
"PATH": "<depending on how you installed the needed tools you may need to paste your PATH here>"
}
}
}
}
Serveurs connexes
Scout Monitoring MCP
sponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
sponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
MasterGo Magic MCP
A standalone MCP service that connects MasterGo design tools with AI models, enabling them to retrieve DSL data directly from design files.
OpenAPI Schema
Exposes OpenAPI schema information to Large Language Models (LLMs). The server loads OpenAPI schema files specified via command line.
SAME (Stateless Agent Memory Engine
Your AI's memory shouldn't live on someone else's server — 12 MCP tools that give it persistent context from your local markdown, no cloud, no API keys, single binary.
Pickapicon
Quickly retrieve SVGs using the Iconify API, with no external data files required.
MCP Todo Server
A demo Todo application server built with a clean architecture using MCPServer and JSON Placeholder.
Limetest
A lightweight, AI-powered end-to-end testing framework for CI workflows. Requires an OpenAI API key.
Kafka MCP
A natural language interface to manage Apache Kafka operations.
Accordo MCP Server
Provides dynamic YAML-driven workflow guidance for AI coding agents with structured development workflows, progression control, and decision points.
Gemini Imagen 3.0
Generate high-quality images using Google's Imagen 3.0 model via the Gemini API.
MemGPT MCP Server
A server that provides a memory system for LLMs, enabling persistent conversations with various providers like OpenAI, Anthropic, and OpenRouter.