MCP ZepAi Server
A server for Zep, a long-term memory store for AI applications, requiring a ZEP_API_KEY for access.
MCP ZepAi Server
This repository contains the MCP (Model Context Protocol) server for the ZepAi module. The server is configured to run using a specific command and environment variables as detailed below.
Configuration
Below is an example configuration:
{
"mcpServers": {
"zepai": {
"command": "uv",
"args": [
"run",
"--directory",
"<path_to_the_mcp_directory>",
"python",
"-m",
"mcp_zepai"
],
"cwd": "<path_to_the_mcp_directory>",
"env": {
"ZEP_API_KEY": "<your_api_key_here>"
}
}
}
}
Key Configuration Elements
- command: The command to execute the server, in this case, using
uv. - args: Arguments passed to the command, including the directory and module to run.
- cwd: The current working directory for the server.
- env: Environment variables required for the server, such as
ZEP_API_KEY.
Running the Server
To run the MCP ZepAi server, ensure that the configuration file is correctly set up and execute the following command:
/opt/homebrew/bin/uv run --directory <path_to_mcp_directory> python -m mcp_zepai
Ensure that the ZEP_API_KEY environment variable is set with a valid API key.
相关服务器
Scout Monitoring MCP
赞助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
赞助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
TestRail
Interact with TestRail's core entities such as test cases, runs, and results using a standardized protocol.
mcp-ssh-sre
An MCP server providing read-only server monitoring tools to AI assistants. Runs predefined diagnostic commands over SSH and passes only the results to the LLM - your server credentials and shell are never exposed.
Adobe After Effects MCP
An MCP server that allows AI assistants to interact with Adobe After Effects.
Postman MCP Server
Run Postman collections using Newman, with support for environment and global variables.
Sionic AI Serverless RAG
Integrates LLM applications with RAG data sources and tools using the Model Context Protocol.
302AI Basic MCP Server
A toolkit for enhancing the fundamental capabilities of large language models.
Jetbrains Debugger MCP Plugin
Allows AI assistants to remotely drive the JetBrains debugger via MCP, including breakpoints, stepping, and variable inspection.
Framer Plugin MCP Server
Create and manage Framer plugins with web3 capabilities.
My First MCP
A Model Context Protocol server implemented in Node.js with Docker support.
MCP Server with Ollama Integration
An MCP server that integrates with Ollama to provide tools for file operations, calculations, and text processing. Requires a running Ollama instance.