MemGPT MCP Server
A server that provides a memory system for LLMs, enabling persistent conversations with various providers like OpenAI, Anthropic, and OpenRouter.
MemGPT MCP Server
A TypeScript-based MCP server that implements a memory system for LLMs. It provides tools for chatting with different LLM providers while maintaining conversation history.
Features
Tools
-
chat- Send a message to the current LLM provider- Takes a message parameter
- Supports multiple providers (OpenAI, Anthropic, OpenRouter, Ollama)
-
get_memory- Retrieve conversation history- Optional
limitparameter to specify number of memories to retrieve - Pass
limit: nullfor unlimited memory retrieval - Returns memories in chronological order with timestamps
- Optional
-
clear_memory- Clear conversation history- Removes all stored memories
-
use_provider- Switch between different LLM providers- Supports OpenAI, Anthropic, OpenRouter, and Ollama
- Persists provider selection
-
use_model- Switch to a different model for the current provider- Supports provider-specific models:
- Anthropic Claude Models:
- Claude 3 Series:
claude-3-haiku: Fastest response times, ideal for tasks like customer support and content moderationclaude-3-sonnet: Balanced performance for general-purpose useclaude-3-opus: Advanced model for complex reasoning and high-performance tasks
- Claude 3.5 Series:
claude-3.5-haiku: Enhanced speed and cost-effectivenessclaude-3.5-sonnet: Superior performance with computer interaction capabilities
- Claude 3 Series:
- OpenAI: 'gpt-4o', 'gpt-4o-mini', 'gpt-4-turbo'
- OpenRouter: Any model in 'provider/model' format (e.g., 'openai/gpt-4', 'anthropic/claude-2')
- Ollama: Any locally available model (e.g., 'llama2', 'codellama')
- Anthropic Claude Models:
- Persists model selection
- Supports provider-specific models:
Development
Install dependencies:
npm install
Build the server:
npm run build
For development with auto-rebuild:
npm run watch
Installation
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"letta-memgpt": {
"command": "/path/to/memgpt-server/build/index.js",
"env": {
"OPENAI_API_KEY": "your-openai-key",
"ANTHROPIC_API_KEY": "your-anthropic-key",
"OPENROUTER_API_KEY": "your-openrouter-key"
}
}
}
}
Environment Variables
OPENAI_API_KEY- Your OpenAI API keyANTHROPIC_API_KEY- Your Anthropic API keyOPENROUTER_API_KEY- Your OpenRouter API key
Debugging
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
Recent Updates
Claude 3 and 3.5 Series Support (March 2024)
- Added support for latest Claude models:
- Claude 3 Series (Haiku, Sonnet, Opus)
- Claude 3.5 Series (Haiku, Sonnet)
Unlimited Memory Retrieval
- Added support for retrieving unlimited conversation history
- Use
{ "limit": null }with theget_memorytool to retrieve all stored memories - Use
{ "limit": n }to retrieve the n most recent memories - Default limit is 10 if not specified
Máy chủ liên quan
Alpha Vantage MCP Server
nhà tài trợAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Kitsune MCP
Shape-shifting MCP hub — shapeshift() into 10,000+ servers at runtime. One entry point, no restarts, 7 registries.
pageguard-mcp
Privacy compliance scanner for AI coding tools. Detects tracking technologies, cookies, and third-party data collection from local projects and live websites.
my-mcp-server
A template for building Model Context Protocol (MCP) servers using the mcp-framework for Node.js.
Bazel MCP Server
Exposes the Bazel build system to AI agents, enabling them to build, query, test, and manage dependencies.
MCP Servers
A collection of reference implementations for the Model Context Protocol (MCP), demonstrating how to give LLMs secure access to tools and data using Typescript and Python SDKs.
Coding Assistant Server
A coding assistant server that provides context-aware code suggestions, documentation integration, and technology detection.
PageLens AI
Your AI assistant, plugged into your PageLens audits.
MCPatterns
A server for storing and retrieving personalized coding patterns from a local JSONL file.
cmux-mcp
MCP server for controlling cmux (Ghostty-based terminal) via native CLI. Send commands, read output, send control characters — all in background via Unix socket.
Qwen-Agent
A framework for developing LLM applications with capabilities like tool usage, planning, and memory, based on the Qwen model.