Unichat
A unified interface for various chat AI models including OpenAI, MistralAI, Anthropic, and Google AI, requiring vendor API keys.
Unichat MCP Server in Python
Also available in TypeScript
Send requests to OpenAI, MistralAI, Anthropic, xAI, Google AI, DeepSeek, Alibaba, Inception using MCP protocol via tool or predefined prompts. Vendor API key required
Tools
The server implements one tool:
unichat: Send a request to unichat- Takes "messages" as required string arguments
- Returns a response
Prompts
code_review- Review code for best practices, potential issues, and improvements
- Arguments:
code(string, required): The code to review"
document_code- Generate documentation for code including docstrings and comments
- Arguments:
code(string, required): The code to comment"
explain_code- Explain how a piece of code works in detail
- Arguments:
code(string, required): The code to explain"
code_rework- Apply requested changes to the provided code
- Arguments:
changes(string, optional): The changes to apply"code(string, required): The code to rework"
Quickstart
Install
Claude Desktop
On MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
Supported Models:
A list of currently supported models to be used as
"SELECTED_UNICHAT_MODEL"may be found here. Please make sure to add the relevant vendor API key as"YOUR_UNICHAT_API_KEY"
Example:
"env": {
"UNICHAT_MODEL": "gpt-4o-mini",
"UNICHAT_API_KEY": "YOUR_OPENAI_API_KEY"
}
Development/Unpublished Servers Configuration
"mcpServers": {
"unichat-mcp-server": {
"command": "uv",
"args": [
"--directory",
"{{your source code local directory}}/unichat-mcp-server",
"run",
"unichat-mcp-server"
],
"env": {
"UNICHAT_MODEL": "SELECTED_UNICHAT_MODEL",
"UNICHAT_API_KEY": "YOUR_UNICHAT_API_KEY"
}
}
}
Published Servers Configuration
"mcpServers": {
"unichat-mcp-server": {
"command": "uvx",
"args": [
"unichat-mcp-server"
],
"env": {
"UNICHAT_MODEL": "SELECTED_UNICHAT_MODEL",
"UNICHAT_API_KEY": "YOUR_UNICHAT_API_KEY"
}
}
}
Installing via Smithery
To install Unichat for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install unichat-mcp-server --client claude
Development
Building and Publishing
To prepare the package for distribution:
- Remove older builds:
rm -rf dist
- Sync dependencies and update lockfile:
uv sync
- Build package distributions:
uv build
This will create source and wheel distributions in the dist/ directory.
- Publish to PyPI:
uv publish --token {{YOUR_PYPI_API_TOKEN}}
Debugging
Since MCP servers run over stdio, debugging can be challenging. For the best debugging experience, we strongly recommend using the MCP Inspector.
You can launch the MCP Inspector via npm with this command:
npx @modelcontextprotocol/inspector uv --directory {{your source code local directory}}/unichat-mcp-server run unichat-mcp-server
Upon launching, the Inspector will display a URL that you can access in your browser to begin debugging.
Server Terkait
WebRTC MCP Chat Server
A remote WebRTC chat server with secure temporary rooms and MCP support for background agents.
Reddit
Interact with Reddit by posting comments and searching for posts.
Prompt for User Input MCP Server
Enables AI models to prompt users for input directly within their code editor for interactive conversations.
Unipile
Access and manage messages from multiple platforms using the Unipile API.
any-chat-completions-mcp
Chat with any other OpenAI SDK Compatible Chat Completions API, like Perplexity, Groq, xAI and more
Beyond MCP Server
Provides standardized access to social platform and onchain data using the Neynar API.
WhatsApp
Search, read, and send WhatsApp messages and contacts. Requires a local Go WhatsApp bridge.
MCP Server SSH Client
A local MCP server that securely connects to remote servers over SSH, acting as a proxy for MCP commands.
Slack MCP Client in Go
A Slack bot client that bridges Slack and Model Context Protocol (MCP) servers, enabling LLMs to use MCP tools.
MCP Email Server
Provides IMAP and SMTP access for sending and receiving emails.