Trustwise
Advanced evaluation tools for AI safety, alignment, and performance using the Trustwise API.
🦉 Trustwise MCP Server
The Trustwise MCP Server is a Model Context Protocol (MCP) server that provides a suite of advanced evaluation tools for AI safety, alignment, and performance. It enables developers and AI tools to programmatically assess the quality, safety, and cost of LLM outputs using Trustwise's industry-leading metrics.
💡 Use Cases
- Evaluating the safety and reliability of LLM responses.
- Measuring alignment, clarity, and helpfulness of AI-generated content.
- Estimating the carbon footprint and cost of model inference.
- Integrating robust evaluation into AI pipelines, agents, or orchestration frameworks.
🛠️ Prerequisites
- A Trustwise API Key (get one here)
- Docker; Follow the install instructions
📦 Installation & Running
Claude Desktop
To connect the Trustwise MCP Server to Claude Desktop, add the following configuration to your Claude Desktop settings:
{
"mcpServers": {
"trustwise": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"TW_API_KEY",
"ghcr.io/trustwiseai/trustwise-mcp-server:latest"
],
"env": {
"TW_API_KEY": "<YOUR_TRUSTWISE_API_KEY>"
}
}
}
}
To point to a specific Trustwise Instance - under env, also set the following optional environment variable:
TW_BASE_URL: "<YOUR_TRUSTWISE_INSTANCE_URL>"
e.g "TW_BASE_URL": "https://api.yourdomain.ai"
Cursor
To connect the Trustwise MCP Server to cursor, add the following configuration to your cursor settings:
{
"mcpServers": {
"trustwise": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e",
"TW_API_KEY",
"-e",
"TW_BASE_URL",
"ghcr.io/trustwiseai/trustwise-mcp-server:latest"
],
"env": {
"TW_API_KEY": "<YOUR_TRUSTWISE_API_KEY>"
}
}
}
}
Replace <YOUR_TRUSTWISE_API_KEY> with your actual Trustwise API key.
🧰 Tools
The Trustwise MCP Server exposes the following tools (metrics). Each tool can be called with the specified arguments to evaluate a model response.
🛡️ Trustwise Metrics
| Tool Name | Description |
|---|---|
faithfulness_metric | Evaluate the faithfulness of a response to its context |
answer_relevancy_metric | Evaluate relevancy of a response to the query |
context_relevancy_metric | Evaluate relevancy of context to the query |
pii_metric | Detect PII in a response |
prompt_injection_metric | Detect prompt injection risk |
summarization_metric | Evaluate summarization quality |
clarity_metric | Evaluate clarity of a response |
formality_metric | Evaluate formality of a response |
helpfulness_metric | Evaluate helpfulness of a response |
sensitivity_metric | Evaluate sensitivity of a response |
simplicity_metric | Evaluate simplicity of a response |
tone_metric | Evaluate tone of a response |
toxicity_metric | Evaluate toxicity of a response |
refusal_metric | Detect refusal to answer or comply with the query |
completion_metric | Evaluate completion of the query’s instruction |
adherence_metric | Evaluate adherence to a given policy or instruction |
stability_metric | Evaluate stability (consistency) of multiple responses |
carbon_metric | Estimate carbon footprint of a response |
cost_metric | Estimate cost of a response |
For more examples and advanced usage, see the official Trustwise SDK.
📄 License
This project is licensed under the terms of the MIT open source license. See LICENSE for details.
🔒 Security
- Do not commit secrets or API keys.
- This repository is public; review all code and documentation for sensitive information before pushing.
Похожие серверы
Scout Monitoring MCP
спонсорPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
спонсорAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Remote MCP Server on Cloudflare
An example of a remote MCP server deployable on Cloudflare Workers, without authentication.
Qase
Interact with the Qase API for test management. Requires a QASE_API_TOKEN for authentication.
MCP Gemini CLI
A command-line interface wrapper for the Google Gemini API, enabling interaction with Gemini's Search and Chat tools.
Jupyter Notebook MCP Server
Interact with Jupyter notebooks, allowing for code execution, cell manipulation, and notebook management.
MCP WordPress Server
A comprehensive MCP server for managing WordPress sites, featuring a wide range of tools for performance monitoring, caching, and more.
Argo CD
Interact with Argo CD applications through natural language.
Skills-ContextManager
Don’t pollute your AI agent’s context with 1,000 skills. Use Skills-ContextManager, a self-hosted web UI for managing AI skills and workflows by providing skills to an AI agent via MCP only when needed. Simply add skills to your library and enable or disable them with a toggle. Choose whether a skill is always loaded into context or dynamically activated when the AI agent determines it’s needed.
Node.js Sandbox MCP Server
Run arbitrary JavaScript in an isolated Docker container with on-the-fly npm dependency installation.
MCP Remote Machine Control
Provides remote machine control capabilities, eliminating SSH overhead for token-efficient system operations.
kemdiCode MCP
kemdiCode MCP is a Model Context Protocol server that gives AI agents and IDE assistants access to 124 specialized tools for code analysis, generation, git operations, file management, AST-aware editing, project memory, cognition & self-improvement, multi-board kanban, and multi-agent coordination.