Trustwise

Advanced evaluation tools for AI safety, alignment, and performance using the Trustwise API.

🦉 Trustwise MCP Server

The Trustwise MCP Server is a Model Context Protocol (MCP) server that provides a suite of advanced evaluation tools for AI safety, alignment, and performance. It enables developers and AI tools to programmatically assess the quality, safety, and cost of LLM outputs using Trustwise's industry-leading metrics.

💡 Use Cases

  • Evaluating the safety and reliability of LLM responses.
  • Measuring alignment, clarity, and helpfulness of AI-generated content.
  • Estimating the carbon footprint and cost of model inference.
  • Integrating robust evaluation into AI pipelines, agents, or orchestration frameworks.

🛠️ Prerequisites

📦 Installation & Running

Claude Desktop

To connect the Trustwise MCP Server to Claude Desktop, add the following configuration to your Claude Desktop settings:

{
  "mcpServers": {
    "trustwise": {
      "command": "docker",
      "args": [
        "run",
        "-i",
        "--rm",
        "-e",
        "TW_API_KEY",
        "ghcr.io/trustwiseai/trustwise-mcp-server:latest"
      ],
      "env": {
        "TW_API_KEY": "<YOUR_TRUSTWISE_API_KEY>"
      }
    }
  }
}

To point to a specific Trustwise Instance - under env, also set the following optional environment variable:

TW_BASE_URL: "<YOUR_TRUSTWISE_INSTANCE_URL>"

e.g "TW_BASE_URL": "https://api.yourdomain.ai"

Cursor

To connect the Trustwise MCP Server to cursor, add the following configuration to your cursor settings:

{
  "mcpServers": {
    "trustwise": {
      "command": "docker",
      "args": [
        "run",
        "-i",
        "--rm",
        "-e",
        "TW_API_KEY",
        "-e",
        "TW_BASE_URL",
        "ghcr.io/trustwiseai/trustwise-mcp-server:latest"
      ],
      "env": {
        "TW_API_KEY": "<YOUR_TRUSTWISE_API_KEY>"
      }
    }
  }
}

Replace <YOUR_TRUSTWISE_API_KEY> with your actual Trustwise API key.

🧰 Tools

The Trustwise MCP Server exposes the following tools (metrics). Each tool can be called with the specified arguments to evaluate a model response.

🛡️ Trustwise Metrics

Tool NameDescription
faithfulness_metricEvaluate the faithfulness of a response to its context
answer_relevancy_metricEvaluate relevancy of a response to the query
context_relevancy_metricEvaluate relevancy of context to the query
pii_metricDetect PII in a response
prompt_injection_metricDetect prompt injection risk
summarization_metricEvaluate summarization quality
clarity_metricEvaluate clarity of a response
formality_metricEvaluate formality of a response
helpfulness_metricEvaluate helpfulness of a response
sensitivity_metricEvaluate sensitivity of a response
simplicity_metricEvaluate simplicity of a response
tone_metricEvaluate tone of a response
toxicity_metricEvaluate toxicity of a response
carbon_metricEstimate carbon footprint of a response
cost_metricEstimate cost of a response

For more examples and advanced usage, see the official Trustwise SDK.

📄 License

This project is licensed under the terms of the MIT open source license. See LICENSE for details.

🔒 Security

  • Do not commit secrets or API keys.
  • This repository is public; review all code and documentation for sensitive information before pushing.

Related Servers