Artificial Analysis
An unofficial MCP server for the Artificial Analysis API
Artificial Analysis MCP Server
An MCP (Model Context Protocol) server that provides LLM model pricing, speed metrics, and benchmark scores from Artificial Analysis.
Features
- Get real-time pricing for 300+ LLM models (input/output/blended rates)
- Compare speed metrics (tokens/sec, time to first token)
- Access benchmark scores (Intelligence Index, Coding Index, MMLU-Pro, GPQA, and more)
- Filter by provider (OpenAI, Anthropic, Google, etc.)
- Sort by any metric
Installation
Claude Code
claude mcp add artificial-analysis -e AA_API_KEY=your-key -- npx -y artificial-analysis-mcp
Or install from GitHub:
claude /mcp add https://github.com/davidhariri/artificial-analysis-mcp
Manual Configuration
Add to your Claude settings (~/.claude/settings.json):
{
"mcpServers": {
"artificial-analysis": {
"command": "npx",
"args": ["-y", "artificial-analysis-mcp"],
"env": {
"AA_API_KEY": "your-api-key"
}
}
}
}
Configuration
| Environment Variable | Required | Description |
|---|---|---|
AA_API_KEY | Yes | Your Artificial Analysis API key |
Get your API key at artificialanalysis.ai.
Tools
list_models
List all available LLM models with optional filtering and sorting.
Parameters:
| Name | Type | Required | Description |
|---|---|---|---|
creator | string | No | Filter by model creator (e.g., "OpenAI", "Anthropic") |
sort_by | string | No | Sort field (see below) |
sort_order | string | No | "asc" or "desc" (default: "desc") |
limit | number | No | Maximum results to return |
Sort fields: price_input, price_output, price_blended, speed, ttft, intelligence_index, coding_index, math_index, mmlu_pro, gpqa, release_date
Example usage:
- "List the top 5 fastest models"
- "Show me Anthropic models sorted by price"
- "What are the cheapest models with high intelligence scores?"
get_model
Get detailed information about a specific model.
Parameters:
| Name | Type | Required | Description |
|---|---|---|---|
model | string | Yes | Model name or slug (e.g., "gpt-4o", "claude-4-5-sonnet") |
Returns: Complete model details including pricing, speed metrics, and all benchmark scores.
Example usage:
- "Get pricing for GPT-4o"
- "What are Claude 4.5 Sonnet's benchmark scores?"
Model Data
Each model includes:
- Pricing: Input/output/blended rates per 1M tokens (USD)
- Speed: Output tokens per second, time to first token
- Benchmarks: Intelligence Index, Coding Index, Math Index, MMLU-Pro, GPQA, LiveCodeBench, and more
Development
# Install dependencies
npm install
# Build
npm run build
# Run locally
AA_API_KEY=your-key node dist/index.js
License
MIT
Verwandte Server
Alpha Vantage MCP Server
SponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Cookiecutter MCP UV Container
A Cookiecutter template for creating MCP servers with Apple container support and configurable transport methods.
mcp-graphql
A GraphQL server that supports the Model Context Protocol (MCP), enabling Large Language Models (LLMs) to interact with GraphQL APIs through schema introspection and query execution.
Cursor Chat History MCP
Provides local access to Cursor chat history for AI analysis and insights, with no external services or API keys required.
Metal MCP Server
Search Metal Framework documentation and generate code.
PyPI Query MCP Server
A server to query the Python Package Index (PyPI) for package information, dependencies, and compatibility.
VectorMCP
A Ruby gem for building Model Context Protocol (MCP) servers to expose tools, resources, and prompts to LLM clients.
Gemma MCP Client
A client for Google's Gemma-3 model that enables function calling through MCP.
MCP Server
A backend service providing tools, resources, and prompts for AI models using the Model Context Protocol (MCP).
ChuckNorris
A specialized MCP gateway for LLM enhancement prompts and jailbreaks with dynamic schema adaptation. Provides prompts for different LLMs using an enum-based approach.
Clojars
Obtains latest dependency details for Clojure libraries.