WolframAlpha LLM
Answer math and other queries using the WolframAlpha LLM API.
WolframAlpha LLM MCP Server
A Model Context Protocol (MCP) server that provides access to WolframAlpha's LLM API. https://products.wolframalpha.com/llm-api/documentation

Features
- Query WolframAlpha's LLM API with natural language questions
- Answer complicated mathematical questions
- Query facts about science, physics, history, geography, and more
- Get structured responses optimized for LLM consumption
- Support for simplified answers and detailed responses with sections
Available Tools
ask_llm: Ask WolframAlpha a question and get a structured llm-friendly responseget_simple_answer: Get a simplified answervalidate_key: Validate the WolframAlpha API key
Installation
git clone https://github.com/Garoth/wolframalpha-llm-mcp.git
npm install
Configuration
-
Get your WolframAlpha API key from developer.wolframalpha.com
-
Add it to your Cline MCP settings file inside VSCode's settings (ex. ~/.config/Code/User/globalStorage/saoudrizwan.claude-dev/settings/cline_mcp_settings.json):
{
"mcpServers": {
"wolframalpha": {
"command": "node",
"args": ["/path/to/wolframalpha-mcp-server/build/index.js"],
"env": {
"WOLFRAM_LLM_APP_ID": "your-api-key-here"
},
"disabled": false,
"autoApprove": [
"ask_llm",
"get_simple_answer",
"validate_key"
]
}
}
}
Development
Setting Up Tests
The tests use real API calls to ensure accurate responses. To run the tests:
-
Copy the example environment file:
cp .env.example .env -
Edit
.envand add your WolframAlpha API key:WOLFRAM_LLM_APP_ID=your-api-key-hereNote: The
.envfile is gitignored to prevent committing sensitive information. -
Run the tests:
npm test
Building
npm run build
License
MIT
Related Servers
GPT Researcher
Conducts autonomous, in-depth research by exploring and validating multiple sources to provide relevant and up-to-date information.
Legislative Yuan API
Search for bills, documents, and meeting records from Taiwan's Legislative Yuan API.
Powertools for AWS MCP
Search the Powertools for AWS Lambda documentation across multiple runtimes to find documentation and examples.
Tavily
Search engine for AI agents (search + extract) powered by Tavily
Google Search Console
An MCP server for accessing Google Search Console data, including site performance and indexing status.
Greptile
Code search and querying using the Greptile API.
Ollama Deep Researcher
Conducts deep research using local Ollama LLMs, leveraging Tavily and Perplexity for comprehensive search capabilities.
Vectorize
Vectorize MCP server for advanced retrieval, Private Deep Research, Anything-to-Markdown file extraction and text chunking.
Readeck MCP
An MCP server for advanced research assistance, configurable via environment variables.
Wttr Weather
Fetches weather data from the wttr.in service.