Calculate MCP
Enables LLMs to perform basic calculations using a calculator interface.
Calculate MCP
A Model Context Protocol (MCP) server that provides browser automation capabilities using basic calculator feature. This server enables LLMs to interact with calculator. (I actually made it for a test program)
Use Cases
- The test code for to connect MCP feature.
- The toy projects
Example config
{
"mcpServers": {
"calculate": {
"command": "npx",
"args": [
"-y",
"@wrtnlabs/calculator-mcp@latest"
]
}
}
}
Installation in VS Code
Alternatively, you can install the Playwright MCP server using the VS Code CLI:
# For VS Code
code --add-mcp '{"name":"calculator","command":"npx","args":["-y", "@wrtnlabs/calculator-mcp@latest"]}'
# For VS Code Insiders
code-insiders --add-mcp '{"name":"calculator","command":"npx","args":["-y", "@wrtnlabs/calculator-mcp@latest"]}'
After installation, the Calculator MCP server will be available for use with your GitHub Copilot agent in VS Code.
CLI Options
The Calculator MCP server supports the following command-line options:
--port <port>: Port to listen on for SSE transport
Running headed browser on Linux w/o DISPLAY
When running headed browser on system w/o display or from worker processes of the IDEs,
run the MCP server from environment with the DISPLAY and pass the --port flag to enable SSE transport.
npx @wrtnlabs/calculator-mcp@latest --port 8931
And then in MCP client config, set the url to the SSE endpoint:
{
"mcpServers": {
"calculator": {
"url": "http://localhost:8931/sse"
}
}
}
Programmatic usage with custom transports
import { createServer } from "@wrtnlabs/calculator-mcp";
// ... other import statement
const client = new Client({
name: "test client",
version: "0.1.0",
});
const server = createServer({
name: "calculator",
version: "1.0.0"
});
const [clientTransport, serverTransport] = InMemoryTransport.createLinkedPair();
await Promise.all([
client.connect(clientTransport),
server.connect(serverTransport),
]);
Tools
- add
- sub
- mul
- div
- mod
- sqrt
İlgili Sunucular
Kone.vc
sponsorMonetize your AI agent with contextual product recommendations
Notion ReadOnly MCP Server
An optimized, read-only MCP server for the Notion API with asynchronous processing.
Obsidian MCP Server
An MCP server that allows AI assistants to read from and write to your local Obsidian vault.
Sequential Thinking
A server that facilitates structured, progressive thinking through defined stages.
Brivvy MCP
Connects AI assistants to your Brivvy workspace so they can use your brand voices, templates, and glossary for on-brand content—hosted at mcp.brivvy.io.
MCP Redmine
A server integration for the Redmine project management tool.
Trello MCP Server
Uses a Trello board as a knowledge base to store and retrieve code snippets, notes, and other information.
BrainCTL
Persistent memory for AI agents. Single SQLite file, 192 MCP tools. FTS5 search, knowledge graph, session handoffs, write gate. No server, no API keys, no LLM calls.
WhatsApp Claude Plugin
WhatsApp channel plugin for Claude Code. Connect WhatsApp as a native channel to your Claude Code session — send/receive messages, voice transcription, access control, and remote tool approval. No API keys needed, uses Baileys for WhatsApp Web connectivity.
Make
Turn your Make scenarios into callable tools for AI assistants.
Canvas MCP
Interact with Canvas LMS and Gradescope using AI agents.