mcp-nodejs
A Node.js MCP server example for the OpenWeather API, requiring an API key.
Bring the model context protocol (MCP) in action:
- Generate an MCP server (Node.js) from Postman
- Deploy the server locally
- Test the server from a local MCP client (Cursor)
The MCP server in this example connects to the OpenWeather API.
Actions:
Develop
Generate an MCP server, using Postman.
See the sections below for a description of the actions taken.
Install Postman
sudo snap install postman
Start Postman
postman
Login with an account > free plan.
Generate
In Postman:
- API network > MCP Generator
- Search > Openweathermap (by API Evangelist)
- Select APIs > Add Requests
- Generate
- Download ZIP
- Unzip
The API requests in this example serve as the "tools" that an LLM can use.
Build
In the ./src directory:
npm install
Deploy (MCP server)
Start the MCP server and configure it with an API key to access the underlying API.
See the sections below for a description of the actions taken.
Start
In the ./src directory:
node mcpServer.js
Configure
In this example, we use the OpenWeather API.
In OpenWeather:
- Create an account or sign-in
- Save the API key
- https://openweathermap.org/api
Copy the API key into src/.env (not checked in in this repo):
- OPENWEATHERMAP_API_KEY=your_copied_key
Restart
In the ./src directory:
node mcpServer.js
Deploy (MCP client)
Install and configure Cursor as an MCP client that embeds the above deployed MCP server into its models.
In this example, no specific model binding is selected (we just use the Cursor defaults).
See the sections below for a description of the actions taken.
Install Cursor
Install:
Observe preconfigured models:
- Cursor settings > Models
Configure Cursor
Add the running MCP server to Cursor.
- In Cursor, select Cursor Settings > Tools & Integrations > MCP tools > New MCP Server
- Add args: "your-parent-dirs/mcpServer.js"
{ "mcpServers": { "weather-mcp-agent": { "command": "node", "args": ["your-parent-dirs/mcp-nodejs/src/mcpServer.js"] } } }
Observe a green dot, indicating that the MCP-server is ready to use.

Sometimes it seems additionally selecting the disable/enable switch is required.
Test
Ask a question to the models:
- Ctrl-I > "What's the weather in Amsterdam" > Accept
Based on the question, the LLM internally reasons and decides which tool to use in order to produce an answer.

Credits
This example is based on:
Verwandte Server
Scout Monitoring MCP
SponsorPut performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
SponsorAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
MCP Server
Automate data science stages using your own CSV data files.
FDEP MCP Server
A static code analysis server for enterprise-scale Haskell codebases, providing over 40 comprehensive analysis tools.
Docker
Run and manage docker containers, docker compose, and logs
LLM API Benchmark MCP Server
Benchmark LLM APIs for throughput and Time To First Token (TTFT) under various concurrency levels.
Lenses
Manage, explore, transform and join data across multiple clusters using different flavours of Apache Kafka via Lenses.io (including the free Community Edition)
ConfigCat
interacts with ConfigCat feature flag platform. Supports managing feature flags, configs, environments, products and organizations. Helps to integrate ConfigCat SDK, implement feature flags or remove zombie (stale) flags.
Tripwire
Context injection for AI agents via MCP. Define path-based policies in YAML — when an agent reads a matching file, relevant knowledge is auto-injected. Prevents mistakes before they happen. Works with Claude Code, Cursor, and any MCP client.
Odoo XML-RPC MCP Server
Interact with Odoo instances using the XML-RPC API. Requires configuration via environment variables or config files.
openapi-to-mcp
Expose API endpoints as strongly typed tools from an OpenAPI specification. Supports OpenAPI 2.0/3.0 in JSON or YAML format, from local or remote files.
CodeGraphContext
An MCP server that indexes local code into a graph database to provide context to AI assistants.