urlDNA
Dynamically scan and analyze potentially malicious URLs using the urlDNA.io
urlDNA MCP Server

The urlDNA MCP Server enables native tool use for security-focused LLM agents like OpenAI GPT-4.1 and Claude Desktop, providing a direct interface to interact with the urlDNA threat intelligence platform via API.
Installation & Setup
This project uses uv for fast Python package management.
Prerequisites
Install uv if you haven't already:
# On macOS and Linux
curl -LsSf https://astral.sh/uv/install.sh | sh
# On Windows
powershell -c "irm https://astral.sh/uv/install.ps1 | iex"
# Or with pip
pip install uv
Quick Start
- Clone and setup the project:
git clone <repository-url>
cd urlDNA-mcp-server
uv sync
- Run the MCP server locally (stdio mode):
uv run python urldna_mcp/run.py
- Run the MCP server in SSE mode:
uv run python urldna_mcp/server.py
Development
# Install development dependencies
uv sync --dev
# Run tests (when available)
uv run pytest
# Format code
uv run black .
# Type checking
uv run mypy .
# Lint code
uv run flake8 .
Hosted MCP Server
The urlDNA MCP server is already hosted and available at:
https://mcp.urldna.io/sse
This server is accessible over Server-Sent Events (SSE) protocol, which supports streaming interactions between LLMs and the backend tools.
You can use it directly with any platform or LLM that supports the MCP specification (e.g., Claude Desktop, OpenAI GPT-4.1).
Supported Tools
Scanning
| Tool | Description |
|---|---|
fast_check | Instantly check if a URL has been scanned. Returns SAFE / MALICIOUS / UNRATED. |
new_scan | Submit a URL for a full scan and wait for the result (~30–60s). |
get_scan | Retrieve a complete scan result by ID. |
Search
| Tool | Description |
|---|---|
search | Search scans using CQL (Custom Query Language) across domain, IP, technology, malicious flag, and more. Supports pagination (page 2+ requires PREMIUM). |
Saved Queries
| Tool | Description |
|---|---|
list_queries | List all saved queries for the authenticated user. |
get_query | Retrieve a specific saved query and its filters by ID. |
create_query | Create a new saved query with one or more CQL filter conditions. |
update_query | Update an existing query's name and filters (full replacement). |
delete_query | Permanently delete a saved query by ID. |
query_scans | Retrieve all matching scans for a saved query. |
Brand Monitoring
| Tool | Description |
|---|---|
list_brands | List available brands with optional name search and visibility filter (ALL / FREE / PREMIUM / USER_BRANDS). |
get_brand | Retrieve full details of a specific brand by ID. |
brand_scans | Get all scans associated with a brand. Supports additional CQL filtering. |
API Reference
| Tool | Description |
|---|---|
search_docs | Fetch the full urlDNA OpenAPI and documentations. |
Integration with Claude Desktop
To integrate the urlDNA MCP server in Claude Desktop, update your claude_desktop_config.json:
{
"mcpServers": {
"urlDNA": {
"command": "uv",
"args": [
"--directory",
"C:\\Users\\pripamonti\\urlDNA\\mcp\\urldna_mcp",
"run",
"run.py"
],
"env": {
"x-api-key": "<urlDNA_API_KEY>"
}
}
}
}
Replace
<YOUR_PATH>with the actual path to the project directory and<urlDNA_API_KEY>with your API key from https://urldna.io.
Once configured, you can prompt Claude with natural language, for example:
"Search in urlDNA for malicious scans with title like paypal"
"Create a saved query for mobile scans from Italy that are flagged as malicious"
"Show me all scans associated with the Google brand"
Claude will automatically call the correct tool and return results from the urlDNA platform.
Using the MCP Server with OpenAI GPT-4.1
from openai import OpenAI
# Initialize OpenAI client (assumes OPENAI_API_KEY is set via environment variable)
client = OpenAI()
response = client.responses.create(
model="gpt-4.1", # GPT-4.1 supports native MCP tool use
input=[
{
"role": "system",
"content": [{"type": "input_text", "text": "You are a cybersecurity analyst using urlDNA."}]
},
{
"role": "user",
"content": [{"type": "input_text", "text": "Search in urlDNA for malicious scans with title like paypal"}]
}
],
text={"format": {"type": "text"}},
reasoning={},
tools=[
{
"type": "mcp",
"server_label": "urlDNA",
"server_url": "https://mcp.urldna.io/sse",
"headers": {
"x-api-key": "<URLDNA_API_KEY>" # Replace with your urlDNA API key
},
"allowed_tools": [
# --- Scanning ---
"new_scan", # Submit a URL for a full scan and wait for the result
"get_scan", # Retrieve a scan result by ID
"fast_check", # Lightweight instant safety check (SAFE / MALICIOUS / UNRATED)
# --- Search ---
"search", # Search scans using CQL (Custom Query Language)
# --- Saved Queries (PREMIUM) ---
"list_queries",
"get_query",
"create_query",
"update_query",
"delete_query",
"query_scans",
# --- Brand Monitoring (PREMIUM) ---
"list_brands",
"get_brand",
"brand_scans",
# --- API Reference ---
"search_docs",
],
"require_approval": "never"
}
],
temperature=0.7,
top_p=1,
max_output_tokens=2048,
store=True
)
print(response.output)
Container Deployment
Build and run with Docker:
# Build the container
docker build -t urldna-mcp-server .
# Run the server
docker run -p 8080:8080 -e x-api-key=<URLDNA_API_KEY> urldna-mcp-server
Contributing
- Fork the repository
- Create a feature branch:
git checkout -b feature-name - Install development dependencies:
uv sync --dev - Make your changes and ensure tests pass
- Format code:
uv run black . - Submit a pull request
Contact & Support
For support or API access, visit https://urldna.io or email [email protected].
Servidores relacionados
Bright Data
patrocinadorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
Career Site Jobs
A MCP server to retrieve up-to-date jobs from company career sites.
Web Fetch
Fetches and converts web content, ideal for data extraction and web scraping.
WebScraping.AI
Interact with WebScraping.AI for web data extraction and scraping.
SABIS MCP Server
Access academic grades from the Sakarya University SABIS system via automated web scraping.
Web-curl
Fetch, extract, and process web and API content. Supports resource blocking, authentication, and Google Custom Search.
LinkedIn Profile Scraper
Fetches LinkedIn profile information using the Fresh LinkedIn Profile Data API.
MCP RSS Crawler
Fetches and caches RSS feeds using a SQLite database for use with LLMs via the MCP protocol.
Steel Puppeteer
Provides browser automation capabilities using Puppeteer and Steel, configurable for local or cloud instances.
Fetch
Fetch web content as HTML, JSON, plain text, or Markdown.
Chrome Debug
Automate Chrome via its debugging port with session persistence. Requires Chrome to be started with remote debugging enabled.