Decodo
официальныйEasy web data access. Simplified retrieval of information from websites and online sources.
Decodo MCP Server
This repository provides a Model Context Protocol (MCP) server that connects LLMs and applications to Decodo's platform. The server facilitates integration between MCP-compatible clients and Decodo's services, streamlining access to our tools and capabilities.
Features
- Easy web data access. Simplified retrieval of information from websites and online sources.
- Geographic flexibility. Access content regardless of regional restrictions.
- Enhanced privacy. Browse and collect data while maintaining anonymity.
- Reliable scraping. Advanced techniques to avoid detection and blocks.
- Simple integration. Seamless setup with popular MCP clients like Claude Desktop, Cursor, and Windsurf.
Authenticating
-
Go to decodo.com and start a Web Scraping API plan (free plan available).
-
Once your plan has started, obtain a Web Scraping API basic authentication token from the dashboard.
Connecting to Decodo's MCP server (Claude Code, Cursor)
- Open your preferred MCP client and add the following configuration:
{
"Decodo": {
"url": "https://mcp.decodo.com/mcp",
"headers": {
"Authorization": "Basic <basic_auth_token>"
}
}
}
Claude Desktop
{
"mcpServers": {
"Decodo MCP Server": {
"command": "npx",
"args": ["@decodo/mcp"],
"env": {
"SCRAPER_API_TOKEN": "<web_scraping_api_base64_token>",
"TOOLSETS": "web,ai"
}
}
}
}
Running the MCP server locally
Prerequisites
- Node.js 18.0+
- An MCP client - popular choices are Claude Desktop and Cursor
Step-by-step guide
- Clone this repository:
git clone https://github.com/Decodo/decodo-mcp-server
- Run the following commands in the terminal:
cd decodo-mcp-server
npm install
npm run build
- Take note of your build location:
cd build/
pwd
Adding index.js to the end of this directory, your build file location should look something like
this:
/Users/your.user/projects/decodo-mcp/build/index.js
- Update your MCP client with the server information:
{
"mcpServers": {
"decodo-mcp": {
"command": "node",
"args": ["/Users/your.user/projects/decodo-mcp/build/index.js"],
"env": {
"SCRAPER_API_TOKEN": "<web_scraping_api_base64_token>"
}
}
}
}
Toolsets
Tools are organized into toolsets. You can selectively enable specific toolsets by passing a
comma-separated list via the toolsets query parameter:
"Decodo MCP Server": {
"url": "https://mcp.decodo.com/mcp?toolsets=web,ai",
"headers": {
"Authorization": "Basic <your_auth_token>"
}
}
When no toolsets are specified, all tools are registered.
| Toolset | Tools |
|---|---|
web | scrape_as_markdown, screenshot |
search | google_search, google_ads, google_lens, google_travel_hotels, bing_search |
ecommerce | amazon_search, amazon_product, amazon_pricing, amazon_sellers, amazon_bestsellers, walmart_search, walmart_product, target_search, target_product, tiktok_shop_search, tiktok_shop_product, tiktok_shop_url |
social_media | reddit_post, reddit_subreddit, reddit_user, tiktok_post, youtube_metadata, youtube_channel, youtube_subtitles, youtube_search |
ai | chatgpt, perplexity, google_ai_mode |
Tools
The server exposes the following tools:
| Tool | Description | Example prompt |
|---|---|---|
scrape_as_markdown | Scrapes any target URL, expects a URL to be given via prompt. Returns results in Markdown. | Scrape peacock.com from a US IP address and tell me the pricing. |
screenshot | Captures a screenshot of any webpage and returns it as a PNG image. | Take a screenshot of github.com from a US IP address. |
google_search | Scrapes Google Search for a given query, and returns parsed results. | Scrape Google Search for shoes and tell me the top position. |
google_ads | Scrapes Google Ads search results. | Scrape Google Ads for laptop and show me the top ads. |
google_lens | Scrapes Google Lens image search results. | Search Google Lens for this image: https://example.com/image.jpg |
google_ai_mode | Scrapes Google AI Mode (Search with AI) results. | Ask Google AI Mode: What are the top three dog breeds? |
google_travel_hotels | Scrapes Google Travel Hotels search results. | Search Google Travel Hotels for hotels in Paris. |
amazon_search | Scrapes Amazon Search for a given query, and returns parsed results. | Scrape Amazon Search for wireless keyboard. |
amazon_product | Scrapes Amazon Product page. | Scrape Amazon product B09H74FXNW and show me the details. |
amazon_pricing | Scrapes Amazon Product pricing information. | Get pricing for Amazon product B09H74FXNW. |
amazon_sellers | Scrapes Amazon Seller information. | Get information about Amazon seller A1R0Z7FJGTKESH. |
amazon_bestsellers | Scrapes Amazon Bestsellers list. | Show me Amazon bestsellers in electronics. |
walmart_search | Scrapes Walmart Search for a given query, and returns parsed results. | Scrape Walmart Search for camping tent. |
walmart_product | Scrapes Walmart Product page. | Scrape Walmart product 15296401808. |
target_search | Scrapes Target Search for a given query, and returns parsed results. | Scrape Target Search for kitchen appliances. |
target_product | Scrapes Target Product page. | Scrape Target product 92186007. |
tiktok_post | Scrapes a TikTok post URL for structured data (e.g. engagement, caption, hashtags). | Scrape this TikTok post: https://www.tiktok.com/@nba/video/7393013274725403950 |
tiktok_shop_search | Scrapes TikTok Shop Search for a given query, and returns parsed results. | Scrape TikTok Shop Search for phone cases. |
tiktok_shop_product | Scrapes TikTok Shop Product page. | Scrape TikTok Shop product 1731541214379741272. |
tiktok_shop_url | Scrapes TikTok Shop page by URL. | Scrape this TikTok Shop URL: https://www.tiktok.com/shop/s?q=HEADPHONES |
youtube_metadata | Scrapes YouTube video metadata. | Get metadata for YouTube video dFu9aKJoqGg. |
youtube_channel | Scrapes YouTube channel videos. | Scrape YouTube channel @decodo_official. |
youtube_subtitles | Scrapes YouTube video subtitles. | Get subtitles for YouTube video L8zSWbQN-v8. |
youtube_search | Search YouTube videos. | Search YouTube for "How to care for chinchillas". |
reddit_post | Scrapes a specific Reddit post. | Scrape the following Reddit post: https://www.reddit.com/r/horseracing/comments/1nsrn3/ |
reddit_subreddit | Scrapes Reddit subreddit results. | Scrape the top 5 posts on r/Python this week. |
reddit_user | Scrapes a Reddit user profile and their posts/comments. | Scrape this Reddit user: https://www.reddit.com/user/IWasRightOnce/ |
bing_search | Scrapes Bing Search results. | Search Bing for laptop reviews. |
chatgpt | Search and interact with ChatGPT for AI-powered responses and conversations. | Ask ChatGPT to explain quantum computing in simple terms. |
perplexity | Search and interact with Perplexity for AI-powered responses and conversations. | Ask Perplexity what the latest trends in web development are. |
Parameters
The following parameters are inferred from user prompts:
| Parameter | Description |
|---|---|
jsRender | Renders target URL in a headless browser. |
geo | Sets the country from which the request will originate. |
locale | Sets the locale of the request. |
tokenLimit | Truncates the response content up to this limit. Useful if the context window is small. |
prompt | Prompt to send to AI tools (chatgpt, perplexity). |
search | Activates ChatGPT's web search functionality (chatgpt only). |
xhr | When true, includes XHR or fetch responses in the scrape result where supported (e.g. tiktok_post). |
deviceType | Device type to emulate for the request (desktop, mobile, tablet). |
domain | Domain to use for the request (e.g., amazon.com, amazon.co.uk, bing.com). |
pageFrom | Starting page number for pagination. |
deliveryZip | ZIP code for delivery location (Target, Walmart). |
storeId | Store ID for local inventory (Target, Walmart). |
country | Country for TikTok Shop requests. |
limit | Maximum number of results to return (e.g., YouTube channel videos). |
language_code | Language code for subtitles (e.g., en, es). |
Examples
Scraping geo-restricted content
Query your AI agent with the following prompt:
Scrape peacock.com from a German IP address and tell me the pricing.
This prompt will say that peacock.com is geo-restricted. To bypass the geo-restriction:
Scrape peacock.com from a US IP address and tell me the pricing.
Limiting number of response tokens
If your agent has a small context window, the content returned from scraping will be automatically truncated, in order to avoid context-overflow. You can increase the number of tokens returned within your prompt:
Scrape hacker news, return 50k tokens.
If your agent has a big context window, tell it to return full content:
Scrape hacker news, return full content.
Related repositories
License
All code is released under the MIT License.
Похожие серверы
Bright Data
спонсорDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
Automatic MCP Discovery
AI powered automation toolkit which acts as an agent that discovers MCP servers for you. Point it at GitHub/npm/configure your own discovery, let GPT or Claude analyze the API or MCP or any tool, get ready-to-ship plugin configs. Zero manual work.
Query Table
A financial web table crawler using Playwright that queries data from multiple websites with fallback switching.
BrowserCat
Automate remote browsers using the BrowserCat API.
Social APIS Hub
The unified API for social media data - built for developers and AI agents.
BrowserAct
BrowserAct MCP Server is a standardized MCP service that lets MCP clients connect to the BrowserAct platform to discover and run browser automation workflows, access results/files and related storage, and trigger real-world actions via natural language.
Shufersal MCP Server
Automates shopping on the Shufersal website, enabling LLMs to search for products, create shopping lists, and manage the cart.
ShopGraph
Structured product data from the open web — where platform APIs don't reach. Schema.org + AI extraction. Pay per call via Stripe MPP.
LinkedIn MCP
Scrape LinkedIn profiles and companies, get recommended jobs, and perform job searches.
HasData
HasData APIs - Google SERP, Amazon, Zillow, Indeed, Maps, and more
Web Scout
An MCP server for web search and content extraction using DuckDuckGo.