Discover, extract, and interact with the web - one interface powering automated access across the public internet.
Welcome to the official Bright Data Model Context Protocol (MCP) server, enabling LLMs, agents and apps to access, discover and extract web data in real-time. This server allows MCP clients, such as Claude Desktop, Cursor, Windsurf and others, to seamlessly search the web, navigate websites, take action and retrieve data - without getting blocked.
Feature | Description |
---|---|
search_engine | Scrape search results from Google, Bing or Yandex. Returns SERP results in markdown (URL, title, description) |
scrape_as_markdown | Scrape a single webpage URL with advanced options for content extraction and get back the results in MarkDown language. This tool can unlock any webpage even if it uses bot detection or CAPTCHA. |
scrape_as_html | Scrape a single webpage URL with advanced options for content extraction and get back the results in HTML. This tool can unlock any webpage even if it uses bot detection or CAPTCHA. |
session_stats | Tell the user about the tool usage during this session |
web_data_amazon_product | Quickly read structured amazon product data. Requires a valid product URL with /dp/ in it. This can be a cache lookup, so it can be more reliable than scraping |
web_data_amazon_product_reviews | Quickly read structured amazon product review data. Requires a valid product URL with /dp/ in it. This can be a cache lookup, so it can be more reliable than scraping |
web_data_linkedin_person_profile | Quickly read structured linkedin people profile data. This can be a cache lookup, so it can be more reliable than scraping |
web_data_linkedin_company_profile | Quickly read structured linkedin company profile data. This can be a cache lookup, so it can be more reliable than scraping |
web_data_zoominfo_company_profile | Quickly read structured ZoomInfo company profile data. Requires a valid ZoomInfo company URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_instagram_profiles | Quickly read structured Instagram profile data. Requires a valid Instagram URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_instagram_posts | Quickly read structured Instagram post data. Requires a valid Instagram URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_instagram_reels | Quickly read structured Instagram reel data. Requires a valid Instagram URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_instagram_comments | Quickly read structured Instagram comments data. Requires a valid Instagram URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_facebook_posts | Quickly read structured Facebook post data. Requires a valid Facebook post URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_facebook_marketplace_listings | Quickly read structured Facebook marketplace listing data. Requires a valid Facebook marketplace listing URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_facebook_company_reviews | Quickly read structured Facebook company reviews data. Requires a valid Facebook company URL and number of reviews. This can be a cache lookup, so it can be more reliable than scraping |
web_data_x_posts | Quickly read structured X post data. Requires a valid X post URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_zillow_properties_listing | Quickly read structured zillow properties listing data. Requires a valid zillow properties listing URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_booking_hotel_listings | Quickly read structured booking hotel listings data. Requires a valid booking hotel listing URL. This can be a cache lookup, so it can be more reliable than scraping |
web_data_youtube_videos | Quickly read structured YouTube videos data. Requires a valid YouTube video URL. This can be a cache lookup, so it can be more reliable than scraping |
scraping_browser_navigate | Navigate a scraping browser session to a new URL |
scraping_browser_go_back | Go back to the previous page |
scraping_browser_go_forward | Go forward to the next page |
scraping_browser_click | Click on an element. Avoid calling this unless you know the element selector (you can use other tools to find those) |
scraping_browser_links | Get all links on the current page, text and selectors. It's strongly recommended that you call the links tool to check that your click target is valid |
scraping_browser_type | Type text into an element |
scraping_browser_wait_for | Wait for an element to be visible on the page |
scraping_browser_screenshot | Take a screenshot of the current page |
scraping_browser_get_html | Get the HTML content of the current page. Avoid using the full_page option unless it is important to see things like script tags since this can be large |
scraping_browser_get_text | Get the text content of the current page |
Install nodejs
to get the npx
command (node.js module runner). Installation instructions can be found on the node.js website
Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:
{
"mcpServers": {
"Bright Data": {
"command": "npx",
"args": ["@brightdata/mcp"],
"env": {
"API_TOKEN": "<insert-your-api-token-here>",
"WEB_UNLOCKER_ZONE": "<optional if you want to override the default mcp_unlocker zone name>",
"BROWSER_AUTH": "<optional if you want to enable remote browser control tools>"
}
}
}
}
Make sure you have an account on brightdata.com (new users get free credit for testing, and pay as you go options are available)
Get your API key from the user settings page
Create a Web Unlocker proxy zone called mcp_unlocker
in your control panel
WEB_UNLOCKER_ZONE
(Optional) To enable browser control tools:
brd-customer-[your-customer-ID]-zone-[your-zone-ID]:[your-password]
To use this MCP server with other agent types, you should adapt the following to your specific software:
npx @brightdata/mcp
API_TOKEN=<your-token>
must exist when running the serverWant to try Bright Data MCP without setting up anything?
Check out this playground:
This platform provide an easy way to explore the capabilities of Bright Data MCP without any local setup. Just sign in and start experimenting with web data collection!
Some example queries that this MCP server will be able to help with:
The videos below demonstrate a minimal use case for Claude Desktop:
Some tools can involve reading web data, and the amount of time needed to load the page can vary by quite a lot in extreme circumstances.
To ensure that your agent will be able to consume the data, set a high enough timeout in your agent settings.
A value of 180s
should be enough for 99% of requests, but some sites load slower than others, so tune this to your needs.
This error occurs when your system cannot find the npx
command. To fix it:
macOS:
which node
Shows path like /usr/local/bin/node
Windows:
where node
Shows path like C:\Program Files\nodejs\node.exe
Replace the npx
command with the full path to Node, for example, on mac, it will look as follows:
"command": "/usr/local/bin/node"
We welcome contributions to help improve the Bright Data MCP! Here's how you can help:
For major changes, please open an issue first to discuss your proposed changes. This ensures your time is well spent and aligned with project goals.
If you encounter any issues or have questions, please reach out to the Bright Data support team or open an issue in the repository.
Web content fetching and conversion for efficient LLM usage
Browser automation and web scraping
Enable AI agents to get structured data from unstructured web with AgentQL.
Actors MCP Server: Use 3,000+ pre-built cloud tools to extract data from websites, e-commerce, social media, search engines, maps, and more
Automate browser interactions in the cloud (e.g. web navigation, data extraction, form filling, and more)
Extract web data with Firecrawl
Scrape websites with Oxylabs Web API, supporting dynamic rendering and parsing for structured data extraction.
Integrate real-time Scrapeless Google SERP(Google Search, Google Flight, Google Map, Google Jobs....) results into your LLM applications. This server enables dynamic context retrieval for AI workflows, chatbots, and research tools.
Turn websites into datasets with Scrapezy
Interact with WebScraping.AI for web data extraction and scraping.