Hyperbrowser
Hyperbrowser is the next-generation platform empowering AI agents and enabling effortless, scalable browser automation.
Hyperbrowser MCP Server
This is Hyperbrowser's Model Context Protocol (MCP) Server. It provides various tools to scrape, extract structured data, and crawl webpages. It also provides easy access to general purpose browser agents like OpenAI's CUA, Anthropic's Claude Computer Use, and Browser Use.
More information about the Hyperbrowser can be found here. The hyperbrowser API supports a superset of features present in the mcp server.
More information about the Model Context Protocol can be found here.
Table of Contents
Installation
Manual Installation
To install the server, run:
npx hyperbrowser-mcp <YOUR-HYPERBROWSER-API-KEY>
Running on Cursor
Add to ~/.cursor/mcp.json like this:
{
"mcpServers": {
"hyperbrowser": {
"command": "npx",
"args": ["-y", "hyperbrowser-mcp"],
"env": {
"HYPERBROWSER_API_KEY": "YOUR-API-KEY"
}
}
}
}
Running on Windsurf
Add to your ./codeium/windsurf/model_config.json like this:
{
"mcpServers": {
"hyperbrowser": {
"command": "npx",
"args": ["-y", "hyperbrowser-mcp"],
"env": {
"HYPERBROWSER_API_KEY": "YOUR-API-KEY"
}
}
}
}
Development
For development purposes, you can run the server directly from the source code.
-
Clone the repository:
git clone [email protected]:hyperbrowserai/mcp.git hyperbrowser-mcp cd hyperbrowser-mcp -
Install dependencies:
npm install # or yarn install npm run build -
Run the server:
node dist/server.js
Claude Desktop app
This is an example config for the Hyperbrowser MCP server for the Claude Desktop client.
{
"mcpServers": {
"hyperbrowser": {
"command": "npx",
"args": ["--yes", "hyperbrowser-mcp"],
"env": {
"HYPERBROWSER_API_KEY": "your-api-key"
}
}
}
}
Tools
scrape_webpage- Extract formatted (markdown, screenshot etc) content from any webpagecrawl_webpages- Navigate through multiple linked pages and extract LLM-friendly formatted contentextract_structured_data- Convert messy HTML into structured JSONsearch_with_bing- Query the web and get results with Bing searchbrowser_use_agent- Fast, lightweight browser automation with the Browser Use agentopenai_computer_use_agent- General-purpose automation using OpenAI’s CUA modelclaude_computer_use_agent- Complex browser tasks using Claude computer usecreate_profile- Creates a new persistent Hyperbrowser profile.delete_profile- Deletes an existing persistent Hyperbrowser profile.list_profiles- Lists existing persistent Hyperbrowser profiles.
Installing via Smithery
To install Hyperbrowser MCP Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install @hyperbrowserai/mcp --client claude
Resources
The server provides the documentation about hyperbrowser through the resources methods. Any client which can do discovery over resources has access to it.
License
This project is licensed under the MIT License.
Serveurs connexes
Bright Data
sponsorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
Intercept
Give your AI the ability to read the web. Fetches URLs as clean markdown with 9 fallback strategies. Handles tweets, YouTube, arXiv, PDFs, and regular pages.
Web-curl
Fetch, extract, and process web and API content. Supports resource blocking, authentication, and Google Custom Search.
Crypto News MCP Server
Fetches the latest cryptocurrency news and converts article content from HTML to Markdown.
MCP Webscan Server
Fetch, analyze, and extract information from web pages.
Playwright
Playwright MCP server
Outscraper
Extract data from Google Maps, including places and reviews, using the Outscraper API.
Any Browser MCP
Attaches to existing browser sessions using the Chrome DevTools Protocol for automation and interaction.
Wayback Machine
Access the Internet Archive's Wayback Machine to retrieve archived web pages and check for available snapshots of URLs.
MCP Web Scraper
A production-ready web scraping platform with ML-powered automation, browser automation via Playwright, and persistent caching.
Documentation Crawler
Crawl websites to generate Markdown documentation and make it searchable through an MCP server.