Web Browser MCP Server
Provides advanced web browsing capabilities for AI applications.
✨ Features
🌐 Enable AI assistants to browse and extract content from the web through a simple MCP interface.
The Web Browser MCP Server provides AI models with the ability to browse websites, extract content, and understand web pages through the Message Control Protocol (MCP). It enables smart content extraction with CSS selectors and robust error handling.
🤝 Contribute • 📝 Report Bug
✨ Core Features
- 🎯 Smart Content Extraction: Target exactly what you need with CSS selectors
- ⚡ Lightning Fast: Built with async processing for optimal performance
- 📊 Rich Metadata: Capture titles, links, and structured content
- 🛡️ Robust & Reliable: Built-in error handling and timeout management
- 🌍 Cross-Platform: Works everywhere Python runs
🚀 Quick Start
Installing via Smithery
To install Web Browser Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install web-browser-mcp-server --client claude
Installing Manually
Install using uv:
uv tool install web-browser-mcp-server
For development:
# Clone and set up development environment
git clone https://github.com/blazickjp/web-browser-mcp-server.git
cd web-browser-mcp-server
# Create and activate virtual environment
uv venv
source .venv/bin/activate
# Install with test dependencies
uv pip install -e ".[test]"
🔌 MCP Integration
Add this configuration to your MCP client config file:
{
"mcpServers": {
"web-browser-mcp-server": {
"command": "uv",
"args": [
"tool",
"run",
"web-browser-mcp-server"
],
"env": {
"REQUEST_TIMEOUT": "30"
}
}
}
}
For Development:
{
"mcpServers": {
"web-browser-mcp-server": {
"command": "uv",
"args": [
"--directory",
"path/to/cloned/web-browser-mcp-server",
"run",
"web-browser-mcp-server"
],
"env": {
"REQUEST_TIMEOUT": "30"
}
}
}
}
💡 Available Tools
The server provides a powerful web browsing tool:
browse_webpage
Browse and extract content from web pages with optional CSS selectors:
# Basic webpage fetch
result = await call_tool("browse_webpage", {
"url": "https://example.com"
})
# Target specific content with CSS selectors
result = await call_tool("browse_webpage", {
"url": "https://example.com",
"selectors": {
"headlines": "h1, h2",
"main_content": "article.content",
"navigation": "nav a"
}
})
⚙️ Configuration
Configure through environment variables:
| Variable | Purpose | Default |
|---|---|---|
REQUEST_TIMEOUT | Webpage request timeout in seconds | 30 |
🧪 Testing
Run the test suite:
python -m pytest
📄 License
Released under the MIT License. See the LICENSE file for details.
Serveurs connexes
Bright Data
sponsorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
CodingBaby Browser
A Node.js server that enables AI assistants to control the Chrome browser via WebSocket. Requires the CodingBaby Chrome Extension.
Firecrawl
Scrape, crawl, and extract data from any website using the Firecrawl API.
Documentation Crawler
Crawl websites to generate Markdown documentation and make it searchable through an MCP server.
Trends MCP
Real-time trend data from Google (Search, Images, News, Shopping), YouTube, TikTok, Reddit, Amazon, Wikipedia, X (Twitter), LinkedIn, Spotify, GitHub, Steam, npm, App Store, news sentiment and web traffic via one MCP connection. Free API key, 20 requests/day, no credit card required.
SABIS MCP Server
Access academic grades from the Sakarya University SABIS system via automated web scraping.
Agentic Deep Researcher
A deep research agent powered by Crew AI and the LinkUp API.
rippr
YouTube transcript extraction for AI agents. Clean text, timestamps, or structured JSON from any video. No API keys required.
Postman API V3
MCP server for Postman API v3
Hyperbrowser
Hyperbrowser is the next-generation platform empowering AI agents and enabling effortless, scalable browser automation.
Mozilla Readability Parser
Extracts and transforms webpage content into clean, LLM-optimized Markdown using Mozilla's Readability algorithm.