CrawlForge MCP
CrawlForge MCP is a production-ready MCP server with 18 web scraping tools for AI agents. It gives Claude, Cursor, and any MCP-compatible client the ability to fetch URLs, extract structured data with CSS/XPath selectors, run deep multi-step research, bypass anti-bot detection with TLS fingerprint randomization, process documents, monitor page changes, and more. Credit-based pricing with a free tier (1,000 credits/month, no credit card required).
CrawlForge MCP Server
Professional web scraping and content extraction server implementing the Model Context Protocol (MCP). Get started with 1,000 free credits - no credit card required!
🎯 Features
- 20 Professional Tools: Web scraping, deep research, stealth browsing, content analysis
- Free Tier: 1,000 credits to get started instantly
- MCP Compatible: Works with Claude, Cursor, and other MCP-enabled AI tools
- Enterprise Ready: Scale up with paid plans for production use
- Credit-Based: Pay only for what you use
🚀 Quick Start (2 Minutes)
1. Install from NPM
npm install -g crawlforge-mcp-server
2. Setup Your API Key
npx crawlforge-setup
This will:
- Guide you through getting your free API key
- Configure your credentials securely
- Auto-configure Claude Code and Cursor (if installed)
- Verify your setup is working
Don't have an API key? Get one free at https://www.crawlforge.dev/signup
3. Configure Your IDE (if not auto-configured)
🤖 For Claude Desktop
Add to claude_desktop_config.json:
{ "mcpServers": { "crawlforge": { "command": "npx", "args": ["crawlforge-mcp-server"] } } }
Location:
- macOS:
~/Library/Application Support/Claude/claude_desktop_config.json - Windows:
%APPDATA%/Claude/claude_desktop_config.json - Linux:
~/.config/Claude/claude_desktop_config.json
Restart Claude Desktop to activate.
🖥️ For Claude Code CLI (Auto-configured)
The setup wizard automatically configures Claude Code by adding to ~/.claude.json:
{ "mcpServers": { "crawlforge": { "type": "stdio", "command": "crawlforge" } } }
After setup, restart Claude Code to activate.
💻 For Cursor IDE (Auto-configured)
The setup wizard automatically configures Cursor by adding to ~/.cursor/mcp.json:
Restart Cursor to activate.
📊 Available Tools
Basic Tools (1 credit each)
fetch_url- Fetch content from any URLextract_text- Extract clean text from web pagesextract_links- Get all links from a pageextract_metadata- Extract page metadata
Advanced Tools (2-3 credits)
scrape_structured- Extract structured data with CSS selectorssearch_web- Search the web using Google Search APIsummarize_content- Generate intelligent summariesanalyze_content- Comprehensive content analysisextract_structured- LLM-powered schema-driven extractiontrack_changes- Monitor content changes over time
Premium Tools (5-10 credits)
crawl_deep- Deep crawl entire websitesmap_site- Discover and map website structurebatch_scrape- Process multiple URLs simultaneouslydeep_research- Multi-stage research with source verificationstealth_mode- Anti-detection browser management
Heavy Processing (3-10 credits)
process_document- Multi-format document processingextract_content- Enhanced content extractionscrape_with_actions- Browser automation chainsgenerate_llms_txt- Generate AI interaction guidelineslocalization- Multi-language and geo-location management
💳 Pricing
| Plan | Credits/Month | Best For |
|---|---|---|
| Free | 1,000 | Testing & personal projects |
| Starter | 5,000 | Small projects & development |
| Professional | 50,000 | Professional use & production |
| Enterprise | 250,000 | Large scale operations |
All plans include:
- Access to all 20 tools
- Credits never expire and roll over month-to-month
- API access and webhook notifications
View full pricing
🔧 Advanced Configuration
Environment Variables
Optional: Set API key via environment
export CRAWLFORGE_API_KEY="cf_live_your_api_key_here"
Optional: Custom API endpoint (for enterprise)
export CRAWLFORGE_API_URL="https://api.crawlforge.dev"
Manual Configuration
Your configuration is stored at ~/.crawlforge/config.json:
{ "apiKey": "cf_live_...", "userId": "user_...", "email": "[email protected]" }
📖 Usage Examples
Once configured, use these tools in your AI assistant:
"Search for the latest AI news"
"Extract all links from example.com"
"Crawl the documentation site and summarize it"
"Monitor this page for changes"
"Extract product prices from this e-commerce site"
🔒 Security & Privacy
- Secure Authentication: API keys required for all operations (no bypass methods)
- Local Storage: API keys stored securely at
~/.crawlforge/config.json - HTTPS Only: All connections use encrypted HTTPS
- No Data Retention: We don't store scraped data, only usage logs
- Rate Limiting: Built-in protection against abuse
- Compliance: Respects robots.txt and GDPR requirements
Security Updates
v3.0.3 (2025-10-01): Removed authentication bypass vulnerability. All users must authenticate with valid API keys.
🆘 Support
- Documentation: https://www.crawlforge.dev/docs
- Issues: GitHub Issues
- Email: [email protected]
- Discord: Join our community
📄 License
MIT License - see LICENSE file for details.
🤝 Contributing
Contributions are welcome! Please read our Contributing Guide first.
Built with ❤️ by the CrawlForge team
Website | Documentation | API Reference
Servidores relacionados
Bright Data
patrocinadorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
MCP Browser Console Capture Service
A browser automation service for capturing console output, useful for tasks like public sentiment analysis.
MCP YouTube Transcript Server
Retrieves transcripts from YouTube videos for content analysis and processing.
APIMesh
18 x402-payable web analysis APIs for AI agents — pay per call with USDC on Base, no API keys needed
PlayMCP Browser Automation Server
A server for browser automation using Playwright, providing powerful tools for web scraping, testing, and automation.
Parallel Task MCP
Perform Deep Research and Batch Tasks
Playwright
Provides browser automation capabilities using Playwright. Interact with web pages, take screenshots, and execute JavaScript in a real browser environment.
Cloudflare Playwright
Control a browser for web automation tasks like navigation, typing, clicking, and taking screenshots using Playwright on Cloudflare Workers.
Patchright Lite MCP Server
A server that wraps the Patchright SDK to provide stealth browser automation for AI models.
scrape-do-mcp
MCP Server for Scrape.do - Web Scraping & Google Search with anti-bot bypass
Browser Use MCP Server
An MCP server that allows AI agents to control a web browser using the browser-use library.