JinaAI Reader
Extracts web content using the Jina.ai Reader API.
mcp-jinaai-reader
⚠️ Notice
This repository is no longer maintained.
The functionality of this tool is now available in mcp-omnisearch, which combines multiple MCP tools in one unified package.
Please use mcp-omnisearch instead.
A Model Context Protocol (MCP) server for integrating Jina.ai's Reader API with LLMs. This server provides efficient and comprehensive web content extraction capabilities, optimized for documentation and web content analysis.
Features
- 📚 Advanced web content extraction through Jina.ai Reader API
- 🚀 Fast and efficient content retrieval
- 📄 Complete text extraction with preserved structure
- 🔄 Clean format optimized for LLMs
- 🌐 Support for various content types including documentation
- 🏗️ Built on the Model Context Protocol
Configuration
This server requires configuration through your MCP client. Here are examples for different environments:
Cline Configuration
Add this to your Cline MCP settings:
{
"mcpServers": {
"jinaai-reader": {
"command": "node",
"args": ["-y", "mcp-jinaai-reader"],
"env": {
"JINAAI_API_KEY": "your-jinaai-api-key"
}
}
}
}
Claude Desktop with WSL Configuration
For WSL environments, add this to your Claude Desktop configuration:
{
"mcpServers": {
"jinaai-reader": {
"command": "wsl.exe",
"args": [
"bash",
"-c",
"JINAAI_API_KEY=your-jinaai-api-key npx mcp-jinaai-reader"
]
}
}
}
Environment Variables
The server requires the following environment variable:
JINAAI_API_KEY: Your Jina.ai API key (required)
API
The server implements a single MCP tool with configurable parameters:
read_url
Convert any URL to LLM-friendly text using Jina.ai Reader.
Parameters:
url(string, required): URL to processno_cache(boolean, optional): Bypass cache for fresh results. Defaults to falseformat(string, optional): Response format ("json" or "stream"). Defaults to "json"timeout(number, optional): Maximum time in seconds to wait for webpage loadtarget_selector(string, optional): CSS selector to focus on specific elementswait_for_selector(string, optional): CSS selector to wait for specific elementsremove_selector(string, optional): CSS selector to exclude specific elementswith_links_summary(boolean, optional): Gather all links at the end of responsewith_images_summary(boolean, optional): Gather all images at the end of responsewith_generated_alt(boolean, optional): Add alt text to images lacking captionswith_iframe(boolean, optional): Include iframe content in response
Development
Setup
- Clone the repository
- Install dependencies:
npm install
- Build the project:
npm run build
- Run in development mode:
npm run dev
Publishing
- Update version in package.json
- Build the project:
npm run build
- Publish to npm:
npm publish
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
License
MIT License - see the LICENSE file for details.
Acknowledgments
- Built on the Model Context Protocol
- Powered by Jina.ai Reader API
Servidores relacionados
Bright Data
patrocinadorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
News MCP Server
Real-time news aggregation from AP, BBC, NPR, Hacker News, and Google News
AI Shopping Assistant
A conversational AI shopping assistant for web-based product discovery and decision-making.
Airbnb MCP Server
Search for Airbnb listings and retrieve detailed information without an API key.
Clawpage
Extract and structure any web page into clean JSON.
Selenium MCP Server
Control web browsers using the Selenium WebDriver for automation and testing.
Oxylabs
Scrape websites with Oxylabs Web API, supporting dynamic rendering and parsing for structured data extraction.
LinkRescue
MCP server that exposes LinkRescue's broken link scanning, monitoring, and fix suggestion capabilities to AI agents (Claude, Cursor, etc.).
Haunt API
Extract clean, structured data from any URL — directly from Claude, Cursor, or any MCP-compatible AI.
Puppeteer
Browser automation using Puppeteer, with support for local, Docker, and Cloudflare Workers deployments.
Documentation Crawler
Crawl websites to generate Markdown documentation and make it searchable through an MCP server.