Higress AI-Search MCP Server
Provides an AI search tool to enhance AI model responses with real-time search results from various search engines using the Higress ai-search feature.
Higress AI-Search MCP Server
Overview
A Model Context Protocol (MCP) server that provides an AI search tool to enhance AI model responses with real-time search results from various search engines through Higress ai-search feature.
Demo
Cline
https://github.com/user-attachments/assets/60a06d99-a46c-40fc-b156-793e395542bb
Claude Desktop
https://github.com/user-attachments/assets/5c9e639f-c21c-4738-ad71-1a88cc0bcb46
Features
- Internet Search: Google, Bing, Quark - for general web information
- Academic Search: Arxiv - for scientific papers and research
- Internal Knowledge Search
Prerequisites
Configuration
The server can be configured using environment variables:
HIGRESS_URL(optional): URL for the Higress service (default:http://localhost:8080/v1/chat/completions).MODEL(required): LLM model to use for generating responses.INTERNAL_KNOWLEDGE_BASES(optional): Description of internal knowledge bases.
Option 1: Using uvx
Using uvx will automatically install the package from PyPI, no need to clone the repository locally.
{
"mcpServers": {
"higress-ai-search-mcp-server": {
"command": "uvx",
"args": [
"higress-ai-search-mcp-server"
],
"env": {
"HIGRESS_URL": "http://localhost:8080/v1/chat/completions",
"MODEL": "qwen-turbo",
"INTERNAL_KNOWLEDGE_BASES": "Employee handbook, company policies, internal process documents"
}
}
}
}
Option 2: Using uv with local development
Using uv requires cloning the repository locally and specifying the path to the source code.
{
"mcpServers": {
"higress-ai-search-mcp-server": {
"command": "uv",
"args": [
"--directory",
"path/to/src/higress-ai-search-mcp-server",
"run",
"higress-ai-search-mcp-server"
],
"env": {
"HIGRESS_URL": "http://localhost:8080/v1/chat/completions",
"MODEL": "qwen-turbo",
"INTERNAL_KNOWLEDGE_BASES": "Employee handbook, company policies, internal process documents"
}
}
}
}
License
This project is licensed under the MIT License - see the LICENSE file for details.
関連サーバー
Readeck MCP
An MCP server for advanced research assistance, configurable via environment variables.
Agently MCP
Discover public A2A agents on the Agently platform using its public API.
Hotel Booking
Search and book from over 2 million hotels with shopping and booking capabilities.
Lancelot-MCP
A containerized MCP server for LanceDB vector search, featuring hybrid processing with Gemini and Ollama.
Transit MCP API
Real-time transit, maritime, and aviation telemetry for AI agents.
Joblyst MCP
One search to get german projects and jobs from different plattforms
Financial AI Agent
An AI agent providing unified access to financial market data and news articles.
MCP Knowledge Base
A knowledge base server that processes local documents (PDF, DOCX, TXT, HTML) and answers questions based on their content using similarity search.
gnosis-mcp
Zero-config knowledge base for AI coding agents. Loads your markdown docs into a searchable database and exposes them as MCP tools — search, read, and manage documentation without leaving your editor. Works instantly with SQLite (no setup), upgrades to PostgreSQL + pgvector for hybrid semantic search. Includes skills for searching docs (/gnosis:search), health checks (/gnosis:status), doc management (/gnosis:manage), and first-time setup (/gnosis:setup). 6 MCP tools, 3 resources, FTS5 keyword search, 176 tests.
Perplexity MCP Server
Perform real-time internet research with source citations using the Perplexity API.
