Langflow Document Q&A Server
A document question-and-answer server powered by Langflow.
Langflow-DOC-QA-SERVER
A Model Context Protocol server for document Q&A powered by Langflow
This is a TypeScript-based MCP server that implements a document Q&A system. It demonstrates core MCP concepts by providing a simple interface to query documents through a Langflow backend.
Prerequisites
1. Create Langflow Document Q&A Flow
- Open Langflow and create a new flow from the "Document Q&A" template
- Configure your flow with necessary components (ChatInput, File Upload, LLM, etc.)
- Save your flow
2. Get Flow API Endpoint
- Click the "API" button in the top right corner of Langflow
- Copy the API endpoint URL from the cURL command
Example:
http://127.0.0.1:7860/api/v1/run/<flow-id>?stream=false - Save this URL as it will be needed for the
API_ENDPOINTconfiguration
Features
Tools
query_docs- Query the document Q&A system- Takes a query string as input
- Returns responses from the Langflow backend
Development
Install dependencies:
npm install
Build the server:
npm run build
For development with auto-rebuild:
npm run watch
Installation
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"langflow-doc-qa-server": {
"command": "node",
"args": [
"/path/to/doc-qa-server/build/index.js"
],
"env": {
"API_ENDPOINT": "http://127.0.0.1:7860/api/v1/run/480ec7b3-29d2-4caa-b03b-e74118f35fac"
}
}
}
}
Installing via Smithery
To install Document Q&A Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install @GongRzhe/Langflow-DOC-QA-SERVER --client claude
Environment Variables
The server supports the following environment variables for configuration:
API_ENDPOINT: The endpoint URL for the Langflow API service. Defaults tohttp://127.0.0.1:7860/api/v1/run/480ec7b3-29d2-4caa-b03b-e74118f35facif not specified.
Debugging
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
📜 License
This project is licensed under the MIT License.
Related Servers
Cala
Cala turns internet chaos into structured, verified knowledge that AI agents and LLMs can call as a tool.
SearchAPI Agent
An MCP agent that integrates various search tools using the SearchAPI service. Requires SearchAPI and Google API keys.
Web Search
A server that provides web search capabilities using OpenAI models.
Perplexity MCP Zerver
Interact with Perplexity.ai using Puppeteer without an API key. Requires Node.js and stores chat history locally.
EU Regulations MCP
Query 37 EU regulations (DORA, NIS2, GDPR, AI Act, CRA) with full-text search, cross-regulation comparison, and ISO 27001/NIST CSF control mappings. Auto-updates via EUR-Lex monitoring.
NCBI Literature Search
Search NCBI databases, including PubMed, for scientific literature. Tailored for researchers in life sciences, evolutionary biology, and computational biology.
news-aggregator-mcp-server
Multi-source news aggregation for AI agents — RSS/Atom feeds (16 sources), HackerNews, and GDELT global news intelligence in 65+ languages. No API key required.
Teleport Documentation
Search and query Teleport's documentation using embeddings stored in a local Chroma vector database.
MCP Lucene Server
MCP Lucene Server is a Model Context Protocol (MCP) server that exposes Apache Lucene's full-text search capabilities through a conversational interface. It allows AI assistants (like Claude) to help users search, index, and manage document collections without requiring technical knowledge of Lucene or search engines.
Jina AI Search
Access Jina AI's Search Foundation APIs for web search, news search, and more, tailored for LLMs.