LanceDB
Interact with on-disk documents using agentic RAG and hybrid search via LanceDB.
🗄️ LanceDB MCP Server for LLMS
A Model Context Protocol (MCP) server that enables LLMs to interact directly the documents that they have on-disk through agentic RAG and hybrid search in LanceDB. Ask LLMs questions about the dataset as a whole or about specific documents.
✨ Features
- 🔍 LanceDB-powered serverless vector index and document summary catalog.
- 📊 Efficient use of LLM tokens. The LLM itself looks up what it needs when it needs.
- 📈 Security. The index is stored locally so no data is transferred to the Cloud when using a local LLM.
🚀 Quick Start
To get started, create a local directory to store the index and add this configuration to your Claude Desktop config file:
MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json
Windows: %APPDATA%/Claude/claude_desktop_config.json
{
"mcpServers": {
"lancedb": {
"command": "npx",
"args": [
"lance-mcp",
"PATH_TO_LOCAL_INDEX_DIR"
]
}
}
}
Prerequisites
- Node.js 18+
- npx
- MCP Client (Claude Desktop App for example)
- Summarization and embedding models installed (see config.ts - by default we use Ollama models)
ollama pull snowflake-arctic-embed2ollama pull llama3.1:8b
Demo
Local Development Mode:
{
"mcpServers": {
"lancedb": {
"command": "node",
"args": [
"PATH_TO_LANCE_MCP/dist/index.js",
"PATH_TO_LOCAL_INDEX_DIR"
]
}
}
}
Use npm run build to build the project.
Use npx @modelcontextprotocol/inspector dist/index.js PATH_TO_LOCAL_INDEX_DIR to run the MCP tool inspector.
Seed Data
The seed script creates two tables in LanceDB - one for the catalog of document summaries, and another one - for vectorized documents' chunks. To run the seed script use the following command:
npm run seed -- --dbpath <PATH_TO_LOCAL_INDEX_DIR> --filesdir <PATH_TO_DOCS>
You can use sample data from the docs/ directory. Feel free to adjust the default summarization and embedding models in the config.ts file. If you need to recreate the index, simply rerun the seed script with the --overwrite option.
Catalog
- Document summary
- Metadata
Chunks
- Vectorized document chunk
- Metadata
🎯 Example Prompts
Try these prompts with Claude to explore the functionality:
"What documents do we have in the catalog?"
"Why is the US healthcare system so broken?"
📝 Available Tools
The server provides these tools for interaction with the index:
Catalog Tools
catalog_search: Search for relevant documents in the catalog
Chunks Tools
chunks_search: Find relevant chunks based on a specific document from the catalogall_chunks_search: Find relevant chunks from all known documents
📜 License
This project is licensed under the MIT License - see the LICENSE file for details.
Servidores relacionados
Chroma MCP Server
An MCP server for the Chroma embedding database, providing persistent, searchable working memory for AI-assisted development with features like automated context recall and codebase indexing.
MongoDB Mongoose MCP
An MCP server for MongoDB with optional Mongoose schema support.
Snowflake MCP Server by CData
A read-only MCP server for querying live Snowflake data, powered by CData.
Metabase Server
Integrates with Metabase for data visualization and business intelligence. Requires METABASE_URL, METABASE_USERNAME, and METABASE_PASSWORD environment variables.
AWS Athena MCP Server
An MCP server for querying and interacting with AWS Athena.
CData Zoho Books
A read-only MCP server by CData for querying live data from Zoho Books. Requires an external CData JDBC Driver.
MCP Database Server
Connects to and interacts with various database systems like SQLite, PostgreSQL, SQL Server, and MongoDB.
OpenSearch MCP Server
An MCP server for interacting with OpenSearch clusters, configured via environment variables.
Toronto Open Data Tools
Query, analyze, and retrieve datasets from Toronto's CKAN-powered open data portal.
Tabular MCP Server
An MCP server for local Tabular Models like PowerBI. It allows LLM clients to debug, analyze, and compose DAX queries by connecting to a local Tabular model instance.