SQLite
MCP server for SQLite files. Supports Datasette-compatible metadata!
mcp-sqlite
Provide useful data to AI agents without giving them access to external systems. Compatible with Datasette for human users!
Features
- AI agents can get the structure of all tables and columns in the SQLite database in one command -
sqlite_get_catalog.- The catalog can be enriched with descriptions for the tables and columns using a simple YAML or JSON metadata file.
- The same metadata file can contain canned queries to the AI to use.
Each canned query will be turned into a separate MCP tool
sqlite_execute_main_{tool name}. - AI agents can execute arbitrary SQL queries with
sqlite_execute.
Quickstart using Visual Studio Code
-
Install uv.
-
Install Visual Studio Code if you don't already have it. Turn on GitHub Copilot.
-
Open this repo in VS Code. Open a GitHub Copilot agent mode chat. Check the available tools - you should see MCP Server: sqlite_sample with three available tools.

-
You should be able to ask Copilot in agent mode a question like "Get Titanic survivors of age 28" and get a response.

-
Use the sample MCP configuration file mcp.json and the sample metadata file titanic.yml as a starting point for your own configuration.
Interactive exploration with MCP Inspector and Datasette
The same database and metadata files can be used to explore the data interactively with MCP Inspector and Datasette.
| MCP Inspector | Datasette |
|---|---|
![]() | ![]() |
![]() | ![]() |
MCP Inspector
Use the MCP Inspector dashboard to interact with the SQLite database the same way that an AI agent would:
- Install npm.
- Run:
npx @modelcontextprotocol/inspector uvx mcp-sqlite sample/titanic.db --metadata sample/titanic.yml
Datasette
Since mcp-sqlite metadata is compatible with the Datasette metadata file, you can also explore your data with Datasette:
uvx datasette serve sample/titanic.db --metadata sample/titanic.yml
Compatibility with Datasette allows both AI agents and humans to easily explore the same local data!
MCP Tools provided by mcp-sqlite
- sqlite_get_catalog(): Tool the agent can call to get the complete catalog of the databases, tables, and columns in the data, combined with metadata from the metadata file.
In an earlier iteration of
mcp-sqlite, this was a resource instead of a tool, but resources are not as widely supported, so it got turned into a tool. If you have a usecase for the catalog as a resource, open an issue and we'll bring it back! - sqlite_execute(sql): Tool the agent can call to execute arbitrary SQL. The table results are returned as HTML. For more information about why HTML is the best format for LLMs to process, see Siu et al.
- {canned query name}({canned query args}): A tool is created for each canned query in the metadata, allowing the agent to run predefined queries without writing any SQL.
Usage
Command-line options
usage: mcp-sqlite [-h] [-m METADATA] [-p PREFIX] [-v] sqlite_file
CLI command to start an MCP server for interacting with SQLite data.
positional arguments:
sqlite_file Path to SQLite file to serve the MCP server for.
options:
-h, --help show this help message and exit
-m, --metadata METADATA
Path to Datasette-compatible metadata YAML or JSON file.
-p, --prefix PREFIX Prefix for MCP tools. Defaults to no prefix.
-v, --verbose Be verbose. Include once for INFO output, twice for DEBUG output.
Metadata
Hidden tables
Hiding a table with hidden: true will hide it from the catalog returned by the MCP tool sqlite_get_catalog().
However, note that the table will still be accessible by the AI agent!
Never rely on hiding a table from the catalog as a security feature.
Canned queries
Canned queries are each turned into a separate callable MCP tool by mcp-sqlite.
For example, a query named my_canned_query will become a tool my_canned_query.
The canned queries functionality is still in active development with more features planned for development soon:
Roadmap
| Datasette query feature | Supported in mcp-sqlite? |
|---|---|
| Displayed in catalog | ✅ |
| Executable | ✅ |
| Titles | ✅ |
| Descriptions | ✅ |
| Parameters | ✅ |
| Explicit parameters | ❌ (planned) |
| Hide SQL | ✅ |
| Write restrictions on canned queries | ✅ |
| Pagination | ❌ (planned) |
| Cross-database queries | ❌ (planned) |
| Fragments | ❌ (not planned) |
| Magic parameters | ❌ (not planned) |
Máy chủ liên quan
OceanBase MCP Server
Interact with OceanBase databases, allowing AI assistants to list tables, read data, and execute SQL queries securely.
Global Database
Access comprehensive company data including financial records, ownership structures, and contact information. Search for businesses using domains, registration numbers, or LinkedIn profiles to streamline due diligence and lead generation. Retrieve historical financial performance and complex corporate group structures to support informed business analysis.
CData Cloudant MCP Server
A read-only MCP server by CData for querying live Cloudant data with LLMs. Requires the CData JDBC Driver for Cloudant.
Prometheus MCP Server
A Prometheus MCP server with full API support for comprehensive management and deep interaction with Prometheus beyond basic query support. Written in go, it is a single binary install that is capable of STDIO, SSE, and HTTP transports for complex deployments.
MemFlow MCP
Enables Large Language Models to store and retrieve persistent memories with intelligent search capabilities.
Cryptocurrency Daemon
An MCP server for interacting with cryptocurrency daemon RPC interfaces.
aml
AML Watcher MCP Server gives AI agents direct, real-time access to AML Watcher's proprietary sanctions, watchlist, and PEP databases. Connect your agentic workflows to 215+ sanctions regimes, 2.6M+ PEP profiles, and coverage across 235+ countries, no complex integrations required.
MotherDuck
Query and analyze data with MotherDuck and local DuckDB
MCP Toolbox for Databases
Open source MCP server specializing in easy, fast, and secure tools for Databases.
mcp-parseable-server
MCP server for Parseable observability platform



