JDBC-MCP
Enables AI assistants to interact with various databases through JDBC connections.
JDBC-MCP - Database Access for AI Assistants
This is a fork of the Quarkus MCP JDBC Server with added features:
- Docker deployment - Easily run the MCP SSE or MCP STDIO as a containerized application
- Enhanced security - Write queries (INSERT, UPDATE, DELETE) are disabled by default and need to be explicitly enabled
The Model Context Protocol (MCP) server enables AI assistants to interact with databases through JDBC connections, making it possible for LLMs to inspect, query, create, and modify database content.
New Features
Docker Deployment
You can now run the JDBC-MCP server in a Docker container.
MCP SSE
Simple usage
docker run -d --name jdbc-mcp \
-p 8080:8080 \
-e enable.write.sql=false \
-e jdbc.user=db_user \
-e jdbc.password=db_password \
-e jdbc.url=jdbc:postgresql://host:port/database \
guang1/jdbc-mcp:latest
Using an Environment File
Create a .env file:
enable.write.sql=false
jdbc.user=db_user
jdbc.password=db_password
jdbc.url=jdbc:postgresql://host:port/database
Then run:
docker run -d --name jdbc-mcp \
-p 8080:8080 \
--env-file .env \
guang1/jdbc-mcp:latest
Docker Compose for SSE
Basic Example
services:
jdbc-mcp:
container_name: jdbc-mcp
ports:
- 8080:8080
environment:
- enable.write.sql=false
- jdbc.user=db_user
- jdbc.password=db_password
- jdbc.url=jdbc:postgresql://host:port/database
restart: always
image: guang1/jdbc-mcp:latest
With Environment File
services:
jdbc-mcp:
container_name: jdbc-mcp
ports:
- 8080:8080
env_file:
- .env
restart: always
image: guang1/jdbc-mcp:latest
MCP SSE URL
http://localhost:8080/mcp/sse
MCP STDIO
Must pass quarkus.mcp.server.stdio.enabled=true, quarkus.log.console.enable=false and quarkus.log.console.stderr=false environment variables with -i interactive option.
Simple usage
docker run --rm \
-e enable.write.sql=false \
-e jdbc.user=db_user \
-e jdbc.password=db_password \
-e jdbc.url=jdbc:postgresql://host:port/database \
-e quarkus.mcp.server.stdio.enabled=true \
-e quarkus.log.console.enable=false \
-e quarkus.log.console.stderr=false \
guang1/jdbc-mcp:latest
Using an Environment File
docker run --rm
--env-file .env \
-e quarkus.mcp.server.stdio.enabled=true \
-e quarkus.log.console.enable=false \
-e quarkus.log.console.stderr=false \
guang1/jdbc-mcp:latest
Claude Desktop Configuration (STDIO)
To use the Docker version with Claude Desktop, add this to your claude_desktop_config.json or server_config.json file:
Simple usage
{
"mcpServers": {
"jdbc": {
"command": "docker",
"args": [
"run",
"--rm",
"-e",
"enable.write.sql=false",
"-e",
"jdbc.user=db_user",
"-e",
"jdbc.password=db_password",
"-e",
"jdbc.url=jdbc:postgresql://host:port/database",
"-e",
"quarkus.mcp.server.stdio.enabled=true",
"-e",
"quarkus.log.console.enable=false",
"-e",
"quarkus.log.console.stderr=false",
"-i",
"guang1/jdbc-mcp:latest"
]
}
}
}
Using an Environment File
{
"mcpServers": {
"jdbc": {
"command": "docker",
"args": [
"run",
"--rm",
"--env-file",
".env",
"-e",
"quarkus.mcp.server.stdio.enabled=true",
"-e",
"quarkus.log.console.enable=false",
"-e",
"quarkus.log.console.stderr=false",
"-i",
"guang1/jdbc-mcp:latest"
]
}
}
}
Configuration
Environment Variables
| Variable | Description | Default |
|---|---|---|
jdbc.url | JDBC connection URL (required) | - |
jdbc.user | Database username | - |
jdbc.password | Database password | - |
enable.write.sql | Enable SQL operations that modify data | false |
Supported Database Types
This MCP server supports multiple database types through JDBC. For detailed information on supported JDBC variants, please see the quarkus-mcp-servers/jdbc documentation.
Security Considerations
⚠️ Warning: When enable.write.sql is set to true, the MCP server can execute SQL statements that modify data. Use with caution in production environments.
Consider these best practices:
- Use read-only database users when possible
- Implement network-level access controls
- Run in an isolated network environment
Using with LLMs
This MCP server provides a standardized API that Large Language Models can use to interact with your database. To connect your LLM:
- Configure the LLM to use the MCP server endpoint
- Define the capabilities you want to allow (read-only vs. read-write)
- Ensure your database schema is accessible to the MCP server
Write Query Security
By default, write operations (INSERT, UPDATE, DELETE) are disabled for security. To enable them:
- When using Docker: add
-e enable.write.sql=trueas shown above
Original Documentation
For all other features, including supported databases, general usage, example databases, MCP components, and troubleshooting, please refer to the original repository documentation:
Quarkus MCP JDBC Server Documentation
License
This project is licensed under the Apache License 2.0.
Acknowledgments
- Original project: Quarkiverse MCP Servers
- Built with Quarkus
- Containerized with Docker
Verwandte Server
MCP Memory Server - Python Implementation
A Python implementation of the MCP memory server for knowledge graph storage and retrieval, using JSONL files for persistence.
Global Database
Access comprehensive company data including financial records, ownership structures, and contact information. Search for businesses using domains, registration numbers, or LinkedIn profiles to streamline due diligence and lead generation. Retrieve historical financial performance and complex corporate group structures to support informed business analysis.
ChromaDB
Provides AI assistants with persistent memory using ChromaDB vector storage.
Neon
Interact with the Neon serverless Postgres platform
Qdrant MCP
Semantic search using the Qdrant vector database.
Billy MCP Client
Access live U.S. congressional data from the Congress.gov API.
Fireproof
Immutable ledger database with live synchronization
Canada's Food Guide
A nutrition analysis platform integrating Canada's Food Guide recipes with Health Canada's official nutrition databases.
SEC EDGAR MCP Server
Provides AI assistants with access to the SEC EDGAR database via its API.
Snow Leopard BigQuery MCP
Interact with Google BigQuery databases using natural language queries and schema exploration.