A read-only MCP server to query live Adobe Analytics data. Requires the CData JDBC Driver for Adobe Analytics.
CData's Model Context Protocol (MCP) Server for Adobe Analytics
:heavy_exclamation_mark: This project builds a read-only MCP server. For full read, write, update, delete, and action capabilities and a simplified setup, check out our free CData MCP Server for Adobe Analytics (beta).
We created this read-only MCP Server to allow LLMs (like Claude Desktop) to query live data Adobe Analytics supported by the CData JDBC Driver for Adobe Analytics.
CData JDBC Driver connects to Adobe Analytics by exposing them as relational SQL models.
This server wraps that driver and makes Adobe Analytics data available through a simple MCP interface, so LLMs can retrieve live information by asking natural language questions — no SQL required.
git clone https://github.com/cdatasoftware/adobe-analytics-mcp-server-by-cdata.git
cd adobe-analytics-mcp-server-by-cdata
mvn clean install
This creates the JAR file: CDataMCP-jar-with-dependencies.jarlib
folder in the installation directory, typically:
C:\Program Files\CData\CData JDBC Driver for Adobe Analytics\
/Applications/CData JDBC Driver for Adobe Analytics/
java -jar cdata.jdbc.adobeanalytics.jar --license
Run the command java -jar cdata.jdbc.adobeanalytics.jar
to open the Connection String utility.
Configure the connection string and click "Test Connection"
Note: If the data sources uses OAuth, you will need to authenticate in your browser.
Once successful, copy the connection string for use later.
.prp
file for your JDBC connection (e.g. adobe-analytics.prp
) using the following properties and format:
Prefix=adobeanalytics
ServerName=CDataAdobeAnalytics
ServerVersion=1.0
DriverPath=PATH\TO\cdata.jdbc.adobeanalytics.jar
DriverClass=cdata.jdbc.adobeanalytics.AdobeAnalyticsDriver
JdbcUrl=jdbc:adobeanalytics:InitiateOAuth=GETANDREFRESH;
Tables=
Create the config file for Claude Desktop ( claude_desktop_config.json) to add the new MCP server, using the format below. If the file already exists, add the entry to the mcpServers
in the config file.
Windows
{
"mcpServers": {
"{classname_dash}": {
"command": "PATH\\TO\\java.exe",
"args": [
"-jar",
"PATH\\TO\\CDataMCP-jar-with-dependencies.jar",
"PATH\\TO\\adobe-analytics.prp"
]
},
...
}
}
Linux/Mac
{
"mcpServers": {
"{classname_dash}": {
"command": "/PATH/TO/java",
"args": [
"-jar",
"/PATH/TO/CDataMCP-jar-with-dependencies.jar",
"/PATH/TO/adobe-analytics.prp"
]
},
...
}
}
If needed, copy the config file to the appropriate directory (Claude Desktop as the example). Windows
cp C:\PATH\TO\claude_desktop_config.json %APPDATA%\Claude\claude_desktop_config.json
Linux/Mac
cp /PATH/TO/claude_desktop_config.json /Users/{user}/Library/Application\ Support/Claude/claude_desktop_config.json'
Run or refresh your client (Claude Desktop).
Note: You may need to fully exit or quit your Claude Desktop client and re-open it for the MCP Servers to appear.
java -jar /PATH/TO/CDataMCP-jar-with-dependencies.jar /PATH/TO/Salesforce.prp
Note: The server uses
stdio
so can only be used with clients that run on the same machine as the server.
Once the MCP Server is configured, the AI client will be able to use the built-in tools to read, write, update, and delete the underlying data. In general, you do not need to call the tools explicitly. Simply ask the client to answer questions about the underlying data system. For example:
The list of tools available and their descriptions follow:
In the definitions below, {servername}
refers to the name of the MCP Server in the config file (e.g. {classname_dash}
above).
{servername}_get_tables
- Retrieves a list of tables available in the data source. Use the {servername}_get_columns
tool to list available columns on a table. The output of the tool will be returned in CSV format, with the first line containing column headers.{servername}_get_columns
- Retrieves a list of columns for a table. Use the {servername}_get_tables
tool to get a list of available tables. The output of the tool will be returned in CSV format, with the first line containing column headers.{servername}_run_query
- Execute a SQL SELECT queryIf you are scripting out the requests sent to the MCP Server instead of using an AI Client (e.g. Claude), then you can refer to the JSON payload examples below – following the JSON-RPC 2.0 specification - when calling the available tools.
{
"jsonrpc": "2.0",
"id": 1,
"method": "tools/call",
"params": {
"name": "adobe_analytics_get_tables",
"arguments": {}
}
}
{
"jsonrpc": "2.0",
"id": 2,
"method": "tools/call",
"params": {
"name": "adobe_analytics_get_columns",
"arguments": {
"table": "Account"
}
}
}
{
"jsonrpc": "2.0",
"id": 3,
"method": "tools/call",
"params": {
"name": "adobe_analytics_run_query",
"arguments": {
"sql": "SELECT * FROM [Account] WHERE [IsDeleted] = true"
}
}
}
This MCP server is licensed under the MIT License. This means you are free to use, modify, and distribute the software, subject to the terms and conditions of the MIT License. For more details, please see the LICENSE file in the project repository.
Connect to any relational database, and be able to get valid SQL, and ask questions like what does a certain column prefix mean.
a lightweight, local RAG memory store to record, retrieve, update, delete, and visualize persistent "memories" across sessions—perfect for developers working with multiple AI coders (like Windsurf, Cursor, or Copilot) or anyone who wants their AI to actually remember them.
Embeddings, vector search, document storage, and full-text search with the open-source AI application database
Interact with the data stored in Couchbase clusters using natural language.
Real-time PostgreSQL & Supabase database schema access for AI-IDEs via Model Context Protocol. Provides live database context through secure SSE connections with three powerful tools: get_schema, analyze_database, and check_schema_alignment.
Interact with the Neon serverless Postgres platform
Search, Query and interact with data in your Milvus Vector Database.
Token Metrics integration for fetching real-time crypto market data, trading signals, price predictions, and advanced analytics.
BigQuery database integration with schema inspection and query capabilities
Manage and query databases, tenants, users, auth using LLMs