Kaggle
Interact with the Kaggle API to access datasets, notebooks, and competitions.
Kaggle MCP (Model Context Protocol) Server
This repository contains an MCP (Model Context Protocol) server (server.py) built using the fastmcp library. It interacts with the Kaggle API to provide tools for searching and downloading datasets, and a prompt for generating EDA notebooks.
Project Structure
server.py: The FastMCP server application. It defines resources, tools, and prompts for interacting with Kaggle..env.example: An example file for environment variables (Kaggle API credentials). Rename to.envand fill in your details.requirements.txt: Lists the necessary Python packages.pyproject.toml&uv.lock: Project metadata and locked dependencies foruvpackage manager.datasets/: Default directory where downloaded Kaggle datasets will be stored.
Setup
-
Clone the repository:
git clone <repository-url> cd <repository-directory> -
Create a virtual environment (recommended):
python -m venv venv source venv/bin/activate # On Windows use `venv\Scripts\activate` # Or use uv: uv venv -
Install dependencies: Using pip:
pip install -r requirements.txtOr using uv:
uv sync -
Set up Kaggle API credentials:
- Method 1 (Recommended): Environment Variables
- Create
.envfile - Open the
.envfile and add your Kaggle username and API key:KAGGLE_USERNAME=your_kaggle_username KAGGLE_KEY=your_kaggle_api_key - You can obtain your API key from your Kaggle account page (
Account>API>Create New API Token). This will download akaggle.jsonfile containing your username and key.
- Create
- Method 2:
kaggle.jsonfile- Download your
kaggle.jsonfile from your Kaggle account. - Place the
kaggle.jsonfile in the expected location (usually~/.kaggle/kaggle.jsonon Linux/macOS orC:\Users\<Your User Name>\.kaggle\kaggle.jsonon Windows). Thekagglelibrary will automatically detect this file if the environment variables are not set.
- Download your
- Method 1 (Recommended): Environment Variables
Running the Server
- Ensure your virtual environment is active.
- Run the MCP server:
The server will start and register its resources, tools, and prompts. You can interact with it using an MCP client or compatible tools.uv run kaggle-mcp
Running the Docker Container
1. Set up Kaggle API credentials
This project requires Kaggle API credentials to access Kaggle datasets.
- Go to https://www.kaggle.com/settings and click "Create New API Token" to download your
kaggle.jsonfile. - Open the
kaggle.jsonfile and copy your username and key into a new.envfile in the project root:
KAGGLE_USERNAME=your_username
KAGGLE_KEY=your_key
2. Build the Docker image
docker build -t kaggle-mcp-test .
3. Run the Docker container using your .env file
docker run --rm -it --env-file .env kaggle-mcp-test
This will automatically load your Kaggle credentials as environment variables inside the container.
Server Features
The server exposes the following capabilities through the Model Context Protocol:
Tools
search_kaggle_datasets(query: str):- Searches for datasets on Kaggle matching the provided query string.
- Returns a JSON list of the top 10 matching datasets with details like reference, title, download count, and last updated date.
download_kaggle_dataset(dataset_ref: str, download_path: str | None = None):- Downloads and unzips files for a specific Kaggle dataset.
dataset_ref: The dataset identifier in the formatusername/dataset-slug(e.g.,kaggle/titanic).download_path(Optional): Specifies where to download the dataset. If omitted, it defaults to./datasets/<dataset_slug>/relative to the server script's location.
Prompts
generate_eda_notebook(dataset_ref: str):- Generates a prompt message suitable for an AI model (like Gemini) to create a basic Exploratory Data Analysis (EDA) notebook for the specified Kaggle dataset reference.
- The prompt asks for Python code covering data loading, missing value checks, visualizations, and basic statistics.
Connecting to Claude Desktop
Go to Claude > Settings > Developer > Edit Config > claude_desktop_config.json to include the following:
{
"mcpServers": {
"kaggle-mcp": {
"command": "kaggle-mcp",
"cwd": "<path-to-their-cloned-repo>/kaggle-mcp"
}
}
}
Usage Example
An AI agent or MCP client could interact with this server like this:
- Agent: "Search Kaggle for datasets about 'heart disease'"
- Server executes
search_kaggle_datasets(query='heart disease')
- Server executes
- Agent: "Download the dataset 'user/heart-disease-dataset'"
- Server executes
download_kaggle_dataset(dataset_ref='user/heart-disease-dataset')
- Server executes
- Agent: "Generate an EDA notebook prompt for 'user/heart-disease-dataset'"
- Server executes
generate_eda_notebook(dataset_ref='user/heart-disease-dataset') - Server returns a structured prompt message.
- Server executes
- Agent: (Sends the prompt to a code-generating model) -> Receives EDA Python code.
Servidores relacionados
Vertex AI MCP Server
Interact with Google Cloud's Vertex AI Gemini models for coding assistance and general query answering.
AWS Documentation MCP Server
Access, search, and get recommendations from public AWS documentation.
Ntropy MCP server
Enrich banking data using the Ntropy API.
CORTEX MCP Extensions
Provides direct access to CORTEX infrastructure for orchestration, monitoring, AI flow management, model routing, and Docker administration.
LuckyCola MCP
An open MCP service providing a standardized interface compatible with various MCP clients.
Elastic Email MCP
The Elastic Email MCP Server enables AI agents like GitHub Copilot, ChatGPT, Claude, and other compatible assistants to seamlessly integrate with your Elastic Email account.
Cisco NSO MCP Server
An MCP server for Cisco NSO (Network Services Orchestrator) that exposes NSO data and operations as MCP primitives.
Remote MCP Server on Cloudflare
A remote MCP server designed to run on Cloudflare Workers, featuring OAuth login support.
Mengram
Human-like memory layer for AI agents with semantic, episodic, and procedural memory types, cognitive profiling, knowledge graph, and 12 MCP tools.
Salesforce TypeScript Connector
Interact with Salesforce data using SOQL queries, SOSL searches, and CRUD operations via a TypeScript MCP server.