An AI-powered server that generates PyAirbyte pipeline code and instructions using OpenAI and connector documentation.
The PyAirbyte Managed Code Provider (MCP) service is an AI-powered backend that generates PyAirbyte pipeline code and instructions. It leverages OpenAI and connector documentation to help users quickly scaffold and configure data pipelines between sources and destinations supported by Airbyte. The MCP service automates code generation, provides context-aware guidance, and streamlines the process of building and deploying data pipelines. If you want to learn more on how the service works check out this video.
The easiest way to get started is using our hosted MCP server. Add this to your Cursor MCP configuration file (.cursor/mcp.json
):
{
"mcpServers": {
"pyairbyte-mcp": {
"url": "https://pyairbyte-mcp-7b7b8566f2ce.herokuapp.com/mcp",
"env": {
"OPENAI_API_KEY": "your-openai-api-key-here"
}
}
}
}
Requirements:
Configuration Steps:
.cursor/mcp.json
in your project directory (for project-specific) or ~/.cursor/mcp.json
(for global access)Once configured, you can use the MCP server in your AI assistant by asking it to generate PyAirbyte pipelines.
Simply ask Cline to generate a PyAirbyte pipeline! Here are example prompts:
Basic Examples:
Generate a PyAirbyte pipeline from source-postgres to destination-snowflake
Create a pipeline to move data from source-github to dataframe
Build a PyAirbyte script for source-stripe to destination-bigquery
Generate a data pipeline from source-salesforce to destination-postgres
Create a pipeline that reads from source-github to a dataframe, and then visualize the results using Streamlit
Help me set up a data pipeline from source-salesforce to destination-postgres
source-postgres
, source-github
, source-stripe
, source-mysql
, source-salesforce
)destination-snowflake
, destination-bigquery
, destination-postgres
) OR dataframe
for Pandas analysisThe tool will automatically use your OpenAI API key (configured in the MCP settings) to generate enhanced, well-documented pipeline code with best practices and detailed setup instructions!
Just start by asking Cline to generate a pipeline for your specific use case! 🎯
Generates a complete PyAirbyte pipeline with setup instructions.
Parameters:
source_name
: The official Airbyte source connector name (e.g., 'source-postgres', 'source-github')destination_name
: The official Airbyte destination connector name (e.g., 'destination-postgres', 'destination-snowflake') OR 'dataframe' to output to Pandas DataFramesReturns:
Securely access BigQuery datasets with intelligent caching, schema tracking, and query analytics via Supabase integration.
Query the Materials Project database using the mp_api client. Requires an MP_API_KEY environment variable.
A server that enables LLMs to connect and interact with databases via JDBC, built using the Spring AI MCP framework.
Interact with Tinybird serverless ClickHouse platform
A Model Context Protocol Server for MongoDB
Connects to Supabase platform for database, auth, edge functions and more.
Query your ClickHouse database server.
MCP Server For Apache Doris, an MPP-based real-time data warehouse.
Access financial statements, including income, balance sheets, and cash flow, for companies listed on the Securities Exchange of Thailand (SET).
A desktop application for managing and interacting with the MCP Memory Service, a semantic memory system built on the Model Context Protocol.