YouTube Transcript Extractor
Extracts transcripts from public YouTube videos.
YouTube Transcript Extractor MCP 🎥
A Model Context Protocol (MCP) server that enables AI assistants to extract transcripts from YouTube videos. Built for integration with Cursor and Claude Desktop, this tool allows AI to analyze and work with YouTube video content directly.
Features
- 🎯 Extract transcripts from any public YouTube video
- 🔌 Easy integration with Cursor and Claude Desktop
- 🚀 Built with TypeScript for type safety
- 📦 Simple setup and deployment
- 🛠️ Based on the Model Context Protocol
Prerequisites
- Node.js (v16 or higher)
- pnpm (recommended) or npm
- A YouTube video URL to extract transcripts from
Installation
- Clone the repository:
git clone https://github.com/yourusername/yt-mcp.git
cd yt-mcp
- Install dependencies:
pnpm install
- Build the project:
pnpm run build
Configuration
For Cursor
- Open Cursor Settings
- Navigate to MCP → Add new MCP server
- Configure with these settings:
- Name:
youtube-transcript - Type:
command - Command:
node /absolute/path/to/yt-mcp/build/index.js
- Name:
For Claude Desktop
Add this configuration to your Claude Desktop config:
{
"mcpServers": {
"youtube-transcript": {
"command": "node",
"args": ["/absolute/path/to/yt-mcp/build/index.js"]
}
}
}
Usage
Once configured, the AI can extract transcripts from YouTube videos by calling the tool with a video URL. Example:
// The AI will use this format internally
const transcript = await extractTranscript({
input: "https://www.youtube.com/watch?v=VIDEO_ID"
});
Technical Details
The server is built using:
- @modelcontextprotocol/sdk - For MCP implementation
- youtube-transcript - For transcript extraction
- TypeScript - For type safety and better development experience
Limitations
- Only works with public YouTube videos
- Videos must have captions/subtitles enabled
- Some videos may have auto-generated captions which might not be 100% accurate
Troubleshooting
Common issues and solutions:
-
"Cannot find video ID" error
- Ensure the YouTube URL is complete and correct
- Check if the video is publicly accessible
-
"No transcript available" error
- Verify that the video has captions enabled
- Try a different video to confirm the tool is working
-
Build errors
- Make sure all dependencies are installed
- Check Node.js version (should be v16 or higher)
Contributing
Contributions are welcome! Please feel free to submit a Pull Request. For major changes, please open an issue first to discuss what you would like to change.
License
MIT
Servidores relacionados
Bright Data
patrocinadorDiscover, extract, and interact with the web - one interface powering automated access across the public internet.
Reddit MCP
A server to browse, search, and read content on Reddit using the Reddit API.
Browser Use
An AI-driven browser automation server for natural language control and web research, with CLI access.
Riksdag & Regering MCP
MCP-server that provides LLMs with easy access to open data from the Swedish Government Offices and Parliament.
Intelligent Crawl4AI Agent
An AI-powered web scraping system for high-volume automation and advanced data extraction strategies.
yt-dlp
Download video and audio content from various websites like YouTube, Facebook, and Tiktok using yt-dlp.
Github to MCP
Convert GitHub repositories to MCP servers automatically. Extract tools from OpenAPI, GraphQL & REST APIs for Claude Desktop, Cursor, Windsurf, Cline & VS Code. AI-powered code generation creates type-safe TypeScript/Python MCP servers. Zero config setup - just paste a repo URL. Built for AI assistants & LLM tool integration.
Crawl MCP
An MCP server for crawling WeChat articles. It supports single and batch crawling with multiple output formats, designed for AI tools like Cursor.
Crawl4AI RAG
Integrate web crawling and Retrieval-Augmented Generation (RAG) into AI agents and coding assistants.
Puppeteer MCP Server
Automate browser interactions using Puppeteer, controlling new or existing Chrome instances.
Crawl4AI MCP Server
An MCP server for advanced web crawling, content extraction, and AI-powered analysis using the crawl4ai library.