LinkedIn Profile Scraper
Fetches LinkedIn profile information using the Fresh LinkedIn Profile Data API.
LinkedIn Profile Scraper MCP Server
This MCP server uses the Fresh LinkedIn Profile Data API to fetch LinkedIn profile information. It is implemented as a model context protocol (MCP) server and exposes a single tool, get_profile, which accepts a LinkedIn profile URL and returns the profile data in JSON format.
Features
- Fetch Profile Data: Retrieves LinkedIn profile information including skills and other settings (with most additional details disabled).
- Asynchronous HTTP Requests: Uses
httpxfor non-blocking API calls. - Environment-based Configuration: Reads the
RAPIDAPI_KEYfrom your environment variables usingdotenv.
Prerequisites
- Python 3.7+ – Ensure you are using Python version 3.7 or higher.
- MCP Framework: Make sure the MCP framework is installed.
- Required Libraries: Install
httpx,python-dotenv, and other dependencies. - RAPIDAPI_KEY: Obtain an API key from RapidAPI and add it to a
.envfile in your project directory (or set it in your environment).
Installation
-
Clone the Repository:
git clone https://github.com/AIAnytime/Awesome-MCP-Server cd linkedin_profile_scraper -
Install Dependencies:
uv add mcp[cli] httpx requests -
Set Up Environment Variables:
Create a
.envfile in the project directory with the following content:RAPIDAPI_KEY=your_rapidapi_key_here
Running the Server
To run the MCP server, execute:
uv run linkedin.py
The server will start and listen for incoming requests via standard I/O.
MCP Client Configuration
To connect your MCP client to this server, add the following configuration to your config.json. Adjust the paths as necessary for your environment:
{
"mcpServers": {
"linkedin_profile_scraper": {
"command": "C:/Users/aiany/.local/bin/uv",
"args": [
"--directory",
"C:/Users/aiany/OneDrive/Desktop/YT Video/linkedin-mcp/project",
"run",
"linkedin.py"
]
}
}
}
Code Overview
- Environment Setup: The server uses
dotenvto load theRAPIDAPI_KEYrequired to authenticate with the Fresh LinkedIn Profile Data API. - API Call: The asynchronous function
get_linkedin_datamakes a GET request to the API with specified query parameters. - MCP Tool: The
get_profiletool wraps the API call and returns formatted JSON data, or an error message if the call fails. - Server Execution: The MCP server is run with the
stdiotransport.
Troubleshooting
- Missing RAPIDAPI_KEY: If the key is not set, the server will raise a
ValueError. Make sure the key is added to your.envfile or set in your environment. - API Errors: If the API request fails, the tool will return a message indicating that the profile data could not be fetched.
License
This project is licensed under the MIT License. See the LICENSE file for more details.
相關伺服器
Bright Data
贊助Discover, extract, and interact with the web - one interface powering automated access across the public internet.
youtube-summarize
MCP server that fetches YouTube video transcripts and summarizes them using your LLM client
yt-dlp
Download video and audio from YouTube and other platforms using the yt-dlp tool.
Trends Hub
Aggregates trending topics from over 20 sources in real-time, with customizable fields and RSS feed support.
Primp MCP Server
An MCP server for the Primp HTTP client, enabling browser impersonation for requests and file uploads.
transcriptor-mcp
An MCP server (stdio + HTTP/SSE) that fetches video transcripts/subtitles via yt-dlp, with pagination for large responses. Supports YouTube, Twitter/X, Instagram, TikTok, Twitch, Vimeo, Facebook, Bilibili, VK, Dailymotion. Whisper fallback — transcribes audio when subtitles are unavailable (local or OpenAI API). Works with Cursor and other MCP host
Scrapling Fetch MCP
Fetches HTML and markdown from websites with anti-automation measures using Scrapling.
Nefino
Access the Nefino renewable energy news API.
Secure Fetch
Secure fetch to prevent access to local resources
Chrome MCP Server
Control a Chrome browser instance using the Chrome DevTools Protocol (CDP).
Crawl4AI RAG
Integrate web crawling and Retrieval-Augmented Generation (RAG) into AI agents and coding assistants.