EdgeOne Pages MCP
An MCP server and client implementation for EdgeOne Pages Functions, supporting OpenAI-formatted requests.
EdgeOne Pages: MCP Client and Server Implementation with Functions
Project Overview
This project showcases an intelligent chat application built with EdgeOne Pages Functions technology. It interacts with backend functions through a web interface, implementing a complete Model Context Protocol (MCP) workflow.
The system architecture consists of the following core components:
- MCP Streamable HTTP Server (
functions/mcp-server/index.ts) - MCP Client (
functions/mcp-client/index.ts) - Backend API (
functions/v1/chat/completions/index.ts) as the MCP HOST, responsible for coordinating the entire MCP workflow
Through this architecture, users can access powerful MCP tool capabilities in the browser, enabling intelligent interactions such as "generating online webpages with a single prompt."
Core Features
- Interactive Chat Interface: Modern, responsive web interface built with Next.js and React
- High-Performance Edge Functions: Critical business logic deployed on highly scalable EdgeOne Pages Functions
- Complete MCP Implementation: Model Context Protocol implementation based on the latest specifications, providing powerful context management and request routing capabilities
- OpenAI Format Compatible: Backend API fully supports OpenAI-formatted request and response handling
Streamable HTTP MCP Server
Configure remote MCP services in applications that support Streamable HTTP MCP Server.
{
"mcpServers": {
"edgeone-pages-mcp-server": {
"url": "https://mcp-on-edge.edgeone.site/mcp-server"
}
}
}
Deploy
More Templates: EdgeOne Pages
Local Development
Environment Setup
First, install dependencies and start the development server:
# Install dependencies
npm install
# Or use other package managers
# yarn install / pnpm install / bun install
# Start development server
npm run dev
# Or use other package managers
# yarn dev / pnpm dev / bun dev
Configure environment variables: Copy the .env.example file and rename it to .env, then fill in your AI service interface configuration information.
After starting, visit http://localhost:3000 in your browser to view the application.
Project Structure
- Frontend Interface:
app/page.tsxcontains the main page logic and UI components - Backend Functions: All edge functions are in the
functionsdirectory- Chat API:
functions/v1/chat/completions - MCP Server:
functions/mcp-server - MCP Client:
functions/mcp-client
- Chat API:
Technical Documentation
Learn more about related technologies:
- Next.js Documentation - Next.js framework features and API
- EdgeOne Pages Functions Documentation - Detailed explanation of EdgeOne serverless functions
- Model Context Protocol (MCP) - Implemented based on the 2025-03-26 version of Streamable HTTP transport
相關伺服器
Scout Monitoring MCP
贊助Put performance and error data directly in the hands of your AI assistant.
Alpha Vantage MCP Server
贊助Access financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Interactive Feedback MCP
An MCP server for AI-assisted development tools like Cursor and Claude, supporting interactive feedback workflows with AI.
Tmux MCP Server
Provides persistent shell execution through tmux sessions.
Mobile Xray MCP
Take screenshots and analyze mobile apps with AI assistance directly from your IDE.
Tableau MCP
A suite of tools for developers to build AI applications that integrate with Tableau.
SensorMCP Server
Automate dataset creation and train custom object detection models using natural language.
Figma Copilot
Enables AI assistants to interact with and automate Figma designs programmatically.
Osquery MCP Server
An MCP server for Osquery that allows AI assistants to answer system diagnostic questions using natural language.
Replicate Minimax Image-01
Generate images using the minimax/image-01 model on Replicate.
Test Code Generator
Generates Vitest test code from JSON specifications using boundary value analysis and equivalence partitioning.
MCPShell
A secure bridge for LLMs to safely execute command-line tools via the Model Context Protocol (MCP).