Svelte Documentation
Remote server (SSE/Streamable) for the latest Svelte and SvelteKit documentation
llmctx
llmctx transforms technical documentation into AI-ready formats. It provides a simple way to access condensed, LLM-friendly versions of popular framework and library documentation through preset URLs. This makes it easier for developers to use AI coding assistants like Cursor or Zed with accurate, up-to-date context from official documentation sources.
- Preset URLs for quick access to LLM-optimized documentation
- Regularly updated content from official documentation
Supported presets
Visit llmctx.com to see all presets.
Adding presets
To add a new preset:
-
Fork this repo.
-
Open the
src/lib/presets.tsfile. -
Add a new entry to the
presetsobject with the following structure:[presetKey]: { title: 'Preset Title', owner: 'github-owner', repo: 'github-repo', glob: ['**/*.md', '**/*.mdx', '!**/excluded/**'], // Required, supports glob patterns prompt: 'Optional prompt for additional context' } -
Create a pull request with your changes.
The glob field supports glob patterns, providing flexible file matching:
- Use
**/*.mdto match all Markdown files in any subdirectory - Use
**/*.mdxto match all MDX files in any subdirectory - Use
!patternto exclude files/directories matching the pattern - Combine patterns for fine-grained control, e.g.,
['**/*.md', '**/*.mdx', '!**/excluded/**'] - Patterns are processed in order, so you can include files and then exclude specific ones
Please ensure that the documentation source is reliable and actively maintained.
Feature requests
- Combine multiple documentation sources (e.g., https://llmctx.com/svelte,sveltekit)
- Create a "stack" of frequently used frameworks and libraries
- Implement a checkbox UI for selecting presets and generate custom URLs based on selected documentation
เซิร์ฟเวอร์ที่เกี่ยวข้อง
Alpha Vantage MCP Server
ผู้สนับสนุนAccess financial market data: realtime & historical stock, ETF, options, forex, crypto, commodities, fundamentals, technical indicators, & more
Agent VRM MCP Server
A server that provides VRM avatar functionality for Large Language Models (LLMs) by connecting to an AgentVRM engine.
Bonsai MCP
An MCP server that integrates IFC model support using Bonsai BIM (Blender) and IfcOpenShell.
MCP Hello World
A minimal MCP server mock in TypeScript for testing MCP clients, supporting both STDIO and HTTP/SSE protocols.
Local Code Indexing for Cursor
A Python-based server that locally indexes codebases using ChromaDB to provide semantic search for tools like Cursor.
TeamCity
MCP server for TeamCity, integrates with Claude Desktop and Cursor.
RenderLens
Visual verification MCP server — render code to screenshots, run WCAG accessibility audits, and pixel-diff UI changes. Free, no API key.
Reactive AI Agent Framework
A reactive AI agent framework for creating agents that use tools to perform tasks, with support for multiple LLM providers and MCP servers.
Data Structure Protocol (DSP)
Graph-based long-term memory skill for AI (LLM) coding agents — faster context, fewer tokens, safer refactors
mcp-of-mcps
MCP of MCPs is a meta-server that merges all your MCP servers into a single smart endpoint. It gives AI agents instant tool discovery, selective schema loading, and massively cheaper execution, so you stop wasting tokens and time. With persistent tool metadata, semantic search, and direct code execution between tools, it turns chaotic multi-server setups into a fast, efficient, hallucination-free workflow. It also automatically analyzes the tools output schemas if not exist and preserves them across sessions for consistent behavior.
Code Scalpel
Code Scalpel is an MCP server that upgrades your AI coding agent with surgical, graph-based tools for precise analysis, atomic refactoring, and 99% lower token costs.