Open-Brain

MCP memory server that builds a knowledge graph as you capture thoughts. 16 tools. Self-hostable.

Open Brain

Your personal knowledge infrastructure. Capture thoughts from anywhere, search by meaning.

Open Brain stores your thoughts with AI-generated embeddings so any AI assistant can search your memory by meaning -- not keywords. Thoughts flow in from Telegram, automated pipelines, or any MCP-compatible AI client (Claude, ChatGPT, and others) and are instantly searchable. A knowledge graph auto-links related thoughts, extracts entities, and tracks co-occurrence patterns. Automated maintenance keeps the graph healthy -- deduplicating near-duplicates, tracking theme evolution, synthesizing insights, and archiving stale content.

Everything lives in your own database, so you own your data.

How It Works

flowchart LR
    TG[Telegram Message] --> TGBot[telegram-bot\nEdge Function]
    MCP[AI Client\nClaude/ChatGPT] --> MCPServer[open-brain-mcp\nEdge Function]
    Pipeline[RSS/HF Papers/\nEmergent Mind] --> RunPipeline[run-pipeline\nEdge Function]
    TGBot --> OR1[OpenRouter\nEmbedding + Metadata]
    MCPServer --> OR2[OpenRouter\nEmbedding + Search]
    RunPipeline --> OR3[OpenRouter\nTriage + Embed]
    OR1 --> DB[(Postgres\n+ pgvector)]
    OR2 --> DB
    OR3 --> DB
    MCPServer --> DB
    TGBot --> TGReply[Telegram Reply\nwith Metadata]

Capture Path

When you send a message to the Telegram bot, the telegram-bot Edge Function picks it up via webhook. It sends the message to OpenRouter in parallel for two things: generating a vector embedding (a numerical representation of meaning) and extracting metadata like topics, people mentioned, action items, theme, quality score, and named entities. The thought is checked for semantic duplicates, stored in your database with auto-linked connections to related thoughts, and the bot replies with a summary of what it captured.

Pipeline Path

The run-pipeline Edge Function automatically ingests ideas from RSS feeds (AI newsletters), Hugging Face daily papers, and Emergent Mind (trending arXiv papers). Each item is triaged for relevance, embedded, deduplicated, and stored. Runs on a schedule via GitHub Actions (Supabase deployment) or a built-in cron container (Docker deployment).

Retrieval Path

Any AI client connected via MCP (Model Context Protocol) can search your thoughts by meaning using semantic search, browse by filters (type, topic, person, time), get aggregate statistics, or request a weekly review of themes. The open-brain-mcp Edge Function handles these requests, authenticated with your personal access key.

Knowledge Graph

Every thought is automatically linked to related thoughts via vector similarity. Connections above 0.80 similarity are classified by an LLM into typed relationships (extends, contradicts, is-evidence-for, supersedes). Named entities (people, tools, projects, organizations) are extracted and resolved into a shared entity graph. Co-occurrence edges track which thoughts are retrieved together over time, strengthening connections based on actual usage patterns.

Storage

Everything lives in Postgres with pgvector for fast similarity search. Thoughts are stored with their embeddings (1536-dimensional vectors), metadata, typed connections, and entity references. You can deploy on Supabase (managed hosting) or self-host with Docker Compose.

Deployment Options

Choose how you want to run Open Brain:

Supabase (hosted)Docker Compose (self-hosted)
SetupLink project + run scriptscp .env.example .env + ./start.sh
InfrastructureManaged by SupabaseRuns on your machine/server
SchedulingGitHub ActionsBuilt-in cron container
CostSupabase free tier + OpenRouterJust OpenRouter
GuideContinue belowDocker guide

Supabase Deployment

Prerequisites

  1. Supabase account -- Supabase is a hosted Postgres database with built-in APIs, authentication, and Edge Functions (serverless code). Create a free account at supabase.com. Create a new project -- you will need the project URL (looks like https://abcdef.supabase.co) and the service role key (a long string found under Settings > API).

  2. Supabase CLI -- The command-line tool for managing your Supabase project (applying database migrations, deploying functions, setting secrets).

    npm install -g supabase
    
  3. OpenRouter account -- OpenRouter routes requests to AI models. It is used here for generating embeddings (vector representations of your thoughts) and extracting metadata. Create an account at openrouter.ai and generate an API key from the dashboard.

  4. Telegram bot (recommended) -- The primary way to capture thoughts on the go. Create a bot via @BotFather on Telegram and run the setup script (see below). If you only want MCP access, you can skip this.

Quick Start

1. Clone the repository

git clone https://github.com/YOUR_USERNAME/open_brain.git
cd open_brain

2. Link your Supabase project

cd supabase
supabase link --project-ref YOUR_PROJECT_REF
cd ..

Tip: Your project ref is the subdomain in your Supabase URL. If your URL is https://abcdef.supabase.co, your project ref is abcdef.

3. Run bootstrap

./scripts/bootstrap.sh

Bootstrap walks you through setting up your environment. It prompts for each secret (Supabase URL, service role key, OpenRouter API key, Telegram tokens, etc.), generates a cryptographic MCP access key automatically, and writes everything to .env.local. If you already have a .env.local, it will show your existing values and let you update specific ones.

4. Run deploy

./scripts/deploy.sh

Deploy applies the database schema (creates the thoughts table with vector search indexes), uploads your secrets to Supabase, and deploys all Edge Functions. It shows a step-by-step checklist as each operation completes. At the end, it prints your MCP connection URL and a ready-to-paste Claude Code command.

5. Run validate

./scripts/validate.sh

Validate runs 8 checks against your live deployment to confirm everything works: database access, RPC functions, Edge Function reachability, authentication, thought capture, semantic search, and thought listing. It prints a checklist with pass/fail for each check and a final summary.

Set up Telegram bot (optional)

Create a bot via @BotFather on Telegram, then run the setup script:

./scripts/setup-telegram.sh YOUR_BOT_TOKEN

The script verifies your token, registers the webhook, sets up command autocomplete, and prints the env vars and secrets to configure. Follow the printed instructions to complete setup.

Connect Your AI Client

Once deployed, connect your AI client to start using Open Brain. You need two values:

  • MCP endpoint URL: https://YOUR_REF.supabase.co/functions/v1/open-brain-mcp/mcp (Supabase) or http://localhost:80/functions/v1/open-brain-mcp (Docker)
  • MCP access key: The key generated during setup (stored in .env.local or Docker .env)

Tip: The deploy script (Supabase) or start script (Docker) prints the exact connection command with your values filled in.

Claude Code (CLI -- recommended)

claude mcp add --transport http --header "x-brain-key: YOUR_MCP_KEY" open-brain https://YOUR_REF.supabase.co/functions/v1/open-brain-mcp/mcp

This registers Open Brain as an MCP server that Claude Code can use in any conversation. Replace YOUR_MCP_KEY and YOUR_REF with your actual values.

Claude Code (project .mcp.json)

Add this to a .mcp.json file in your project root to share the connection with your team:

{
  "mcpServers": {
    "open-brain": {
      "type": "http",
      "url": "https://YOUR_REF.supabase.co/functions/v1/open-brain-mcp/mcp",
      "headers": {
        "x-brain-key": "${MCP_ACCESS_KEY}"
      }
    }
  }
}

Note: The ${MCP_ACCESS_KEY} syntax uses environment variable expansion so your key stays out of version control. Set the MCP_ACCESS_KEY environment variable on each machine that uses this config.

Claude Desktop

Claude Desktop does not support remote MCP servers via configuration files. Instead:

  1. Open Claude Desktop > Settings > Connectors
  2. Click Add a new connector
  3. Enter the MCP endpoint URL: https://YOUR_REF.supabase.co/functions/v1/open-brain-mcp/mcp
  4. Configure the x-brain-key authentication header with your MCP access key

ChatGPT (Pro/Team/Enterprise/Edu)

  1. Go to Settings > Connectors > Advanced > Developer Mode
  2. Add the MCP server URL: https://YOUR_REF.supabase.co/functions/v1/open-brain-mcp/mcp
  3. Configure the x-brain-key authentication header with your MCP access key

Usage Examples

Telegram Capture

Send any message to your bot and Open Brain automatically processes it:

You: Just had a great meeting with Sarah about the Q3 product roadmap.
     She wants to prioritize the mobile app redesign.

Bot: Captured!
  Type: meeting_note
  Theme: personal
  Topics: q3-roadmap, mobile-app-redesign
  Quality: 0.7
  People: Sarah
  Action items: Prioritize mobile app redesign
  Why: Records a product strategy decision with clear ownership
  Related: "Product planning session notes..." (82% similar)

Every message is embedded as a vector, enriched with extracted metadata, checked for duplicates, auto-linked to related thoughts, and entities are resolved into a knowledge graph.

Semantic Search

Ask any connected AI client to search your brain:

You: Search my brain for anything about product roadmap discussions

Claude: I found 3 relevant thoughts:
  1. (0.89 similarity) Meeting with Sarah about Q3 product roadmap...
  2. (0.82 similarity) Product planning session notes...
  3. (0.76 similarity) Quarterly goals discussion...

Semantic search finds thoughts by meaning -- even if you used different words. Asking about "product planning" will find thoughts about "roadmap discussions" because the meanings are similar.

Weekly Review

Get an AI-generated summary of your recent thinking:

You: Give me a weekly review of my recent thoughts

Claude: Here's your weekly review:
  Themes: Product planning, team meetings, technical architecture
  Open loops: Mobile redesign decision pending, API migration timeline
  Connections: Sarah mentioned in 3 meetings this week, all about mobile

The weekly review analyzes the past 7 days of thoughts and synthesizes themes, open loops, connections between ideas, and gaps in your thinking.

Available Tools

ToolDescription
search_thoughtsSemantic search with optional graph expansion (1-hop traversal)
list_thoughtsBrowse thoughts filtered by type, topic, person, theme, quality, time
thought_statsAggregate statistics: counts, type/theme breakdown, top topics/people
capture_thoughtSave a new thought from any AI client (with auto-embedding)
get_connectionsGraph traversal from a thought (typed links: extends, contradicts, etc.)
list_entitiesBrowse extracted entities (people, tools, projects, orgs) by frequency
weekly_reviewAI-generated summary of themes, open loops, and next steps
analyzeGraph analysis: hubs, density, sources, co-occurrence, themes
dedup_reviewDuplicate candidates with similarity zone histogram
refresh_salienceRecompute all salience scores
update_thoughtRewrite content (re-embeds, re-extracts metadata)
delete_thoughtPermanent delete (cascades connections)
serendipity_digestResurface forgotten high-quality thoughts
pipelinePipeline monitoring: health status, run history, merge audit
review_staleReview and act on stale thought candidates
migration_guideInstructions for importing memories from other platforms

See docs/cookbook.md for detailed usage patterns, tool compositions, and non-obvious behaviors.

Skills (Claude Code Workflows)

Open Brain includes Claude Code skills -- structured multi-phase workflows that compose the MCP tools above into higher-level analysis. Skills are auto-discovered from .claude/skills/ and invoked as slash commands.

SkillWhat it does
/discoverIncremental pattern discovery across recent thoughts. Builds on previous reports (EVOLVED/NEW/STALE classification), dispatches parallel research agents, correlates with project priorities.
/pulsePipeline and data health report. 9 parallel MCP calls, rubric-scored (GREEN/YELLOW/RED), cross-run memory for tracking findings over time, 6 cross-metric pattern detectors.
/brain-healthKnowledge graph health report. 12 parallel MCP calls covering theme attention, graph density, hub health, co-occurrence alignment, dedup pressure, synthesis output, and entity landscape.

See docs/skills/README.md for detailed descriptions and usage.

Automated Maintenance

Open Brain runs background maintenance to keep the knowledge graph healthy. These jobs run automatically -- via GitHub Actions (Supabase deployment) or the built-in cron container (Docker deployment).

JobFrequencyPurpose
RSS/HF Papers/Emergent Mind ingestion2x dailyIngest ideas from configured sources
Pipeline monitoring2x dailyHealth checks with Telegram alerting on failures
Dream dedup2x dailyMerge near-duplicate thoughts (>0.92 similarity auto-merged, 0.85-0.92 LLM-confirmed)
Graph analysis cacheDailyPre-compute hub, density, and co-occurrence analysis
Dream themesWeeklyTrack theme velocity, lifecycle transitions (emerging/active/declining), centroid drift
Dream decayWeeklyArchive stale thoughts via tiered scoring + LLM confirmation
Dream synthesisWeeklyGenerate cross-cutting insights from thought clusters
Co-occurrence decayWeeklyDecay unused co-occurrence edges

GitHub Actions workflow files are included in docs/workflows/ as reference for customizing schedules.

Benchmark

Open Brain includes a LongMemEval benchmark harness that measures retrieval quality across 500 questions in 6 categories. Each question gets its own isolated brain -- conversations are ingested as thoughts, then questions test whether the right memories can be retrieved and answered.

Baseline results (threshold=0.4, limit=20, graph expansion on):

CategoryScoreWhat it tests
single-session-assistant69.6%Recall assistant responses from a single conversation
single-session-user55.7%Recall user statements from a single conversation
knowledge-update52.6%Surface the latest version of updated information
temporal-reasoning30.8%Answer questions about when things happened
multi-session18.8%Synthesize across multiple conversations
single-session-preference3.3%Recall casually mentioned preferences
Overall37.2%

Vector search handles direct factual recall well. Multi-session synthesis and temporal reasoning are the main areas for improvement -- they require retrieval strategies beyond single-query vector similarity.

Run the benchmark against your own instance:

pip install -r benchmark/requirements.txt
cp benchmark/.env.example benchmark/.env  # fill in keys
python -m benchmark longmemeval run --concurrency 5

Results land in benchmark/results/. See benchmark/longmemeval/config.py for tunable parameters (threshold, retrieval limit, models).

Project Structure

open-brain-server/
  supabase/
    migrations/                   # Database migrations (applied with supabase db push)
    functions/
      _shared/                    # Shared modules (supabase-client, openrouter, types, errors, auto-link, entities, dream-*)
      telegram-bot/               # Telegram capture (primary capture path)
      open-brain-mcp/             # MCP server (16 tools)
        tools/                    # Individual tool implementations
      run-pipeline/               # Automated RSS/HF Papers/Emergent Mind ingestion
      monitor-pipeline/           # Pipeline health monitoring with Telegram alerts
      refresh-graph-analysis/     # Graph analysis cache computation
  benchmark/                        # LongMemEval retrieval quality benchmark
    longmemeval/                    # Harness: retrieve → reader → judge → score
    results/                        # Baseline results (summary markdown)
    tests/                          # 60 tests
  docker/                         # Docker Compose self-hosting (6 services)
  pipeline/                       # Python-based local pipeline (Reddit, RSS, briefing)
  scripts/                        # Setup and deployment automation
  tests/                          # Integration tests
  docs/
    cookbook.md                    # MCP tool usage patterns and compositions
    skills/                       # Skill documentation
    workflows/                    # GitHub Actions reference (scheduling)
    writing-a-source.md           # Guide for adding pipeline sources
  .claude/
    skills/                       # Claude Code skills (auto-discovered)
      discover/                   # Incremental pattern discovery
      pulse/                      # Pipeline health report
      brain-health/               # Knowledge graph health report

License

MIT

Related Servers

NotebookLM Web Importer

Import web pages and YouTube videos to NotebookLM with one click. Trusted by 200,000+ users.

Install Chrome Extension