Thoughtbox
next-generation MCP reasoning tool. successor to Waldzell AI's Clear Thought.
Thoughtbox
Multi-agent collaborative reasoning that's auditable. Thoughtbox is a Docker-based MCP server where AI agents coordinate through shared workspaces — claiming problems, proposing solutions, reviewing each other's work, and reaching consensus. Every step is recorded as a structured thought in a persistent reasoning ledger that can be visualized, exported, and analyzed.
Local-First: Runs entirely on your machine. All data stays at ~/.thoughtbox/ — nothing leaves your network.
Observatory UI showing a reasoning session with 14 thoughts and a branch exploration (purple nodes 13-14) forking from thought 5.
Multi-Agent Collaboration
The Hub is the coordination layer. Agents register with role-specific profiles, join shared workspaces, and work through a structured problem-solving workflow — all via the thoughtbox_hub MCP tool.
The workflow: register → create workspace → create problem → claim → work → propose solution → peer review → merge → consensus
Workspace primitives:
- Problem — A unit of work with dependencies, sub-problems, and status tracking (open → in-progress → resolved → closed)
- Proposal — A proposed solution with a source branch reference and review workflow
- Consensus — A decision marker tied to a thought reference for traceability
- Channel — A message stream scoped to a problem for discussion
Agent Profiles: MANAGER, ARCHITECT, DEBUGGER, SECURITY, RESEARCHER, REVIEWER — each provides domain-specific mental models and behavioral priming.
27 operations across identity, workspace management, problems, proposals, consensus, channels, and status reporting.
Auditable Reasoning
Every thought is a node in a graph — numbered, timestamped, linked to its predecessors, and persisted across sessions. This creates an auditable trail of how conclusions were reached.
Agents can think forward, plan backward, branch into parallel explorations, revise earlier conclusions, and request autonomous critique via MCP sampling. Each pattern is a first-class operation:
| Pattern | Description | Use Case |
|---|---|---|
| Forward | Sequential 1→2→3→N progression | Exploration, discovery, open-ended analysis |
| Backward | Start at goal (N), work back to start (1) | Planning, system design, working from known goals |
| Branching | Fork into parallel explorations (A, B, C...) | Comparing alternatives, A/B scenarios |
| Revision | Update earlier thoughts with new information | Error correction, refined understanding |
| Critique | Autonomous LLM review via MCP sampling | Self-checking, quality gates |
See the Patterns Cookbook for comprehensive examples.
Real-Time Observability
The Observatory is a built-in web UI at http://localhost:1729 for watching reasoning unfold live.
- Live Graph — thoughts appear as nodes in real-time via WebSocket
- Branch Navigation — branches collapse into clickable stubs; drill in and back out
- Detail Panel — click any node to view full thought content
- Multi-Session — switch between active reasoning sessions
- Deep Analysis — analyze sessions for reasoning patterns, cognitive load, and decision points
The full observability stack includes OpenTelemetry tracing, Prometheus metrics, and Grafana dashboards.
Knowledge & Reasoning Tools
Knowledge Graph — Persistent memory across sessions. Capture insights, concepts, workflows, and decisions as typed entities with typed relations (BUILDS_ON, CONTRADICTS, SUPERSEDES, etc.) and visibility controls (public, agent-private, team-private).
Mental Models — 15 structured reasoning frameworks (five-whys, pre-mortem, steelmanning, trade-off matrix, decomposition, and more) that provide process scaffolds for different problem types.
Notebooks — Interactive literate programming combining documentation with executable JavaScript/TypeScript in isolated environments.
Client Compatibility
Thoughtbox is currently optimized for Claude Code. We are actively working on supporting additional MCP clients. Due to variation in capability support across the MCP ecosystem — server features (prompts, resources, tools), client features (roots, sampling, elicitation), and behaviors like
listChangednotifications — we implement custom adaptations for many clients.
If you're using a client other than Claude Code and encounter issues, please open an issue describing your client and the problem.
Installation
Thoughtbox runs as a Docker-based MCP server. It requires Docker and Docker Compose.
Quick Start
git clone https://github.com/Kastalien-Research/thoughtbox.git
cd thoughtbox
docker compose up --build
This starts Thoughtbox and the full observability stack. The MCP server listens on port 1731 and the Observatory UI is available at http://localhost:1729.
MCP Client Configuration
Since Thoughtbox uses HTTP transport, configure your MCP client to connect via URL.
Claude Code
Add to your ~/.claude/settings.json or project .claude/settings.json:
{
"mcpServers": {
"thoughtbox": {
"url": "http://localhost:1731/mcp"
}
}
}
To connect through the observability sidecar (adds OpenTelemetry tracing):
{
"mcpServers": {
"thoughtbox": {
"url": "http://localhost:4000/mcp"
}
}
}
Cline / VS Code
Add to your MCP settings or .vscode/mcp.json:
{
"servers": {
"thoughtbox": {
"url": "http://localhost:1731/mcp"
}
}
}
Usage Examples
Forward Thinking — Problem Analysis
Thought 1: "Users report slow checkout. Let's analyze..."
Thought 2: "Data shows 45s average, target is 10s..."
Thought 3: "Root causes: 3 API calls, no caching..."
Thought 4: "Options: Redis cache, query optimization, parallel calls..."
Thought 5: "Recommendation: Implement Redis cache for product data"
Backward Thinking — System Design
Thought 8: [GOAL] "System handles 10k req/s with <100ms latency"
Thought 7: "Before that: monitoring and alerting operational"
Thought 6: "Before that: resilience patterns implemented"
Thought 5: "Before that: caching layer with invalidation"
...
Thought 1: [START] "Current state: 1k req/s, 500ms latency"
Branching — Comparing Alternatives
Thought 4: "Need to choose database architecture..."
Branch A (thought 5): branchId="sql-path"
"PostgreSQL: ACID compliance, mature tooling, relational integrity"
Branch B (thought 5): branchId="nosql-path"
"MongoDB: Flexible schema, horizontal scaling, document model"
Thought 6: [SYNTHESIS] "Use PostgreSQL for transactions, MongoDB for analytics"
Environment Variables
| Variable | Description | Default |
|---|---|---|
DISABLE_THOUGHT_LOGGING | Suppress thought logging to stderr | false |
THOUGHTBOX_DATA_DIR | Base directory for persistent storage | ~/.thoughtbox |
THOUGHTBOX_PROJECT | Project scope for session isolation | _default |
THOUGHTBOX_TRANSPORT | Transport type (stdio or http) | http |
THOUGHTBOX_STORAGE | Storage backend (fs or memory) | fs |
THOUGHTBOX_OBSERVATORY_ENABLED | Enable Observatory web UI | false |
THOUGHTBOX_OBSERVATORY_PORT | Observatory UI port | 1729 |
THOUGHTBOX_OBSERVATORY_CORS | CORS origins for Observatory (comma-separated) | (none) |
THOUGHTBOX_AGENT_ID | Pre-assigned Hub agent ID | (none) |
THOUGHTBOX_AGENT_NAME | Pre-assigned Hub agent name | (none) |
THOUGHTBOX_EVENTS_ENABLED | Enable event emission | false |
THOUGHTBOX_EVENTS_DEST | Event destination | stderr |
PORT | HTTP server port | 1731 |
HOST | HTTP server bind address | 0.0.0.0 |
NODE_ENV | Node environment | (none) |
PROMETHEUS_URL | Prometheus endpoint (Docker) | http://prometheus:9090 |
GRAFANA_URL | Grafana endpoint (Docker) | http://localhost:3001 |
Development
For local development (requires Node.js 22+):
npm install
npm run build
npm run dev # Development with hot reload
Testing
npx vitest run # Unit tests
npm test # Agentic tests — full suite (build + run)
npm run test:tool # Agentic tests — tool-level only
npm run test:quick # Agentic tests — quick (no build)
npm run test:behavioral # Behavioral contract tests
Docker Compose
docker compose up --build starts the full stack:
| Service | Port | Description |
|---|---|---|
| thoughtbox | 1731 (MCP), 1729 (Observatory) | Core MCP server + Observatory UI |
| mcp-sidecar | 4000 | Observability proxy with OpenTelemetry |
| otel-collector | 4318 (HTTP), 8889 (metrics) | OpenTelemetry Collector |
| prometheus | 9090 | Metrics storage + alerting |
| grafana | 3001 | Dashboards and visualization |
Persistent data is stored in named volumes: thoughtbox-data, prometheus-data, grafana-data.
Architecture
src/
├── index.ts # Entry point (stdio/HTTP transport selection)
├── server-factory.ts # MCP server factory with tool registration
├── tool-registry.ts # Progressive disclosure (stage-based tool enabling)
├── tool-descriptions.ts # Stage-specific tool descriptions
├── thought-handler.ts # Thoughtbox tool logic with critique support
├── gateway/ # Always-on routing tool
│ ├── gateway-handler.ts # Routes to handlers with stage enforcement
│ └── operations.ts # Gateway operations catalog
├── init/ # Init workflow and state management
│ ├── tool-handler.ts # Init tool operations
│ └── state-manager.ts # Session state persistence
├── sessions/ # Session tool handler
├── sampling/ # Autonomous critique via MCP sampling
│ └── handler.ts # SamplingHandler for LLM critique requests
├── persistence/ # Storage layer
│ ├── storage.ts # InMemoryStorage with LinkedThoughtStore
│ └── filesystem-storage.ts # FileSystemStorage with atomic writes
├── observatory/ # Real-time visualization
│ ├── ui/ # Self-contained HTML/CSS/JS
│ ├── ws-server.ts # WebSocket server for live updates
│ └── emitter.ts # Event emission for thought changes
├── hub/ # Multi-agent collaboration
│ ├── identity.ts # Agent registration
│ ├── workspace.ts # Workspace management
│ ├── problems.ts # Problem tracking with dependencies
│ ├── proposals.ts # Solution proposals with reviews
│ ├── consensus.ts # Decision recording
│ ├── channels.ts # Problem-scoped messaging
│ ├── hub-handler.ts # Hub operation dispatcher
│ └── operations.ts # 27-operation catalog
├── knowledge/ # Knowledge graph memory
├── events/ # Event emission system
├── mental-models/ # 15 reasoning frameworks
├── notebook/ # Literate programming engine
├── observability/ # Prometheus/Grafana integration
└── resources/ # Documentation and patterns cookbook
Storage
Thoughtbox supports two storage backends:
- InMemoryStorage: Default for development, uses
LinkedThoughtStorefor O(1) thought lookups - FileSystemStorage: Persistent storage with atomic writes and project isolation
Data is stored at ~/.thoughtbox/ by default:
~/.thoughtbox/
├── config.json # Global configuration
└── projects/
└── {project}/
└── sessions/
└── {date}/
└── {session-id}/
├── manifest.json
└── {thought-number}.json
Contributing
We welcome contributions! See CONTRIBUTING.md for:
- Development setup
- Commit conventions (optimized for
thick_readcode comprehension) - Testing with vitest and agentic scripts
- Pull request process
License
MIT License — free to use, modify, and distribute.
Related Servers
Etsy
A TypeScript-based MCP server for interacting with the Etsy API, featuring a simple notes system.
MCP GameBoy Server
Allows Large Language Models to interact with a GameBoy emulator via the serverboy MCP server.
FastMCP Calculator Server
A calculator server that performs basic math operations like addition, subtraction, multiplication, division, power, and square root.
Umami MCP server
MCP server exposing Umami analytics (Cloud + self-hosted)
VFX MCP
A powerful video editing server using ffmpeg-python to process external video files.
SO-ARM100 Robot Control with MCP
Control SO-ARM100 and LeKiwi robot arms using LLM-based AI agents.
OpenHue MCP Server
Control Philips Hue lights using LLM interfaces via the OpenHue CLI.
Bitnovo Pay
MCP server for Bitnovo Pay integration with AI agents. Provides cryptocurrency payment capabilities through Bitnovo Pay API. Features include payment creation, status checking, QR code generation, and webhook management with support for multiple tunnel providers (ngrok, zrok, manual).
企业经营分析洞察服务
Provides in-depth analysis of enterprise operations, including business status, development, and market performance.
Memory Bank MCP
A production-ready Model Context Protocol (MCP) server that provides a powerful, vector-native memory bank for AI agents. Built with the Protocol-Lattice Go Agent Framework, this server offers persistent, searchable, and shareable memory with multiple database backends.