Tadpole OS is a local-first runtime for orchestrating autonomous teams of AI agents — without sending your data to the cloud. You define a goal, assign a hierarchy of specialized agents (each with its own model, memory, and toolset), and the engine coordinates them in parallel. With Agent 99: Self-Annealing, the system autonomously extracts architectural wisdom from every mission, persisting institutional knowledge into its long-term memory for cross-mission learning. Whether you're automating a financial audit, a marketing pipeline, or a software deployment, Tadpole OS gives you the infrastructure to run complex, multi-step AI workflows with full visibility, strict budget controls, and the option to keep every token of data completely private.
Note
Status: Production-Ready / 100% Audit Verified
Version: 1.1.4
Documentation Level: Sovereign (Last Hardened: 2026-04-12)
- 🚀 Getting Started
- 🎯 Test Missions
- 🧩 Core Features
- 🤖 Multi-Model Management
- 📚 Documentation Excellence
- 🔱 Sovereign Forking Protocol
- 🛡️ License
Ready to take the swarm for a spin? We've curated a set of Small Missions to help you verify and experience the core skills of Tadpole OS, from basic model connectivity to autonomous background execution.
Tadpole OS is organized around six capability pillars. Each pillar covers a distinct domain — click "Full details" to expand the complete feature list for that area.
A sovereign multi-monitor dashboard built for high-density swarm oversight.
- Multi-Tab Interface — persistent navigation across operations, missions, and configs
- Unified Observability Sidebar — combines System Log, Neural Waterfall, and Swarm Pulse into one context-aware panel
- Detachable Portals — spread tactical sectors across multiple physical displays
- 10Hz Swarm Pulse — sub-millisecond binary telemetry for real-time agent performance and system health
- Reconnection Intelligence — WebSocket auto-reconnects on settings change, no refresh needed
Full details →
- Multi-Tab Sovereign Interface: Persistent navigation via a multi-tab bar, allowing seamless switching between operations, missions, and configurations.
- Observability Stack (Unified Sidebar): A smart, context-aware
Observability_Sidebarthat unifies the System Log, Neural Waterfall, and Swarm Pulse. It automatically synchronizes with detachable portals to prevent duplicate data streams. - Multi-Monitor Optimized: Support for Detachable Portals, enabling operators to spread tactical sectors across multiple physical displays for high-density oversight.
- Unified Tactical Header: Dynamic, context-aware header featuring a Smooth Carousel for telemetry metrics and action buttons.
- Live Stat Metrics: Real-time visualization of agent performance, recruitment velocity, and system health.
- Reconnection Intelligence: WebSocket auto-connects when settings (URL/Key) change—no refresh needed.
- Zustand Reactive Stores: Unitary source of truth for settings, agents, and providers.
- Lazy Singleton Socket: Performance-optimized socket initialization via Proxy.
- Command Palette: Global
Cmd/Ctrl+KorCmd/Ctrl+/navigation for agents and actions. - Neural Lineage Breadcrumbs: Real-time visibility into the swarm's recursive hierarchy.
- Swarm Visualizer (God-View): High-performance 2D Force-Graph with animated MessagePack pulse traces and real-time node "heartbeat" indicators.
The core engine for recruiting, coordinating, and governing hierarchical agent swarms.
- CEO/COO Patterns — explicit recursive delegation from strategic to tactical layers
- Hierarchical Recruitment — agents spawn ephemeral sub-agents via the
recruit_specialistMCP tool - Parallel Swarming — concurrent multi-agent execution via
FuturesUnordered - Lifecycle Hooks —
pre-tool/post-toolexecutable governance for security auditing - Triple-Slot Model Routing — every agent has Primary, Secondary, and Tertiary model slots with independent providers and prompts
Full details →
- Strategic Intent Handoffs: Parent agents inject tactical context into sub-agent neural pathways.
- Recursive Swarm Protocols: Explicit CEO (Agent of Nine) and COO (Tadpole Alpha) patterns for strategic-to-tactical delegation.
- Hierarchical Recruitment: High-level agents can recruit ephemeral sub-agents via the
recruit_specialistMCP tool. - Parallel Swarming: Concurrent multi-agent recruitment via
FuturesUnordered. - Lifecycle Hooks:
pre-toolandpost-toolexecutable governance for security auditing. - Intelligence Forge: Configure multiple modalities (LLM, REASONING, Vision) for the same model node.
- Local LLM Support (Ollama): Seamless integration with local Ollama nodes for zero-latency, truly private execution.
- Mission Analysis (Agent 99): Automated post-mission debriefs and Self-Annealing logic that extracts architectural wisdom from logs and persists it to institutional memory.
- Triple-Slot Configuration: Every agent possesses three model slots (Primary, Secondary, Tertiary) with granular per-slot provider, system prompt, and temperature control.
- Manual Routing: Using the "Zap" (Skill) icon on an agent node, operators can manually dispatch specific skills to specific slots.
- Recursive Swarm Auto-Registration: Parent agents automatically discover and register sub-agent capabilities into the AI Services registry, enabling cross-mission intelligence reuse.
Persistent, cross-session memory powered by a split-brain SQLite + LanceDB architecture.
- Persistent Vector Memory — LanceDB-backed semantic search across sessions
- Mission RAG Scopes — sandboxed per-mission search spaces, aggregated before synthesis
- Identity & Long-Term Memory — global
IDENTITY.mdandLONG_TERM_MEMORY.mdinjected at runtime - Cross-Mission Pattern Recognition — Agent 99 detects behavioral drift and deduplicates swarm memory
- Orphaned Scope Cleanup — background Tokio service sweeps completed mission vector stores automatically
Full details →
- Persistent Vector Memory: Powered by LanceDB for cross-session institutional knowledge.
- Mission RAG Scopes: Agents spin up localized, sandboxed semantic search spaces per-mission to aggregate findings before synthesis.
- Persistent Memory & Identity: Global
directives/IDENTITY.mdanddirectives/LONG_TERM_MEMORY.mdinjection. - Mission Analysis & Wisdom Write-Back (Agent 99): Automated post-mission debriefs featuring Self-Annealing logic. The system distills architectural insights from mission logs and autonomously updates
directives/LONG_TERM_MEMORY.md, ensuring the swarm learns and evolves across missions. - Split-Brain Memory Architecture: SQLite handles deterministic relational data (logs, budgets), while LanceDB handles high-dimensional vector embeddings — a top-tier industry standard for RAG applications.
- Orphaned Scope Cleanup: A Tokio background service automatically sweeps completed/failed mission vector stores, preventing unbounded disk bloat.
- Unified Persistence: All agents and skill templates persist strictly via a unified SQLite schema — no disparate JSON registry or manual file synchronization required.
Zero-trust governance with a hard gate for sensitive actions and full local-only operation.
- Sapphire Shield — strict schema validation; flags
budget:spendandshell:executefor mandatory human approval - Privacy Mode — hard gate that blocks all external cloud traffic for 100% data sovereignty
- Continuity Scheduler — cron-driven background swarm execution with per-mission budget caps and auto-disable safety
- Graceful Degradation — Null Provider fallback ensures the UI degrades gracefully (Amber Badging) rather than hard-crashing
- CodeQL Verified — automated vulnerability scanning on every commit
Full details →
- Sapphire Shield (Phase 1): Strict schema validation and an automated security gate that flags
budget:spendorshell:executefor mandatory human oversight. Includes BroadcastChannel session synchronization forNeural_Vaultstate consistency across multiple browser tabs. - Autonomous Continuity (Phase 2): Cron-driven Continuity Scheduler enabling background, proactive swarm execution with strict per-mission budget caps and auto-disable safety mechanisms for runaway failure prevention.
- Graceful Degradation (Phase 3): Trait-driven Null Provider fallback ensures integration tests pass seamlessly without API keys and the UI degrades gracefully (Amber Badging) rather than hard-crashing during provider outages.
- Privacy Mode (Shield): A "Hard Gate" that explicitly blocks all external cloud traffic for 100% data sovereignty.
- Pre-deployment Verification Suite:
verify_all.pyenforces documentation parity, security audit, system health, and latency benchmarking before every deploy. - Yield Phase Transition: Rigid lock-step transitions for secure
Security_Huboverwatch. - ADG Governance: 100% automated sync between code and docs via the
parity_guard.pysystem.
A Rust-native engine engineered for sub-millisecond initialization and zero idle overhead.
- 10Hz Binary Telemetry — Sub-millisecond
tadpole-pulse-v1protocol using MessagePack for zero-latency swarm updates - Neural Context Pruning — automatic truncation based on TPM limits using
tiktoken-rs - Raw Tool Output — tool handlers return results directly, eliminating redundant per-tool LLM synthesis calls
- Cached Context —
IDENTITY.mdandLONG_TERM_MEMORY.mdloaded once at startup viaparking_lot::Mutex - Unified Provider Dispatch — all LLM calls route through a single
dispatch_to_provider(); adding a new provider is a one-line edit
Full details →
- Shared Rate Limiting: Model-level RPM/TPM quotas enforced across all concurrent agents.
- Neural Context Pruning: Automatic truncation based on TPM limits using
tiktoken-rs. - Unified Provider Dispatch (PERF-09): All LLM provider calls route through a single
dispatch_to_provider()— adding a new provider is a one-edit operation. - Raw Tool Output (PERF-10): Tool handlers return raw results directly instead of making redundant per-tool synthesis LLM calls, eliminating token cost doubling.
- Cached Context (PERF-11):
IDENTITY.mdandLONG_TERM_MEMORY.mdloaded once at startup;parking_lot::Mutexfor all synchronous locks; bounded tool cache (64 entries);VecDequeoversight ledger. - First Principles Architecture: Lazy-loaded core engine resources with
tokio::sync::OnceCellfor sub-millisecond initialization and zero idle RAM consumption. - Performance Analytics Hub: Persistent benchmarking system for tracking latency (mean, p95, p99) and performance deltas across all agent actions.
- Real-Time Telemetry: Corrected TPM calculation and bridged backend-frontend token usage mapping for 100% accurate display.
- HATEOAS & Pulse Alignment: REST Level 3 discovery and binary WebSocket multiplexing for sub-millisecond parity.
A plug-and-play ecosystem for extending agent capabilities and running fully air-gapped swarms.
- Unified MCP Skill Model — standardized tool discovery and execution via the Model Context Protocol
- Capability Import Engine — one-click
.mdskill import with a structured preview gate before registration - Zero-Conf Swarm Discovery — one-click local network scanning via mDNS (
_tadpole._tcp.local.) - Sovereign Intelligence Store — browse and pull Ollama/Hugging Face models with integrated VRAM profiling
- In-App Swarm Template Store — browse and install industry-specific agent swarms directly from the dashboard
Full details →
- Unified MCP Skill Model: Standardized tool discovery and execution via the Model Context Protocol. Support for native
.mdskill/workflow imports. - Capability Import Engine: One-click import for documented skills and workflows via the "Import .md" gate. Includes a structured Import Preview for safety verification before registration.
- Zero-Conf Swarm Discovery: One-click local network scanning via mDNS (
_tadpole._tcp.local.). - Privacy Mode (Shield): A "Hard Gate" that explicitly blocks all external cloud traffic for 100% data sovereignty.
- Sovereign Intelligence Store: Browse and pull models (Ollama/Hugging Face) with integrated VRAM Hardware Profiling to prevent node-level OOM failures.
- Multi-Bunker Registry: Unified dashboard for multi-node swarm management across your entire local network.
- Neural Voice Integration: Local-first Piper TTS, Whisper STT, Silero VAD, SQLite-backed audio cache, and real-time PCM streaming over binary WebSocket.
- In-App Store: Browse and install industry-specific agent swarms directly from the dashboard.
- GitHub Native Hub: Leverages public GitHub repositories as a decentralized distribution hub for swarm templates, installed natively via Git.
- Automated Settings Migration: Seamless transition from legacy OpenClaw configurations to the Tadpole OS unified settings store.
Tadpole OS provides "One-Click" deployment for specialized industry swarms via our Template Ecosystem.
- 💸 Financial Services: Automated bookkeeping and real-time audit trails.
- 📣 Digital Marketing: Full-funnel content and ad optimization.
- 🏠 Real Estate: High-velocity lead and transaction management.
- 🏭 Manufacturing: Multi-tier templates for Job Shops (25 seats) up to Smart Factories (100 seats).
- ⚖️ Legal & Medical: Specialized compliance-first intake and review swarms.
👉 Explore the Template Registry
- Rust (1.80+): Required for compiling the backend.
- Node.js 20+: Required for the React dashboard.
- AI Provider Key (Optional but Recommended): Configure at least one provider key (
GOOGLE_API_KEY,OPENAI_API_KEY,ANTHROPIC_API_KEY, orGROQ_API_KEY) or run local-only via Ollama.
-
Install Frontend Dependencies:
npm install
-
Prepare Backend: The backend is located in
server-rs/. It will be compiled on the first run via thenpmscripts. -
Configure Environment: Create a
.envfile in the root directory:NEURAL_TOKEN=your_secret_token_here GOOGLE_API_KEY=your_api_key_here
Important
Required: NEURAL_TOKEN must be set — the engine panics at startup if it is missing.
Sovereign Warning: At startup, the engine performs a diagnostic check for configured AI providers. If no keys are found, it will issue a warning and guide you to docs/GETTING_STARTED.md.
Start the high-performance Rust engine:
npm run engineThe engine will start on http://localhost:8000.
Note
Production Access: When deployed via Docker, the engine serves both the API and the React dashboard on this same port (8000). You do not need port 5173 for production.
In a separate terminal, start the React development server:
npm run devBefore pushing changes or deploying to a Production Bunker, you MUST run the full verification suite to ensure 100% system parity and security compliance:
python execution/verify_all.py . --url http://localhost:8000This gate enforces:
- Documentation Parity (P0): Zero-drift detection between codebases and distributed
.mddocumentation viaparity_guard.py. - Security Audit (P0): Deep vulnerability scanning, secret detection, and dependency risk analysis.
- System Health (P1): Complete unit and integration test suite pass.
- Performance Benchmarking: Latency p95/p99 validation for the core engine.
Here is what a Full-Capacity Swarm looks like when utilizing all potential offerings (10 Providers, 25 Agents, 10 Clusters, Recursion Depth 10).
graph TD
classDef cluster fill:#222,stroke:#444,stroke-width:2px,color:#fff;
classDef node fill:#333,stroke:#666,color:#eee;
subgraph Cluster0 ["Executive Core (Google)"]
CEO["ID 1: CEO (Google Gemini 3 Pro)"]
end
CEO -->|issue_alpha_directive| COO["ID 2: COO (Claude Opus 4.5)"]
CEO -->|delegate| CTO["ID 3: CTO (GPT-5.3 Codex)"]
subgraph Cluster1 ["Operations Hub (Anthropic)"]
COO --> Ops1["ID 22: HR Manager"]
COO --> Ops2["ID 10: Support Lead"]
end
subgraph Cluster2 ["Engineering Sector (Groq/OpenAI)"]
CTO --> Eng1["ID 7: DevOps"]
CTO --> Eng2["ID 8: Backend Dev"]
end
subgraph Cluster3 ["Marketing/Sales (Meta/xAI)"]
CEO --> CMO["ID 4: CMO"]
CMO --> Mark1["ID 17: Copywriter"]
CMO --> Mark2["ID 19: SEO Specialist"]
end
subgraph Cluster4 ["Security Center (Mistral)"]
Eng1 --> Sec1["ID 12: Security Auditor"]
end
%% Depth levels
subgraph Level1 ["Depth 1 (Alphas)"]
COO
CTO
CMO
end
subgraph Level2 ["Depth 2 (Leads)"]
Eng1
Ops2
Mark1
end
subgraph Level3 ["Depth 3 (Specialists)"]
Sec1
Mark2
end
subgraph Level4 ["Depth 4 (Sub-Specialists)"]
%% Spawning dynamically
Sub1["Dynamic Researcher"]
end
subgraph Level10 ["Depth 10 (Terminal Layer)"]
%% Hard recursion limit
Sub2["Max Depth Agent"]
end
Sub1 --> Sub2
Sec1 --> Sub1
class Cluster0,Cluster1,Cluster2,Cluster3,Cluster4 cluster;
class CEO,COO,CTO,CMO,Ops1,Ops2,Eng1,Eng2,Mark1,Mark2,Sec1,Sub1,Sub2 node;
| Cluster | Focus | Provider (Sample) | Model Capacity |
|---|---|---|---|
| Executive Core | Strategic Direction | Pro / Flash | |
| Operations Hub | Orchestration | Anthropic | Opus / Sonnet |
| Engineering Sector | Implementation | Groq / OpenAI | Llama / Codex |
| Marketing/Sales | Growth | Meta / xAI | Maverick / Grok |
| Security Center | Auditing | Mistral | Medium / Large |
| Product Lab | R&D | DeepSeek | V3 / R1 |
| Finance Sector | Fiscal Control | Perplexity (API) | Sonar |
| Human Capital | HR/Logistics | OpenSource | Qwen / GLM |
| Legal/Compliance | Risk Mgmt | Local | Ollama/Llama3 |
| Logistics | Fleet Mgmt | Azure | GPT-4o |
- Distributed Compute: 10 clusters allow for complete physical or logical isolation of workloads, preventing mission interference.
- Provider Redundancy: If a primary provider (e.g., Google) hits a rate limit, the swarm can dynamically reassign agents to 9 other fallback providers.
- Recursive Density: With a depth of 10, a single strategic prompt can fan out into over 100 concurrent sub-tasks if each layer recruits just 3-4 specialists.
- Token Expenditure: At this scale, real-time cost monitoring (USD budget gate) becomes critical, as the "Neural Pulse" generates massive event bus traffic.
Tadpole OS supports Multi-Model Routing for every agent, allowing for manual failover or task-specific specialized model selection.
- Triple-Slot Configuration: Every agent possesses three model slots (Primary, Secondary, Tertiary).
- Granular Control: Each slot can have its own provider (e.g., Gemini 1.5 Pro on Primary, Llama 3 on Secondary), system prompt, and temperature.
- Manual Routing: Using the "Zap" (Skill) icon on an agent node, operators can manually dispatch specific skills to specific slots.
- Persistent Active Slot: The frontend tracks the
activeModelSlot(1, 2, or 3) to ensure the agent's "Neural Presence" reflects its current active intelligence.
Tadpole OS provides a comprehensive "Strategic Success" suite for both the Overlord (Entity 0) (AKA Human-in-the-loop) and AI assistants:
- System Architecture: Technical deep dive and data flows.
- Getting Started Guide: Hardware requirements and first deployment.
- Swarm Orchestration: Guide to designing hierarchical, autonomous intelligence clusters.
- Conceptual Glossary: Standardized terminology for Human-AI alignment.
- AI Codebase Map: Relationship guide optimized for AI development and navigation.
- Contributing Standards: Guidelines for joining the swarm.
- Project Roadmap: Future milestones and strategic goals.
- Support Guide: How to get help and report issues.
- API Reference: Complete REST and WebSocket endpoint documentation.
- Troubleshooting: Comprehensive diagnostics for swarm and engine issues.
- Release Process: Parity-guarded deployment and versioning protocols.
To ensure fast triaging and high-quality responses, please use our Issue Templates when submitting feedback:
Tadpole OS is built to be a foundational layer for your own AI endeavors. To create your own instance:
- Fork this Repository: Click the 'Fork' button to create your own tacticial branch.
- Setup Local Environment: Follow the Getting Started Guide.
- Customize your Agents: Use the SQLite database or Swarm Template ecosystem to dynamically deploy your own specialists.
- Deploy: Use the
deploy-bunker-1.ps1ordeploy-bunker-2.ps1script to push to your own Swarm Bunker.
We encourage "Overlords" to share their unique agent configurations and skill-sets back to the core via Pull Requests.
Tip
AI Readiness: You can ask any AI assistant (like Claude or Gemini) to read the CODEBASE_MAP.md and GLOSSARY.md to get an instant, expert-level understanding of this project's inner workings.
- Rust Engine (
server-rs/): The Axum-based server handling WebSockets, agent routing, and AI execution. - Agent Memory (
server-rs/src/agent/memory.rs): The LanceDB and Arrow-powered vector memory module utilizing Geminitext-embedding-004. - Persistence (
server-rs/src/agent/persistence.rs): Unified SQLite tracking for Agents, Sub-Agents, and their configurations. - Runner (
server-rs/src/agent/runner/mod.rs&analysis.rs): The native Tokio-based loop managing LLM communication, lock-step phase transitions, and post-mission synthesis. - Oversight Dash: The Overlord (Entity 0) (AKA Human-in-the-loop) safety gate for approving/rejecting sensitive actions.
MIT
