TabTab OS is the AI Agents Operating System that runs your Tab — six coordinated agents handling development, marketing, sales, outreach, support, and operations. It runs locally on dedicated hardware, hosted at the Bench.
Every TabTab deployment runs on dedicated AI hardware — your data, your runtime, your platform. We provision, configure, and operate the hardware on your behalf. The local-first architecture is deliberate: most agent operations run on local LLMs on the Tab’s hardware, with cloud APIs used selectively. As accelerated compute matures and clustering technologies improve, the ecosystem scales through hardware aggregation rather than cloud API dependency. This is the architectural foundation of our local-first thesis.
The principles of TabTab OS — local LLMs, owned data, no cloud lock-in, sealed nodes — work on any sufficiently capable local-first hardware. For the initial Tab cohort, we chose dedicated local hardware for three reasons:
Unified memory shares resources between CPU, GPU, and Neural Engine. 30B+ parameter models run at full context locally on a single machine smaller than a hardcover book.
Dedicated hardware sits in a rack for years without maintenance. Tabs need infrastructure, not workstations. No server racks, no cooling, no IT department to maintain it.
New machines ship with predictable security guarantees, hardware enclaves, and Tailscale support. One vendor, one supply chain, one threat model.
Every Tab runs on a dedicated node — hosted, managed, and monitored at the Bench in Irvine. One machine per Tab. No multi-tenancy.
Hosted at the Bench. Runs the directors and heavy brains. Indexes the vault around the clock. Active cooling for sustained inference.
Active cooling allows sustained multi-hour inference runs — massive vector embeddings, overnight vault indexing — without thermal throttling.
No screen, no battery, no keyboard. Every gram is inference power. Ideal for running 30B+ parameter models at full context locally.
Hosted at the Bench, monitored 24/7. Runs headless, SSH-disabled, Tailscale-only. The silent anchor of your org chart.
Every TabTab node ships with defense-in-depth security that eliminates the most common attack vectors in AI agent deployments. Your Tab’s data never leaves your dedicated Tab unless you say so.
SSH disabled at factory. No open ports. Tailscale loopback binding only. The node is invisible to the public internet.
Infrastructure email accounts isolate inbound data. Verified sender lists prevent prompt injection from malicious emails reaching agent context.
The Steward handles strategy via a strict JSON schema. The Builder handles execution. They communicate through structured handoffs — no free-form LLM-to-LLM chatter that can be hijacked.
WireGuard encryption via Tailscale. All node-to-node traffic is end-to-end encrypted. Zero data traverses the public internet in cleartext.
Every AI deployment falls somewhere on this spectrum. Tabs sit deliberately in the middle — dedicated hardware, your data, but with the operational simplicity of a managed service.
The current architecture — one node, one Tab — is the starting point, not the ceiling. Unified memory architecture and the emerging ecosystem around local model clustering (Exo, RDMA-style interconnects, Thunderbolt fabric) point toward a future where multiple machines can pool inference capacity without cloud dependency.
We are not shipping distributed inference today. What we are doing is making every architectural decision with that future in mind: standardized hardware, local-first model execution, sealed nodes with mesh networking already in place. When the clustering layer is production-ready, every Tab in the portfolio is positioned to benefit without re-architecture.
This is the difference between a platform that happens to run locally and a platform designed for a local-first future. The hardware decisions today are the infrastructure decisions of tomorrow.
Hardware procured new. Configured at the Bench. Sealed and deployed. Three-hour onboarding loads your knowledge vault. By Friday, your Tab is operating.