Local LLMs. Full IDE. Autonomous agents. Image generation. Model training. Cloud deploy. All offline. All on your machine. Nothing sent anywhere.
Inference client. Code editor. Image tool. Deploy pipeline. Training platform. Five apps, five subscriptions, five surfaces for your IP to leak. Nobody built them to work together — because nobody had to, until now.
Context switch between inference, IDE, image generation, and deployment. Every handoff is friction, lost state, and engineer time. The tools were never designed to share context with each other.
AI editors route completions through remote APIs. Image tools process your work on third-party servers. Your source code, prompts, and IP flowing through infrastructure you didn’t choose.
You build locally, then hand off to entirely separate tooling for containers and cloud. The AI work you did has no direct path to what ships. You rebuild the pipeline every project.
Workbench doesn’t fill a gap. It closes all three. One app, running offline, combining six tool categories the industry kept separate for years — because combining them required building all six from scratch.
Each of these would ship as a standalone product. In Workbench, they share state, context, and models — making each one exponentially more useful than it would be alone.
GGUF inference with Flash Attention v3 and speculative decoding. Any model, any size. Auto-configured on first launch — no CUDA setup, no quantization guesswork. Mix local and cloud in the same session.
mbsd daemon (JSON-RPC 2.0, TCP:3031) + TS & Python SDKs
Monaco editor with full LSP, real PTY terminal, DAP debugger, and native Git. Every AI completion uses the local model in Pillar 1 — no additional API call, no latency, no context switch.
Describe a goal. The agent decomposes it, edits files across your codebase, runs terminal commands, calls APIs, tests results, iterates — live audit trail every step. Safety gates prevent anything destructive without your approval.
Complete Stable Diffusion pipeline and a professional Voice Studio — on your GPU, with no per-image cost, no API keys, no prompts shared anywhere. Generate until you’re satisfied.
Fine-tune any model on your own data. Watch it train. Export and load it directly into the inference engine. Cloud GPU rental is one click away when local hardware is not enough.
Every project built in Workbench ships to production without switching tools. Docker, Kubernetes, and 10 cloud providers in one panel. Cost estimates shown before you commit.
mbsd + mbs CLI + TS & Python SDKs for pipeline automationBecause all six pillars share state, the outcomes aren’t just the sum of the parts. The agent knows your codebase. Your fine-tuned model answers the agent’s questions. Deployment follows directly from the build.
Give the agent a goal. It scaffolds the project, writes the code, runs tests, fixes errors, and deploys to the cloud — while you watch every step in the audit log. No context switches. No different tools.
Local inference means completions in milliseconds. The model sees your whole codebase through semantic injection — not just the open file. Episodic memory carries project history across every session.
Upload a dataset, pick a base model, run LoRA fine-tuning on your GPU. Watch loss curves live. When training finishes, load it directly into the inference engine — no export pipeline, no cloud bill.
GSD-2 runs multi-phase agentic tasks autonomously. Checkpoint and resume on crash. Safety gates prevent destructive actions without your approval. Define the workflow once — Workbench runs it on schedule.
Stable Diffusion on your GPU with no per-image cost. Combine ControlNet conditioning with LoRA style adapters. Voice Studio’s 22-voice TTS for narrated content — zero data leaves your machine.
Multi-model chat mixes local and cloud providers in one session. Run adversarial debates, chain outputs between models, or benchmark your fine-tuned model against a frontier model — without swapping tools.
Not a concept. Not a prototype. Built phase by phase — each shipped, audited, and functional. Here are the numbers that matter.
mbsd daemon (JSON-RPC 2.0, TCP:3031) + TS & Python SDKsEvery feature on this page is implemented and functional. No vaporware, no stubbed commands.
Every inference, image gen, and agent action runs on your hardware by default. Nothing leaves unless you choose.
cargo clippy: 0 warnings, 0 errors. All former stubs replaced with real implementations.
Active community since early alpha. Every release shaped by real-world developer feedback.
Private beta. All six pillars included, all features unlocked. Request access and we’ll send download instructions as soon as your request is reviewed.
Invite-only. Fill in your details — we’ll send access instructions as soon as your request is reviewed.
Free during private beta. All six pillars with no feature gates. Premium tier ($10/mo) planned for cloud API credit bundles and team features — local inference, agents, training, and image generation will remain free forever.