A journey to safe, governed, autonomous software development — from concept to production-ready platform.
When the Switch Product Development organization onboarded Claude Code and its powerful foundation models into our software development lifecycle, it fundamentally changed how we build software. AI agents could now write code, debug issues, and complete complex tasks autonomously with predictable quality guarantees. But this leap forward surfaced a critical question for any enterprise organization: How do we harness this transformative technology while maintaining safety, consistency, and control?
In the course of a month, the team recognized this gap and responded with notable speed. We built a complete end-to-end AI Ops platform from scratch — a system that provides the governance, infrastructure, and specifications needed to use AI assistants and agentic coding tools safely, consistently, and productively at scale.
The question was never whether to adopt AI code assistants — it was how to use them safely, consistently, and productively at enterprise scale.
SGA (Switch Governed Agents) eases the governance challenge by transforming organizational policy into runtime guidance that AI assistants automatically follow. Instead of asking developers to continuously remember and teach AI agents our rules and standards, we compile a central policy registry into governance artifacts that load automatically.
Policy maintainers define rules, behaviors, skills, and agent templates in a central registry stored in Git. Developers run sga sync to pull the latest policy, then sga apply to compile it into their workspace. This generates:
@backend-implementer or @security-reviewer that automatically embed relevant rules/sga-security for explicit verificationThis solves three critical problems: governance without friction (developers don't need to remember rules), consistency across teams (everyone uses the same compiled artifacts), and adaptability (policy updates propagate automatically).
CNAM (Cloud-Native Agentic Memory) addresses a fundamental limitation of current AI assistants: they lack memory between sessions. Every interaction starts fresh, requiring users to repeatedly provide context. CNAM provides a multi-tenant memory system that AI assistants can query and update as they work through MCP.
Memories exist at two scopes: personal memories visible only to the individual user, and organizational memories shared across teams — enabling both private knowledge accumulation and collective institutional learning. The system classifies memories by type: factual (infrastructure standards, configurations, architectural decisions), experiential (debugging insights, lessons learned), and working (temporary session context that auto-expires). Built on PostgreSQL, Qdrant for vector search, and Ollama for embeddings — a production-grade cloud-native service with admin console, API key provisioning, and audit logging.
Runcell enables the most powerful — and potentially risky — capability of AI assistants: autonomous execution. It creates isolated workspace containers where task execution and experimentation happens safely.
The architecture introduces a critical distinction: Outer Agent versus Inner Agent. The outer agent runs in your terminal, orchestrating work and creating workspaces. The inner agent runs inside those workspaces with its own API key, working autonomously on delegated tasks. This separation provides security isolation and enables parallel autonomous workflows.
Two deployment modes are supported. Local mode runs on individual machines using Podman for fast, network-independent workspace creation. Remote mode connects to a centralized Runcell server that orchestrates workspaces as Kubernetes pods, providing centralized management, shared infrastructure, and full audit logging.
Security is baked in through multiple layers: containers run as non-root users with dropped capabilities, seccomp profiles restrict dangerous syscalls, cgroups enforce resource limits, and networks can be fully isolated. Credentials are managed securely through the macOS Keychain and injected into containers at runtime.
Simplex is a workflow specification designed specifically for autonomous agents. It occupies a crucial middle ground between natural language (too ambiguous) and programming languages (too prescriptive). Simplex captures what needs to be done and how to know when it's done, without prescribing how to do it.
The specification uses landmarks — structural markers like FUNCTION, RULES, DONE_WHEN, EXAMPLES, and ERRORS — that agents recognize and orient around. Five pillars guide the design:
The true power of the platform emerges when these components work in concert. Here's the end-to-end workflow:
sga apply to compile policy into their workspace, generating role-specific agents, auto-triggering skills, and baseline governance context.The Model Context Protocol (MCP) serves as the integration layer, standardizing how AI assistants communicate with these systems. CNAM provides MCP tools for memory operations, Runcell provides MCP tools for workspace and file operations, and SGA provides governance context that all agents automatically inherit.
What this platform enables next is transformative. Organizations can define Simplex specifications for complex software systems and have autonomous agents build them end-to-end, with governance automatically enforced, context accumulated across iterations, and execution safely isolated.
Teams can share knowledge through CNAM's org-scoped memories, avoiding the rediscovery of solutions to known problems. Policy can evolve centrally and propagate automatically, keeping governance in sync with organizational learning.
The broader vision for AI Ops is clear: create the infrastructure that enables safe, governed, productive use of AI across the entire software development lifecycle. Now we expand — integrating with CI/CD pipelines, extending to testing and deployment, enabling multi-agent swarm architectures, and building observability into AI-assisted workflows.
The question is no longer "Can we use AI assistants?" The question is "How can we use AI assistants to transform software development?" With our AI Ops platform, we have the answer.
Special thanks to Logan Broitman, Ahmed Schrute, Steven Shurtliff, Nate Lantz, Leah Hunnewell, Jared Bischoff, and Travis Harrison for their contributions to making this platform a reality.