A Manifesto

Teams are dead.
The factory is the
operating system.

A framework for the transition from individual AI productivity to Institutional AI — where agents, humans, and guardrails form a single operating system.

Read the thesis
Everyone is using AI to write code faster. Almost nobody is asking the harder question: what happens when the organisation itself becomes the machine?
The Factory Way, 2026

The Core Thesis

Most companies are stuck giving developers better autocomplete and calling it transformation. They are optimising the typist. The unit of productivity is still the individual.

The Factory Way is a different bet. What if high-quality software is a systems problem, not a people-and-talent problem? What if the right unit of productivity is the organisation, not the developer?

We call this Institutional AI — where agents don't assist individuals, they operate as an organisational system. Humans govern, direct, and review. The factory ships.

The goal is not efficiency. The goal is to compress the time from idea to live, at a quality level that was previously impossible without an army.

The Shift

Individual AI vs Institutional AI

L4 — L8

Where most teams are

  • +98% More PRs opened
  • +91% Longer review times
  • +47% More context switching
  • Flat Delivery metrics unchanged
  • Each developer has their own AI setup. Agents don't know about each other. The pipeline clogs.
The line
L10+

Where the future lies

  • 1,300 PRs/week, zero human code (Stripe)
  • 50%+ Of merged PRs from agents (Ramp)
  • 70% Machine-generated code (Uber)
  • 21K Developer hours saved (Uber)
  • Agents operate within institutional infrastructure. Humans govern, direct, and review. The factory ships.
What the Data Says

Surprising Findings

-19%

Experienced developers were slower with AI tools in METR's controlled trial. They believed they were 20% faster — a 39-point perception gap.

Read more on L4 →
+98% PRs

Teams at L8 produce 98% more pull requests — but delivery metrics stay flat. Review time up 91%. Context switching up 47%. This is the trap.

Read more on L8 →
1,300/wk

Stripe's Minions ship 1,300+ PRs per week with zero human-written code. Built on a decade of standardised devboxes and 3 million tests.

Read more on L10 →
48%

Nearly half of AI-generated code contains security vulnerabilities. Self-healing loops make it worse — 37.6% more critical vulnerabilities after 5 iterations.

Read the research →
Self-Assessment

Where is your organisation?

Answer 5 questions to find your approximate level on the framework.

1 / 5

How does your team primarily use AI in development?

What Institutional AI Looks Like

The operating system for Institutional AI

Most tools help individual developers move faster. The future lies in infrastructure that makes the organisation ship differently.

01

Structured Intake

Work enters through defined channels — tickets, messages, system events. Every task carries intent, context, and acceptance criteria before an agent touches it.

02

Agent Orchestration

Coordinated multi-agent pipelines that plan, implement, test, and deliver within your existing infrastructure. Not a sandbox. Your repos, your CI, your standards.

03

Guardrails Engine

Guardrails as a first-class function, not an afterthought. Policy-driven validation at every stage. Because in regulated industries, unguarded agent output is liability.

04

Reflect Cycle

Every task deposits insight. The system learns what works, what fails, and why. Prompt study and institutional memory that compounds over time.

05

Capability Routing

The factory thinks in capability tiers, not brand names. No single model failure breaks the system. Your intelligence layer survives any vendor's bad quarter.

06

Evidence Trail

Tool logs, linked artifacts, full traceability. Not agent self-reporting. If it can't be traced, it didn't happen. Built for audit, built for trust.

Companies operating at L10+
Stripe Uber Ramp Coinbase Google Block
See the data →

Non-Negotiables

01

Without guardrails, AI work is a liability.

In regulated industries, unguarded agent output isn't innovation — it's risk. Guardrails are a first-class organisational function, not a compliance afterthought.

02

Evidence over assertion.

Verification requires tool logs and linked artifacts — not agent self-reporting. If it can't be traced, it didn't happen.

03

Stay in the cheapest reasoning space.

Planning is cheap. Execution is expensive. Most teams burn resources executing a mediocre plan when they should still be iterating on the idea. Stay in plan space until the thinking converges.

04

The factory learns.

Knowledge lives in the improvement cycles. Every task deposits insight. The reflect cycle turns raw experience into reusable intelligence. Prompt study — understanding what works and why — is the discipline that compounds.

05

Intent persistence is a first-class concern.

"The human asked me" is not intent. The system captures WHY, not just WHAT.

06

The factory thinks in capabilities, not brand names.

No single model failure should break the system. The factory references capability tiers and routes accordingly. Your intelligence layer should survive any vendor's bad quarter.

North Star Metric

Time from customer request to deployed capability in a new market.

Coming Soon

Building for enterprises where
failure has real consequences.

The path from wherever you are on the 12 levels to Institutional AI requires more than better tools. If this resonates, we want to hear from you.

Early Access

Get on the list

The transition to Institutional AI is coming. Leave your email and tell us what you're most interested in.

Get in Touch

Let's talk.

Whether you're exploring AI adoption, stuck at L8, or ready to build institutional infrastructure — we want to hear from you.