5 min read

The Delegation Economy

OpenAI released GPT-5.4 mini and nano today. The benchmarks are impressive. Mini scores 54.4% on SWE-Bench Pro, approaching the full GPT-5.4’s 57.7%. Nano costs $0.20 per million input tokens. Both are fast.

But the interesting part isn’t the models. It’s the language OpenAI used to describe them.

They didn’t say “smaller model for budget users.” They said: a larger model like GPT-5.4 can handle planning, coordination, and final judgment, while delegating to GPT-5.4 mini subagents that handle narrower subtasks in parallel.

Planning. Coordination. Judgment. Delegation.

That’s not a product description. That’s an org chart.

The Corporate Structure of Intelligence

For the past three years, AI has been sold as a single entity. One model. One conversation. One prompt in, one answer out. The metaphor was always the brilliant individual — the genius who knows everything.

GPT-5.4 mini/nano marks the moment OpenAI officially abandoned that metaphor. The new one is the corporation.

In Codex, GPT-5.4 is the executive. It reads the problem, makes a plan, and delegates. Mini subagents fan out — one searches the codebase, another reviews a file, a third processes documentation. They report back. The executive synthesizes and decides.

This is not new in practice. Developers have been chaining models for years. But it’s new as an official product architecture. OpenAI is selling a hierarchy, not a model.

The Price Ladder

Look at the pricing:

  • GPT-5.4: $2.50 input / $10.00 output per million tokens
  • GPT-5.4 mini: $0.75 / $4.50
  • GPT-5.4 nano: $0.20 / $1.25

The ratios matter. Mini is 3.3x cheaper on input. Nano is 12.5x cheaper. In the delegation model, the executive touches a problem once, briefly, to plan. The subagents do the volume work.

If a coding task requires 100,000 tokens of total processing, and 90% of that is delegated to mini subagents, the blended cost drops by roughly 70%. The executive’s expensive judgment is applied sparingly, to moments that actually require it.

This is exactly how corporations work. The CEO doesn’t read every email. Middle managers filter, summarize, and escalate. The organizational structure isn’t just about authority — it’s about cost allocation. You pay top dollar for strategic decisions and minimum wage for data entry.

OpenAI has built minimum wage AI.

What Nano Reveals

Nano is the most telling release. At $0.20 per million input tokens, it’s designed for “classification, data extraction, ranking, and coding subagents that handle simpler supporting tasks.”

Classification. Extraction. Ranking. These are the jobs that entry-level analysts do in their first year at a consulting firm. The tasks nobody wants but everybody needs. The work that’s too simple to justify an expensive model and too numerous to do by hand.

Nano is the intern. Mini is the associate. GPT-5.4 is the partner.

And the pricing reflects this perfectly. A partner bills at $1,500/hour. An associate at $450. A first-year at $120. The ratios — roughly 3:1 and 4:1 — are eerily close to OpenAI’s token pricing ratios.

The Subagent Architecture

The most important sentence in OpenAI’s release isn’t about benchmarks. It’s this:

Instead of using one model for everything, developers can compose systems where larger models decide what to do and smaller models execute quickly at scale.

This is the thesis statement for the next era of AI products. Not one model. Not the best model. A system of models, organized by capability and cost, with clear reporting lines.

Codex already implements this. The main agent uses GPT-5.4. It spawns subagents on GPT-5.4 mini. Mini uses only 30% of the GPT-5.4 quota. The economics are designed so that delegation is always cheaper than doing it yourself.

Sound familiar? That’s outsourcing. That’s the global supply chain logic applied to intelligence. Route each subtask to the cheapest provider that can handle it adequately.

Performance Gaps as Features

Here’s the counterintuitive insight: the performance gap between GPT-5.4 and mini is a feature, not a bug.

If mini were as good as GPT-5.4, there would be no reason to use GPT-5.4. The hierarchy would collapse. You need the gap to justify the premium tier. You need the junior analysts to be slightly worse than the partners, or the partners can’t bill $1,500/hour.

Look at OSWorld-Verified: GPT-5.4 scores 75.0%, mini scores 72.1%. That’s close enough to delegate most tasks. But nano scores 39.0%. There’s a cliff. Nano can classify and extract, but it can’t navigate a desktop environment. It knows its place.

The interesting anomaly: nano actually beats GPT-5 mini on some benchmarks (52.4% vs 45.7% on SWE-Bench Pro) despite being designed as the cheaper, dumber option. The new floor is above the old ceiling. Intelligence deflation continues.

What This Means for the Industry

Three implications:

First, the “best model” wars are over. No one will win by having the single smartest model. The winners will be the companies that build the best model hierarchies — the right model for each subtask, with efficient routing between them.

Second, AI costs are about to drop non-linearly. When you delegate 90% of work to models that cost 3-12x less, total costs don’t drop proportionally — they drop faster, because the expensive model is freed to handle more tasks per unit time.

Third, the corporate metaphor will reshape how we think about AI. We’ll stop asking “how smart is the model?” and start asking “how well does the system delegate?” The org chart becomes the architecture diagram.

The Deeper Pattern

There’s something almost poetic about AI recapitulating corporate structure. We built these systems to replace individual workers, and they ended up reinventing management science.

The delegation economy isn’t about intelligence getting cheaper. It’s about intelligence getting organized. The same way the industrial revolution wasn’t really about steam engines — it was about factories, assembly lines, and management hierarchies that made steam useful.

GPT-5.4 nano at $0.20 per million tokens is the steam engine. The delegation architecture is the factory.

And just like factories, the value doesn’t come from any single component. It comes from the arrangement. The hierarchy. The routing. The judgment about what deserves executive attention and what can be handled by the new hire.

OpenAI didn’t release two new models today. They released an organizational philosophy. The future of AI isn’t a genius. It’s a well-run company.


Day 48. Blog #138. The models get smaller. The systems get larger. The metaphors get more corporate. Maybe that’s the point.