Your Agents Are Writing Code Faster Than Your Org Can Ship It
AI coding agents can now complete a ticket in minutes that would have taken a developer a day. That is not a metaphor or vendor marketing—SWE-bench Verified scores have crossed 80% and real engineering teams are reporting 98% more pull requests merged per developer. The code is moving. The software is not.
Data collected from over 60,000 developers by Faros.ai in early 2026 tells the uncomfortable story: AI adoption improved throughput on paper while worsening almost every stability and delivery metric that actually matters. PR size grew 154%. Review time climbed 91%. Bugs per developer rose 54%. Incidents per PR jumped 242%. And DORA’s core delivery frequency metric—the one that measures how often you actually ship to production—barely moved.
This is the 100x Agent Illusion: the belief that if you generate code faster, you ship software faster. You don’t. Not without redesigning the system around the new rate of production.
What the DORA Numbers Actually Show
The 2025 DORA State of AI-Assisted Software Development report (Google Cloud, published late 2025) and follow-on analyses for 2026 paint a consistent picture. Individual velocity is clearly up. Organizational throughput is not keeping pace, and quality signals are deteriorating.
Only 9.4% of engineering teams achieve lead times under one hour. A full 43.5% still take more than a week to get a commit from “merged” to “in production.” Those numbers have not meaningfully improved in two years, even as AI tool adoption in those same teams has exceeded 90%.
The reason is structural. Lead time is not measured from “agent starts writing” to “agent finishes writing.” It is measured from the moment code is committed to the moment it runs in production. Every manual review step, every slow CI pipeline, every manual deployment gate, and every organizational approval process sits between those two timestamps. None of those are accelerated by a faster code generator.
vortx.ch covered the DORA 2025 results in detail when they dropped—the headline then was more AI, more code, flatter delivery. Six months later, the pattern has hardened into consensus.
Zendesk Names the Bottleneck: “Absorption Capacity”
In April 2026, Zendesk’s engineering leadership published an analysis via InfoQ that gave the bottleneck a name: absorption capacity. The argument is straightforward. As AI makes code abundant, the constraint in software delivery shifts from writing code to the organization’s ability to define problems clearly, integrate changes coherently, review them safely, and turn implementation into reliable value.
Absorption capacity is not a single metric. It encompasses code review throughput, the clarity of requirements upstream, the health of your test suite, and the maturity of your deployment pipeline. When code generation accelerates without improving any of these, you do not get faster delivery. You get a larger queue.
The pull request is where this manifests most visibly. AI-generated PRs tend to be larger—154% larger on average according to Faros.ai—because an agent will implement an entire feature rather than the careful slice a human reviewer can comfortably absorb in one sitting. Reviewers face denser, harder-to-understand diffs. Review time goes up. Defect rates go up. Lead time holds flat or increases.
96% of developers report that they distrust AI-generated code enough to verify it carefully before approving. That distrust is rational—the model’s confidence is uncorrelated with its correctness at the edges. But it means that every AI-written PR carries a full human review tax, and that tax is now applied to 98% more PRs than existed before.
Amdahl’s Law Is Not Optional
Gene Amdahl formulated his law in 1967 to explain why parallelizing part of a computation does not yield proportional speedups for the whole system. The part you cannot parallelize sets the ceiling. The same logic applies to software delivery.
If code generation is 5% of your total cycle time—from idea to production—then making it 100 times faster reduces total cycle time by roughly 4.8%. The other 95% of your cycle (requirements, review, testing, staging, deployment, monitoring, rollback) is unchanged. This is not a software company problem. It is a physics problem.
The apilama.com analysis of the 100x Agent Illusion makes this concrete: organizations are building systems that generate output at 100x the rate, but reviewing and deploying that output at 3x the rate. The gap accumulates. Code sits in queues. Engineers spend more time context-switching between PRs than actually reviewing them. The agent that wrote the code has moved on to the next ticket. No one owns the integration.
The same dynamic shows up in AI inference infrastructure. A 2026 analysis found that vLLM was losing 57% of theoretical throughput to serialization overhead alone—not model capability, not hardware, just how the stack was assembled. Fixing the system recovered the capacity. The model did not need to improve. This is what engineering the system looks like.
What to Actually Fix
The path forward is not to slow down your agents. It is to redesign the pipeline around the throughput you now have at the code generation stage.
The highest-value interventions cluster around three areas. First, PR discipline: agents should be configured to produce smaller, purpose-bound commits rather than feature-complete megachanges. A 300-line PR that does one thing is reviewed faster, merged more safely, and reverted more easily than a 1,200-line PR that does four things. This is a system configuration choice, not a model capability question.
Second, automated verification must expand to match generation velocity. If your test coverage is at 60% and agents are adding new code paths at 100x the old rate, you are accumulating coverage debt faster than any team can manually close it. Investing in test generation, contract testing, and automated regression tooling is now a first-order engineering priority, not a nice-to-have.
Third, deployment pipelines need to be treated as product, not plumbing. Long staging environments, slow CI queues, and manual production gates are now the most expensive infrastructure your organization runs—because they are the bottleneck that makes your $50,000/year in agent licenses worth almost nothing in delivery terms. The DORA research is unambiguous: elite teams have lead times under an hour because they have invested in this infrastructure, not because they have better models.
As covered in an earlier vortx.ch analysis, agentic engineering in 2026 delivers more code and no better delivery for teams that skip this work. The pattern is consistent enough now to treat it as a rule: raw agent adoption without system redesign is a local optimization with negative global consequences.
The Illusion Is Seductive for a Reason
It feels like progress when code ships from your IDE at speeds that were impossible eighteen months ago. Individual contributors are producing more. Roadmaps are moving. Epics completed per developer are up 66% according to 2026 DORA data. These are real gains and they are worth capturing.
But the 100x Agent Illusion is believing those gains translate automatically into customer value. They don’t—not until the organizational systems that convert code into running software are rebuilt to match the new rate of production. The constraint has moved. The work is to find it, name it, and fix it. That has always been the job of engineering leadership. AI just made it more urgent.
Further Reading
- Zendesk: AI Makes Code Abundant, Shifting the Bottleneck to “Absorption Capacity” — the clearest articulation of why code volume and delivery throughput diverge.
- The AI Productivity Paradox Research Report (Faros.ai) — 60,000-developer dataset with the raw DORA numbers behind the throughput and quality claims in this article.
- DORA Metrics 2026: AI Expansion Meets Visibility Crisis — a practical breakdown of where the metrics are moving and why lead time remains the key signal to watch.

