I’ve been tracking our engineering metrics closely since we rolled out AI coding assistants last year, and the data is both exciting and frustrating. CircleCI’s 2026 State of Software Delivery report shows a 59% increase in average engineering throughput across 28 million data points. That’s massive. But here’s the uncomfortable truth: most organizations—including mine—are leaving the majority of those productivity gains on the table.
The Throughput Paradox
Our team’s experience mirrors what the research shows. We saw a 15.2% increase in throughput on feature branches—developers are clearly moving faster. They’re experimenting more, iterating quicker, and shipping more code to review. But when I looked at main branch throughput, we actually declined 6.8%.
That disconnect was a wake-up call.
The problem isn’t the AI tools. The problem is that everything downstream of code generation—pull request reviews, QA validation, security scanning, deployment approvals—was built for a different velocity. When coding accelerates, pull request volume increases, review queues grow, QA becomes saturated, and security validation lags. The entire delivery system needs to adapt, and most of us haven’t done that work yet.
The Measurement Blind Spot
According to Waydev’s 2026 analysis, the strategic question for CTOs and VPs of Engineering isn’t whether to adopt AI, but how to build the organizational visibility required to extract AI’s full value across the entire delivery cycle. The bottleneck has shifted, and most leaders haven’t updated their dashboards to reflect it.
Four metrics belong on every engineering leader’s dashboard right now:
-
Main branch success rate - This is the clearest signal of whether your delivery system is keeping pace with AI-generated volume. The industry benchmark is 90%. The current average is 70.8%. We’re at 73%, which tells me we have work to do.
-
PR review time - Research shows PR review time increased 91% even as coding speed improved. That’s your bottleneck screaming at you.
-
Mean time to recovery (MTTR) - For teams running AI-assisted workflows, this is where productivity gains either hold or disappear.
-
Deployment frequency - More code should mean more value delivered to customers. If your deployment cadence hasn’t changed, you’re not capturing the gains.
What Actually Works
Companies that are successfully capturing AI productivity gains aren’t just tracking different metrics—they’re fundamentally rethinking their delivery systems:
-
Dropbox tracks daily and weekly active AI users, AI tool satisfaction, time saved per engineer, and spend. They’re connecting AI adoption to business outcomes, not just developer satisfaction.
-
Leading teams are investing in hardened security templates, automated validation, and pre-configured guardrails so that security doesn’t become the bottleneck.
-
The most effective organizations are strengthening review practices and retraining teams to understand that AI-generated code still requires the same rigor as human-written code.
The Uncomfortable Questions
Here’s what I’m wrestling with:
-
Are we measuring the right things? Traditional metrics like story points and velocity don’t capture where AI creates value or where delivery systems break down.
-
Have we adapted our processes? If developers can write code 30% faster but we’re still shipping at the same pace, the problem isn’t the developers.
-
What’s the ROI? CFOs are starting to ask hard questions about AI tool spend. Without clear metrics linking AI adoption to delivery outcomes or business results, those budgets are at risk.
-
Are we building the right things faster, or just building faster? Throughput gains only matter if we’re solving the right problems for customers.
What I’m Curious About
I’d love to hear from other engineering leaders about:
- What metrics are you tracking to measure AI’s impact on delivery, not just coding?
- Where have you found the bottlenecks when AI accelerates coding?
- How are you adapting your review, QA, and deployment processes to match the new velocity?
- What does “value capture” actually mean for your organization—is it faster releases, better quality, cost savings, or something else?
The 59% throughput increase is real. The question is whether we’re building the organizational systems to actually realize that value.
Sources: