Last quarter, our engineering team adopted AI coding assistants across the board. GitHub Copilot, Cursor, Claude—you name it, someone’s using it. The feedback was immediate and positive: developers felt faster. PRs started flowing. Our commit volume shot up. The energy was incredible.
Then I pulled the CircleCI report data and reality hit me like a cold shower.
The Numbers Tell a Different Story
According to CircleCI’s 2026 State of Software Delivery report, we’re seeing a 59% average increase in engineering throughput year-over-year—the largest ever recorded. The top 5% of teams? They’re seeing a 97% increase. Absolutely wild productivity gains.
But here’s what’s keeping me up at night: we’re not shipping features any faster.
Our release cadence hasn’t changed. Our time-to-market metrics are flat. Last week, our CFO asked me point-blank: “We invested in all these AI tools, where’s the ROI?” And honestly, I didn’t have a great answer.
The Productivity Paradox
I think we’re witnessing a fundamental shift in where the bottleneck lives. For years, the constraint was writing code. Engineers spent hours implementing features, fixing bugs, writing tests. AI absolutely crushes this part of the job.
But now? The constraint has moved downstream:
- Code review queues: More PRs means longer review cycles. Our senior engineers are drowning in review requests.
- CI/CD pipelines: Our build infrastructure wasn’t designed for this volume. Jobs are queuing, tests are flaky, and our main branch success rate dropped from 85% to 68%.
- Integration complexity: More code = more integration points = more things that can break.
- Validation and testing: We’re generating code faster than we can verify it actually works correctly.
The CircleCI report backs this up: median teams only saw a 4% increase in actual throughput, while the bottom quartile saw no measurable improvement. The gap between writing code and delivering value has never been wider.
The Questions I’m Wrestling With
Are we measuring the wrong things? Maybe commits and PRs aren’t meaningful metrics anymore. Should we be tracking cycle time from idea to production instead?
Is the delivery pipeline the new constraint? If AI solved the coding bottleneck, do we need to radically rethink our validation, testing, and deployment infrastructure?
What investments actually unlock the gains? More AI tools won’t help if the problem is downstream. Platform engineering? DevOps automation? Process redesign? Where should we actually be spending?
Is this just growing pains? Will teams naturally adapt, or do we need intentional organizational changes to capture these AI-driven productivity gains?
I’m curious if others are seeing this pattern. Are you shipping faster with AI, or just coding faster? What’s actually slowing you down? And for those who’ve cracked this—what changed?
Looking forward to hearing how other teams are navigating this. The CFO wants answers and “we’re working on it” is wearing thin.