I’ve been using AI coding assistants for the past year, and something’s been bugging me. ![]()
My fingers fly across the keyboard now. AI autocompletes entire functions. I feel fast—like I’m coding at warp speed. But here’s the thing: our sprint velocity hasn’t budged. We’re still shipping features at the same pace we did before Claude Code, Copilot, and all the other AI tools became part of our workflow.
So I started digging into the data, and wow—the numbers tell a wild story.
The Productivity Paradox
Research from 2026 shows that developers report being 30% faster at writing code with AI assistants. That’s massive! But when you measure actual delivery velocity—how fast teams ship features to production—the improvement is only about 8%.
Where did the other 22% go? ![]()
Even wilder: a randomized trial by METR found that developers using AI were actually 19% slower on average, yet they were convinced they’d been faster. Before the experiment, they predicted AI would make them 24% faster. After finishing (slower!), they still believed AI had sped them up by ~20%.
We’re living in a perception bubble.
The Bottleneck Problem
Here’s what I think is happening, and I’m seeing this on my own team:
AI speeds up code generation, but everything downstream is drowning.
- Code review queues are backing up (PRs are 154% larger on average now!)
- Testing takes longer because there’s more code to test
- Bug rates went up 9% with AI-generated code
- Integration and deployment processes weren’t designed for this volume
It’s like we upgraded one machine on the assembly line to super-speed, but forgot about all the other machines. Now we’ve got a massive pile-up at code review. ![]()
The bottleneck moved—it didn’t disappear.
My Design Perspective
As someone who came from design into product building, this reminds me of a classic UX mistake: optimizing one part of the user journey while ignoring the end-to-end experience.
Faster coding is great! But if we’re still waiting days for code review, or if we’re shipping buggy code that needs hot fixes, or if we’re building features that don’t move business metrics—what did we actually optimize?
The research shows that top-performing organizations (the ones who’ve adapted their full SDLC) are seeing 20-60% productivity gains. But most companies? Still stuck at 5-10% because they bought AI tools but didn’t upgrade their review, testing, and integration processes.
Questions for the Community
I’m curious what you’re all seeing:
-
Where’s your bottleneck? Is it code review? Testing? Requirements? Something else?
-
Have you adapted your processes for the AI era, or are you still using pre-AI workflows with AI-speed code generation?
-
What metrics are you tracking? Are you measuring coding speed, delivery speed, or business outcomes?
-
Are we building faster, or just building more? Is the extra code actually valuable?
I have a sneaking suspicion we’re optimizing the wrong part of the stack. Would love to hear if others are seeing the same patterns—or if you’ve cracked the code on actually translating AI coding speed into real delivery velocity. ![]()
Sources for the nerds: