Building on the productivity measurement discussion, I want to dig into something Luis and Michelle both mentioned: pipeline bottlenecks.
The 5% Problem
Research shows that AI coding assistants speed up the coding portion of software delivery—which represents roughly 5-10% of the total pipeline from idea to production (Axify analysis).
The other 90-95%?
- Requirements clarification and design
- Code review
- Testing and QA
- Integration and deployment
- Documentation and knowledge transfer
We optimized 5% of the pipeline and wondered why the whole system didn’t speed up.
This is basic systems thinking: A pipeline is only as fast as its slowest step. Making the fastest step even faster doesn’t help if everything else is unchanged.
Where I’m Seeing the Backlog Build
In our design systems work, we’re now seeing:
Before AI:
- Dev time: 40% of cycle time
- Review time: 25%
- Testing: 20%
- Deployment: 15%
After AI:
- Dev time: 25% of cycle time (improved!)
- Review time: 35% (bottleneck shifted here)
- Testing: 25% (longer queue because more code)
- Deployment: 15% (unchanged)
Result: Overall cycle time only improved 8%, even though coding time improved 37%.
The faster coding created a backlog in review and testing that we weren’t prepared for.
The Uncomfortable Question
Should we:
A) Speed up the rest of the pipeline to match AI coding velocity?
- Invest in automated code review tools
- Scale QA team and test automation
- Improve CI/CD infrastructure for faster deployments
B) Slow down coding to match downstream capacity?
- Gate AI usage to prevent overwhelming review/test
- Maintain current process and accept modest gains
- Let natural equilibrium emerge
C) Rethink the entire pipeline for the AI era?
- Merge coding and review into single step (pair programming with AI?)
- Shift testing left (AI generates tests alongside code)
- Automate deployment gates that are currently manual
I honestly don’t know which is right.
The Systems View
Michelle’s original thread asked if we’re measuring wrong things. But maybe we also need to ask: Are we optimizing the wrong part of the system?
If review and testing are the real bottlenecks, shouldn’t our AI investment focus there instead of making coding (which wasn’t the bottleneck) even faster?
What I’m Curious About
- Where are bottlenecks appearing in your delivery pipelines post-AI?
- Are you investing to eliminate bottlenecks, or accepting that AI gains will be modest until you do?
- What’s working in terms of pipeline reoptimization for the AI era?
From a design perspective, I suspect quality gates are where AI gains will disappear unless we fundamentally rethink how we ensure quality at higher code volumes.
Thoughts?