LinearB just published their 2026 Software Engineering Benchmarks Report, and the numbers are striking. They analyzed 8.1 million pull requests from 4,800 engineering teams across 42 countries - the largest empirical dataset of developer productivity ever published.
Here’s what elite looks like in 2026:
The New Elite Thresholds
| Metric |
Elite |
Median |
Red Flag |
| Cycle Time |
<48 hours |
83 hours |
124+ hours |
| PR Size |
<105 lines |
~200 lines |
500+ lines |
| Daily Focus Time |
6+ hours |
4.2 hours |
<3 hours |
| Merged PRs/Month |
15+ |
12.4 |
<8 |
| Lead Time |
<24 hours |
3.8 days |
7+ days |
Why This Matters for Engineering Leaders
I’ve been tracking our team’s metrics against these benchmarks, and it’s been eye-opening. A few observations:
1. Small PRs Are Non-Negotiable
Elite teams keep PRs under 105 lines. This isn’t arbitrary - smaller PRs mean faster reviews, fewer merge conflicts, and easier debugging when issues arise. Yet most teams I talk to struggle with engineers who ship 500+ line PRs as “normal.”
2. Protected Focus Time Separates Elite from Average
6+ hours of protected deep work daily for elite teams vs 4.2 hours for median. That’s 9 extra hours per week of uninterrupted coding. Over a year, that’s 468 hours of additional focused development time per engineer.
3. Cycle Time Is the Leading Indicator
Sub-48 hour cycle time correlates strongly with all other positive metrics. If your cycle time is trending up, everything else is probably suffering too.
The Hard Questions
- How does your team compare to these benchmarks?
- What’s blocking you from reaching elite thresholds?
- Is anyone actually measuring these metrics systematically?
I’ll share our journey from ~120 hour cycle times to ~60 hours in the comments - but I’m curious what patterns others are seeing.
Sources: LinearB 2026 Benchmarks Report, ByteIota Analysis
Keisha, these benchmarks hit close to home. Let me share our actual numbers and what moved the needle.
Where We Started (Q1 2025)
- Cycle time: 127 hours (yikes - deep in red flag territory)
- Average PR size: 340 lines
- Focus time: ~3.5 hours daily
- Lead time: 6.2 days
Where We Are Now (Q4 2025)
- Cycle time: 68 hours (still not elite, but solid progress)
- Average PR size: 145 lines
- Focus time: ~5.1 hours daily
- Lead time: 2.8 days
What Actually Worked
-
Mandatory PR size limits - We set a hard ceiling at 200 lines with automated checks. Exceptions require manager approval and justification. Initially engineers pushed back, but within a month it became natural.
-
No-meeting mornings - Every engineer has protected focus time from 8am-12pm, no exceptions. This alone improved focus time by 1.5 hours daily.
-
Review SLAs - PRs must be picked up within 4 hours during work hours. We track this and discuss it in retros.
-
Smaller batches - Instead of one big feature PR, we ship in 3-4 incremental PRs. Takes discipline but cycle time improved dramatically.
What We’re Still Struggling With
- Getting to sub-48 hour cycle time feels like it requires architectural changes, not just process changes
- Financial services compliance reviews add 8-12 hours to every PR touching sensitive systems
- Some senior engineers still resist small PRs - “it’s artificial chunking”
The Financial Services Reality
In regulated industries, some of these benchmarks feel aspirational. Our compliance requirements add overhead that pure tech companies don’t face. But that’s not an excuse - it’s a constraint to optimize around.
What’s worked for others in regulated environments?
Love seeing this data, Keisha. Let me put on my statistician hat and add some context.
On the Methodology
8.1 million PRs is an impressive sample size, but it’s worth noting:
- Self-selection bias: Companies using LinearB’s tooling may already be more metrics-conscious
- Industry distribution matters: Are these mostly tech companies? Enterprises? The median will differ dramatically
- Time window: Were these PRs from one quarter or multiple years? Seasonality affects these metrics
This doesn’t invalidate the benchmarks - but it means you should treat them as directional, not gospel.
The Statistical Reality of “Elite”
By definition, “elite” means top performers. If everyone achieved sub-48 hour cycle time, the benchmark would just move. This is important because:
-
Diminishing returns - Getting from 120 hours to 70 hours is much easier than getting from 70 to 48. The effort required grows exponentially as you approach elite thresholds.
-
Context matters - A 48-hour cycle time for a 5-person startup is different than for a 500-person enterprise. Team size, codebase complexity, and domain all influence achievable benchmarks.
-
Confidence intervals - I’d want to see the variance, not just the median. Is the median 83 hours with a tight distribution, or are there massive outliers skewing it?
What I Actually Look For
When analyzing our team’s metrics:
- Trend direction matters more than absolute numbers
- Variance within the team is often more actionable than the average
- Correlation analysis - which metrics actually predict outcomes we care about (incident rate, customer satisfaction, retention)?
One Concern
These benchmarks can create perverse incentives. I’ve seen teams game cycle time by:
- Opening PRs earlier (before code is really ready)
- Approving without thorough review to hit SLAs
- Avoiding complex changes that would take longer
Measure carefully, act thoughtfully.
This thread is excellent. Let me add the executive perspective on how these benchmarks translate to strategic decisions.
How I Use Benchmarks With My Board
When presenting to the board, raw metrics like “cycle time” don’t resonate. What does resonate:
-
Time to value - “We ship customer-facing improvements 40% faster than last year. That means we can respond to competitive threats in 2 days, not 5.”
-
Cost of delay - “Every day of delay on Feature X costs us $50K in potential revenue. Our cycle time improvements mean we capture $750K more annually.”
-
Engineering leverage - “Our 30-person team ships at the velocity of a 45-person team at industry benchmarks. That’s $2.5M in avoided headcount.”
The Strategic Questions These Benchmarks Enable
- Are we investing in the right areas? (Platform engineering, DevEx)
- How much engineering capacity do we really need for our roadmap?
- Where’s the bottleneck - technical debt, process, or people?
- Can we confidently commit to customer timelines?
What I Watch Closely
| Metric |
Why It Matters Strategically |
| Cycle time trend |
Leading indicator of organizational health |
| PR size distribution |
Tells me if we’re doing sustainable engineering |
| Focus time |
Proxy for meeting culture and process overhead |
| Lead time variance |
Predictability for planning |
A Warning About Executive Dashboards
I’ve seen engineering leaders present beautiful metrics to executives, only to have those metrics weaponized later. “You said cycle time was 48 hours, why can’t you ship this feature by Monday?”
Educate your executives on what these metrics mean and their limitations. Rachel’s point about gaming is real - make sure leadership understands the tradeoffs.
My Ask
If you’re an engineering leader: How do you communicate these metrics to non-technical stakeholders? What’s worked, what’s backfired?