Analysis

Why AI Coding Assistants Haven't Sped Up Delivery (And What Actually Does)

Published: April 2026 | 11 min read

Two years into the great AI coding experiment, the data is in — and it's more complicated than the marketing claims. Despite GitHub Copilot, Claude Code, Cursor, and their competitors being adopted by millions of developers, a growing body of research suggests that AI coding tools have not, on average, meaningfully improved software delivery velocity. Understanding why is reshaping how engineering leaders think about productivity.

The Surprising Finding

QCon London 2026 opened with a presentation that went viral in engineering circles: "AI Coding Assistants Haven't Sped Up Delivery Because Coding Was Never the Bottleneck." The talk synthesized data from over 200 engineering teams across a variety of company sizes and industries, measuring delivery metrics before and after AI tool adoption.

The headline findings:

Perhaps most striking: the teams that saw the most improvement from AI coding tools were not the ones using them most intensively. They were the ones that had already solved other bottlenecks.

Why Coding Isn't the Bottleneck

The finding makes more sense when you examine where software delivery time actually goes. For a typical feature request, the coding phase — the actual writing of code — represents a minority of total elapsed time. Here's where time actually disappears:

1. Requirements and Specifications (20-30%)

Before a line of code is written, someone needs to decide what to build. Requirements gathering, specification writing, stakeholder alignment, and scope negotiation consume significant time. AI coding tools are excellent at generating code — but they're not good at resolving ambiguous requirements that human stakeholders haven't yet clarified.

When AI accelerates coding, it often just surfaces the requirements bottleneck faster. Teams finish coding quicker only to realize they need more clarification — creating a backlog at the requirements stage.

2. Code Review and Testing (15-25%)

AI generates code quickly, but that code still requires human review, testing, and validation. In many organizations, review and testing bottlenecks have actually worsened after AI tool adoption, because:

3. Integration and Coordination (15-20%)

Modern software development is a coordination-intensive activity. Pull requests need approval. Features need to be integrated with other teams' work. Environments need to be provisioned. CI/CD pipelines need to pass. These coordination activities are largely unaffected by AI coding tools.

In some cases, AI coding acceleration has made coordination harder — if one team writes code 2x faster but the integration and review pipeline hasn't changed, you get longer wait times at those handoff points.

4. Deployment and Infrastructure (10-15%)

Getting code to production — infrastructure provisioning, environment setup, deployment pipelines, monitoring configuration — is often a separate bottleneck that AI coding tools don't directly address. Some tools can help generate infrastructure-as-code, but operational concerns remain human-led.

5. The Coding Phase Itself (20-30%)

Here's the counterintuitive part: for many teams, the coding phase is not even the majority of the time. And within coding, the time breakdown looks like this:

What the High-Performers Did Differently

The QCon research identified a subset of teams (~15%) that did see meaningful delivery improvements from AI tools. Their common characteristics:

They Addressed Bottlenecks in Order

High-performing teams didn't just add AI tools — they used AI-generated time savings to reveal and then fix upstream bottlenecks. When AI accelerated coding, they invested that gained time in faster requirements clarification, more efficient review processes, and better coordination. They treated AI tools as part of a broader system improvement, not a standalone solution.

They Used AI Selectively, Not Universally

Interestingly, the teams that saw the most improvement were not using AI for every task. They had developed heuristics for when AI assistance added value vs. when it added noise. For example:

They Maintained Code Review Rigor

Teams that treated AI-generated code with more scrutiny, not less, saw better outcomes. They developed review checklists specific to AI-generated code — looking for common failure modes like plausible-but-wrong function calls, missing edge cases in generated tests, and overly generic implementations that work but aren't optimal.

They Measured the Right Things

High-performing teams tracked metrics beyond simple velocity. They measured code quality trends, bug rates by source (AI-generated vs. human-written), time spent in review states, and lead time broken down by phase. This visibility allowed them to identify where AI was actually helping and where it was creating new problems.

The Code Quality Paradox

One of the most counterintuitive findings: AI coding tools have, in aggregate, increased the volume of code written without proportionally improving its quality or reducing its long-term maintenance burden. This creates what researchers are calling the "code quality paradox."

The paradox works like this:

  1. AI makes it cheap and fast to write code
  2. Teams therefore write more code for any given feature
  3. More code means more surface area for bugs and more code to maintain
  4. The perceived productivity gain from faster coding is partially offset by higher maintenance costs downstream

The data bears this out: several organizations reported that AI coding tools led to a net increase in technical debt, as teams took the path of writing more code rather than cleaner code when AI made "just write another helper function" the path of least resistance.

What Actually Speeds Up Delivery

Based on the research and case studies, here are the interventions that consistently show up in teams that have meaningfully improved delivery:

1. Address Requirements Bottlenecks

Invest in better upfront specification: clearer user stories, explicit acceptance criteria, and pre-mortems on ambiguous requirements. When requirements are clear, AI coding tools are far more effective — because the ambiguity that would normally surface mid-coding is resolved before the AI generates code.

2. Parallelize Review, Not Just Coding

If coding is faster but review is serial, you get longer queues. High-performing teams invest in making reviews parallelizable: smaller PRs, clearer review guidelines, automated pre-checks that reduce reviewer burden, and explicit review time SLAs.

3. Invest in Test Infrastructure

The teams seeing the most value from AI coding tools are those that have invested heavily in test infrastructure — fast unit tests, good integration test coverage, and CI pipelines that give fast feedback. AI generates code fast; the only way to validate that code faster is fast tests. Without fast, reliable tests, AI-generated code creates a verification bottleneck.

4. Use AI for Code Improvement, Not Just Code Generation

The teams getting the best ROI are using AI for refactoring, dead code elimination, and test coverage improvement — not just feature coding. AI is often better at improving existing code than writing new code, because the context is richer and the "correct" answer is more constrained.

5. Reduce Handoff Overhead

For organizations where the biggest delays are in coordination and handoffs between teams, process improvements matter more than coding tools. Cross-functional team structures, reduced approval layers, and clear ownership reduce the non-coding time that no tool can eliminate.

The Honest Assessment

AI coding tools are genuinely useful — but the industry oversold them as a solution to software delivery productivity. The reality is more nuanced: AI coding tools excel at specific tasks within the coding phase, but the coding phase is not where most delivery time goes.

This doesn't mean AI coding tools are not worth using. It means the productivity gains come from being thoughtful about where you apply them, not from assuming they will accelerate your entire delivery pipeline by default.

The teams that will see the most benefit from AI coding tools in 2026 are those that use them as part of a broader strategy to identify and address delivery bottlenecks — starting with the biggest ones, not the most convenient ones.

Affiliate Link: GitHub Copilot | Claude Code | Cursor

Affiliate Disclosure: This page contains affiliate links. If you purchase through our links, we may earn a commission at no extra cost to you.