
The Bain Technology Report 2025 reveals why AI coding tools deliver only 10-15% productivity gains. Learn why companies aren't seeing ROI and how to fix it with lifecycle-wide transformation.

The promise was enticing: deploy AI coding assistants, watch productivity soar, and transform software development overnight. Yet nearly two years into the generative AI revolution, most companies are asking the same question: "Where's the payoff?"
Two important reports, Bain & Company's 2025 Technology Report and Faros AI's AI Productivity Paradox Report, reveal why AI gains have stalled and, more importantly, what separates companies capturing real value from those stuck in pilot purgatory.
The Bain Technology Report 2025 found that while two-thirds of software firms have rolled out generative AI tools, teams see only 10-15% productivity boosts. And the time saved rarely translates into business value. Most leaders are still asking, "Where's the payoff?"
Faros's AI Engineering Report 2026 investigates this phenomenon with two years of telemetry from 22,000 developers across more than 4,000 teams. The picture has evolved. Organizational throughput gains are now real: epics completed per developer are up 66%, and AI is finally moving roadmaps, not just individual task counts. The gains Bain identified as stalling are beginning to arrive.
But they are arriving alongside something the productivity dashboards are not showing. For every pull request merged, the probability of a production incident has more than tripled. Bugs per developer are up 54%. 31% more code is reaching production with no review at all. Individual developers are working faster and organizations are shipping more. The question is whether what they are shipping is ready. We call this the Acceleration Whiplash.
{{whiplash}}
The Bain Technology Report 2025 reveals that writing and testing code accounts for only 25-35% of the time from initial idea to product launch. Speeding up coding while leaving requirements gathering, planning, deployment, and maintenance unchanged creates a bigger bottleneck, not a faster pipeline.
Faros's telemetry makes this bottleneck visible and measurable. While tasks involving code specifically rose 210% at the team level, median time in PR review is up 441%, and 31% more PRs are merging with no review at all.
Consider what we're observing in real development teams: A developer using AI blasts through three tickets before lunch. But those PRs now sit in the review queue for days because the reviewers are underwater. The testing pipeline, built for a slower cadence, starts failing. The deployment process can't keep up with daily merges.
This is Amdahl's Law in action: A system's speed is determined by its slowest component. AI just exposed where those slowest components really are.
The Bain Technology Report 2025 emphasizes that "speeding up these (coding) steps does little to reduce time to market if others remain bottlenecked." Faros AI's data shows exactly how severe this effect is.
Companies investing millions in AI coding tools see their organizational DORA metrics stay stubbornly flat, not because AI doesn't work, but because the rest of the development lifecycle hasn't evolved to absorb the acceleration.
The bottleneck problem compounds when you look at code quality. Faros's 2026 analysis found that high AI adoption correlates with a 51.3% increase in average pull request size, a 54% increase in bugs per developer, up from just 9% in our prior dataset, and an 861% increase in code churn—the ratio of lines deleted to lines added in a given quarter—means significantly more code is being removed relative to what is being added.
AI-generated code tends to be more verbose, less incremental, and harder to review, placing even greater burden on already-overwhelmed review processes.
Faros AI also observed fragmented adoption patterns that prevent gains from scaling:
When Bain asked companies what the hardest part of AI adoption was, three out of four pointed to the same challenge: Getting people to change how they work.
Our conversations with CTOs at Fortune 1000 companies reveal why:
{{cta}}
The Bain Technology Report 2025's central argument is that "real value comes from applying generative AI across the entire software development life cycle, not just coding." The report gives two examples of how leading companies like Netflix and Goldman Sachs prove this works.
Netflix implemented "shift left" approaches to ensure rapidly generated code isn't stuck waiting on slow tests.
Goldman Sachs integrated AI into its internal development platform and fine-tuned it on the bank's codebase, extending benefits from autocomplete to automated testing and code generation.
These companies didn't just add AI to existing workflows, they rebuilt workflows around AI:
The Bain Technology Report 2025 shows these organizations are achieving 25-30% productivity gains, far above the 10% from basic code assistants, because they addressed the entire lifecycle, not just coding.
Faros AI's research reveals three operational characteristics that separate winners from those stuck in pilot mode:
Based on this research, Faros AI has developed the GAINS™ framework (Generative AI Impact Net Score) to help organizations realize AI's potential. The Bain Technology Report 2025 refers to this as operationalizing the "AI-native reinvention of the software development life cycle."
GAINS™ leverages live telemetry across your SDLC to ten dimensions that actually move outcomes, including adoption and usage, velocity and flow efficiency, quality and safety, onboarding and platform maturity, organizational structure and strategic alignment. It locates the constraint, recommends how to treat the cause, and proves the result with numbers engineering and finance both trust.
When teams use GAINS™, the conversation changes. Instead of arguing about the value and the limitations of AI tools, you can see where adoption is strong and where it is weak, whether review wait time (not coding time) is dominating lead time, that test instability (not developer speed) is driving long development cycles.
This type of clarity makes the Bain Technology Report 2025's playbook actionable. You modernize where it matters, enable people by role and services, put guardrails around AI-authored code, and track how the time you reclaimed shows up as business value on the next quarterly scorecard.
As the Bain Technology Report 2025 notes, when organizations pair generative AI with end-to-end transformation, productivity gains don't just look bigger. They are bigger, routinely in the twenty-five to thirty percent range, and increasingly durable.
{{cta}}
The urgency just went up. The Bain Technology Report 2025 emphasizes that "an even bigger leap is on the horizon as AI evolves from assistant to autonomous agent—a shift that could redefine software development and widen the gap" between leaders and laggards.
If a 210% increase in code-related tasks and a 441% surge in median review time reflects what happens when humans use AI as a primary authoring tool, what happens when autonomous agents submit PRs independently with no human initiating each change?
Within 12 months, agentic AI will require centralized control planes providing visibility and governance across human and agent workflows. The gap is widening now.
If you recognize your organization in that opening scene with busy engineers, crowded Jira boards, flat business metrics—you don't have an AI problem. You have a system problem.
The good news is that systems can be changed.
The Bain Technology Report 2025 offers the macro lens; Faros AI data shows you exactly where to start.
And do it now, while the gap between operators and dabblers is still crossable.
The companies that move decisively today, instrumenting the lifecycle, modernizing the platform, enabling their people, and governing for speed and safety, are already seeing the difference between coding faster and delivering faster. The rest will keep adding code to queues.
{{cta}}
You're experiencing what both the Bain Technology Report 2025 and Faros AI's data confirm: Individual velocity increases don't automatically translate to organizational gains. Developers may be coding faster and merging more PRs, but if review time increases and your testing/deployment pipelines can't keep pace, the gains get absorbed by downstream bottlenecks. The system's speed is determined by its slowest component. AI just exposed where your constraints really are.
The biggest mistake is treating AI as a point solution rather than a lifecycle transformation. Companies deploy tools but don't redesign the workflows around them. They speed up coding while leaving review, testing, and deployment processes unchanged. As the Bain Technology Report 2025 found, writing code is only 25-35% of the development lifecycle—accelerating that one piece without addressing the other 65-75% creates bottlenecks, not breakthroughs.
Yes, our data shows review time increases by 91% on average when PR volume surges. This happens because AI-generated code tends to be more verbose (154% larger PRs on average) and contain more bugs and security vulnerabilities, making reviews more complex. You need to redesign your review process: implement smaller PR batching, update routing to handle volume, shift quality checks earlier, and potentially expand review capacity.
Our data shows several patterns: less tenured engineers adopt more aggressively (they need AI to navigate unfamiliar codebases), while senior engineers remain skeptical (they handle complex, context-dependent work where current AI struggles). Most developers only use autocomplete while advanced features remain untapped. Without centralized enablement, role-specific training, and shared playbooks, adoption stays fragmented and surface-level.
GAINS™ (Generative AI Impact Net Score) is a diagnostic that evaluates ten dimensions critical to AI transformation: adoption patterns, velocity, flow efficiency, quality, safety, developer satisfaction, onboarding, platform maturity, organizational structure, and strategic alignment. It uses live telemetry across your SDLC to identify where constraints actually are - whether that's review wait time, test instability, or inadequate enablement - so you can treat root causes rather than symptoms.
Current AI coding assistants are copilots, meaning they suggest code step-by-step with heavy human involvement. Agentic AI will autonomously reason, plan, and execute multi-step tasks across the SDLC with minimal human intervention. If a 210% increase in code-related tasks and a 441% surge in median review time reflects what happens when humans use AI as a primary authoring tool, what happens when autonomous agents submit PRs independently with no human initiating each change? Organizations that haven't addressed foundational gaps in review capacity, testing infrastructure, deployment pipelines, and governance structures will be overwhelmed within 12 months.
The Bain Technology Report 2025 found that three out of four companies say this is the hardest part. Success requires: (1) Clear strategic direction from leadership on where and how AI should be used, (2) Role-specific training; new grads need different skills than senior architects, (3) Internal playbooks and communities of practice to share what works, (4) Visible celebration of wins to build momentum, (5) Making AI competency part of the job, not optional. Grassroots enthusiasm without structure leads to shallow adoption and low ROI.
Nothing's wrong with your tools, but your system hasn't adapted. Our data shows deployment frequency improves slightly, but lead time actually increases (driven by longer reviews), while change failure rate and MTTR stay flat. The 2025 DORA Report shows AI amplifies team dysfunction as often as capability.
DORA metrics won't improve until you address the full lifecycle: Review bottlenecks, testing constraints, deployment automation, and how you redeploy saved capacity. This is why lifecycle-wide transformation is essential.
It depends on your approach. Companies treating AI as a tool see minimal gains that plateau quickly. Companies treating it as a transformation by modernizing platforms, redesigning workflows, providing structured enablement, and measuring continuously are starting to see measurable improvements within 2-3 quarters and gains that compound over time. The key is not waiting for perfect conditions but moving forward with discipline: Test, learn, adapt.
Start with the GAINS™ diagnostic to identify your specific constraints. Don't assume you know where the problems are. Our data shows the bottlenecks are rarely where organizations think they are. Once you know whether your constraint is review capacity, test infrastructure, inadequate enablement, or strategic misalignment, you can focus investment where it actually matters. Most organizations spread resources too thin trying to fix everything. High performers identify the constraint and treat that first.
{{cta}}




Tokenmaxxing—treating AI token consumption as a productivity metric—is repeating the lines-of-code mistake. Data from 22,000 developers points to a better way to measure AI engineering impact.