Executive Summary

  • Ninety-five percent of corporate AI initiatives produce no measurable return on investment, yet enterprise spending continues to accelerate — a pattern that points to a systemic misdiagnosis, not a technology shortfall.

  • Only 6% of organizations qualify as AI "high performers" in McKinsey's 2025 global survey, and their distinguishing trait is not superior models or bigger budgets — it is fundamental workflow redesign.

  • A new category of productivity loss called "workslop" — AI-generated content that shifts cognitive labor onto recipients — costs the average organization with 10,000 employees roughly $9 million per year in invisible rework, according to research from BetterUp and Stanford.

  • Forty-two percent of companies abandoned most of their AI initiatives in 2025, up from 17% the prior year, while employees report that workloads have increased, not decreased, since AI adoption began.

  • The bottleneck is not what AI can do. It is what organizations are willing to reorganize, retrain, and govern to let AI do.

The Setup

Here is a number that should stop a boardroom cold: organizations waste 41% of daily work time on activities that create no value for the enterprise. That finding, from Deloitte's 2025 Global Human Capital Trends survey of nearly 10,000 leaders across 93 countries, predates the current AI spending boom. It describes the organizational baseline into which companies are now pouring generative AI tools — a baseline of duplicated effort, performative busyness, and processes no one has examined in a decade.

The investment itself is not small. Enterprises directed $30 to $40 billion toward generative AI in the past year alone, according to Forbes. Adoption is broad: 88% of employees now report using AI at work, per the EY 2025 Work Reimagined Survey. And yet 95% of those initiatives yield no measurable return, according to research from the MIT Media Lab.

The standard explanation — that the technology is immature — no longer holds. Controlled studies consistently show 15% to 55% task-completion time reductions when AI tools are properly deployed, as documented in the ICLE's empirical review. The tools work. The question is why the organizations deploying them do not.

This article argues that the answer is organizational capacity: the ability to redesign workflows, retrain people, establish governance, and absorb change. Most companies have diagnosed an AI capability problem. What they actually have is a capacity crisis — and no amount of additional spending on models, platforms, or pilots will resolve it.

The Context

The disconnect between technology investment and productivity return is not new. It has a name — the productivity paradox — and it has appeared in every major wave of enterprise technology since the mainframe era.

In 1987, the economist Robert Solow observed that computers were showing up everywhere except in the productivity statistics. The pattern repeated with enterprise resource planning systems in the 1990s and cloud computing in the 2010s. In each case, the lag between adoption and measurable output gains was not months. It was a matter of years, sometimes a full decade, and the unlock was never a better version of the technology. It was an organizational redesign: new workflows, new roles, new management practices built around what the technology made possible.

The AI cycle appears to be following the same script, but at compressed timescales and higher stakes. Generative AI reached 26.4% workplace penetration by the second half of 2024, according to data cited in the Penn Wharton Budget Model — a diffusion rate faster than any prior general-purpose technology. That speed is the problem. Organizations that took five to seven years to adopt cloud computing are now expected to absorb a more disruptive technology in 18 months, often without changing a single reporting line or workflow.

The structural preconditions for this failure were already in place. Workers spent an average of 257 hours annually navigating inefficient processes and another 258 hours on duplicative work and unnecessary meetings — roughly twelve full workweeks per year — according to Asana's research cited in Deloitte's analysis. Time spent in collaborative activities has increased by more than 50% over the past two decades, while only 22% of organizations report being highly effective at simplifying work. These are not conditions into which a productivity tool can be dropped and expected to function.

Meanwhile, the dominant AI deployment pattern concentrated 50% to 70% of budgets on sales and marketing pilots — the most visible but often least structurally sound use cases — while back-office functions with clearer ROI potential were underinvested, per the MIT Media Lab's findings reported by Forbes. The result: visible activity, minimal value creation, and a growing graveyard of abandoned proof-of-concept projects.

This is the environment — bloated, underdesigned, and structurally resistant to change — into which the world's largest companies decided to inject the most powerful productivity technology in a generation.

The Analysis

The Capacity-Capability Confusion

The most consequential misdiagnosis in enterprise AI is the conflation of capability with capacity. AI capability — what the models can do — has advanced remarkably. Capacity — an organization's ability to absorb, deploy, and sustain that capability — has not kept pace.

McKinsey's 2025 State of AI survey makes the distinction stark. Only 6% of respondents qualify as "high performers" — organizations attributing significant EBIT impact to AI. The defining characteristic of these organizations is not model sophistication or investment scale. High performers are nearly three times as likely as others to have fundamentally redesigned individual workflows. They have deployed twice as many use cases, and three-quarters are scaling or have scaled AI, compared with one-third of other organizations.

The inverse is equally telling. Among all respondents, 39% report any level of EBIT impact from AI, and most of those say the impact accounts for less than 5% of total EBIT. Nearly two-thirds of organizations have not yet begun scaling AI across the enterprise. The technology is present. The organizational architecture to use it is not.

This confusion drives a specific and expensive pathology: organizations respond to disappointing AI results by purchasing more AI. They add models, expand pilot programs, and hire prompt engineers. What they do not do is examine the workflows, incentive structures, and governance frameworks that determine whether those tools produce value or noise. It is the equivalent of buying a faster car and placing it on a road with no lanes, no signs, and no speed limits — then blaming the engine when traffic does not improve.

The Real Cost of Workslop

One of the most concrete manifestations of the capacity gap is a phenomenon researchers have labeled "workslop" — AI-generated output that appears polished but lacks substance, effectively shifting cognitive labor from the sender to the recipient.

Research from BetterUp Labs and Stanford, reported by CNBC, found that approximately 40% of workers received workslop in the past month. Recipients estimated that 15% of all material they receive now qualifies as low-effort AI-generated content. Each instance triggers an average of one hour and 56 minutes of downstream rework — reading, interpreting, correcting, or discarding content that should never have been sent. The invisible cost: roughly $186 per worker per month, or $9 million annually for a 10,000-person organization.

The damage extends beyond time. According to the same research, 53% of recipients reported annoyance, 38% confusion, and 22% said they felt offended. Nearly half of the workers who received work slop viewed the sender as less creative, capable, and dependable afterward. Roughly a third felt less inclined to collaborate with that person again. As Stanford's Jeff Hancock observed, generative AI has eliminated the effort barrier that previously constrained low-quality output: producing subpar work once "still required work. Now, that effort is eliminated," as reported by CNBC.

Workslop is not a fringe annoyance. It is a systemic symptom of deploying powerful generation tools without corresponding norms for quality, review, and purpose — a governance vacuum that only organizational design can fill.

Why AI Programs Stall in the Middle

A persistent myth holds that AI initiatives fail at the technical layer — bad data, wrong model, poor integration. The evidence points elsewhere. According to S&P Global/451 Research, 42% of companies abandoned most of their AI initiatives in 2025, up from 17% the prior year. The average organization scraps 46% of its proof-of-concept projects before reaching production.

The top challenges cited were not technical. They were confident in accuracy (29%), budget constraints (29%), staff resistance (28%), customer resistance (27%), and skill shortages (27%). Organizations with above-average failure rates considered fewer factors when prioritizing use cases (4.2 versus 4.9 for lower-failure organizations) and were 77% more likely to cite reputational damage as a concern.

This is the stall point: not at the start, where enthusiasm is high, and pilots are cheap, and not at the technical layer, where models are increasingly capable. AI programs stall in the middle — at the transition from proof of concept to production — where organizational realities collide with experimental assumptions. Scaling requires training people who did not volunteer. It requires changing processes owned by managers who were not consulted. It requires governance structures that do not yet exist. These are capacity challenges, not capability challenges.

The EY Work Reimagined Survey quantifies the gap: 88% of employees use AI, but only 5% use it in advanced ways that transform their work. Only 12% receive sufficient training. And 64% report that their workloads have actually increased since AI adoption — the opposite of what the investment was supposed to produce. Companies are missing up to 40% of potential productivity gains due to these talent and strategy gaps.

The Governance and Trust Vacuum

The absence of governance is not an oversight. It is a structural failure that compounds every other problem. Between 23% and 58% of employees across sectors are bringing their own AI solutions to work — shadow AI — according to the EY survey. These tools operate outside any quality framework, training protocol, or data governance policy.

The BCG/GPT-4 experiment documented in the ICLE review illustrates the stakes. When consultants used GPT-4 on tasks within the model's capability boundary, performance improved. When they used it on tasks just beyond that boundary, performance declined — because workers relied on plausible but incorrect outputs. The researchers described a "jagged technological frontier": AI exhibits uneven capabilities across tasks that appear similar in difficulty, and without verification protocols, workers cannot distinguish where the frontier lies.

This finding has direct governance implications. Without clear guidelines on where AI is reliable, which outputs require human review, and how quality is measured, organizations do not merely fail to capture value. They actively destroy it — generating confident-sounding errors at machine speed. McKinsey's survey found that 51% of organizations using AI have experienced at least one negative consequence, with nearly one-third reporting consequences stemming from AI inaccuracy. Only 28% of organizations have achieved what EY calls "Talent Advantage" — effective integration of talent and technology. The rest are operating in a trust vacuum where neither employees nor leaders have confidence in the system they have built.

Where This Argument Gets Complicated

The strongest counter to the organizational capacity thesis comes from macroeconomic data. U.S. productivity grew roughly 2.7% in 2025, nearly double the 1.4% annual average over the prior decade, as Stanford's Erik Brynjolfsson documented in Fortune. Fourth-quarter GDP tracked at 3.7% growth while job gains were revised sharply downward — the classic signature of a productivity surge. Brynjolfsson argues that we are "transitioning from an era of AI experimentation to one of structural utility," entering the "harvest phase" of the J-curve, where earlier investments begin to yield measurable output.

The Penn Wharton Budget Model projects that generative AI will increase GDP by 1.5% by 2035 and nearly 3% by 2055, with task-level labor cost savings averaging 25% today and growing to 40% over the coming decades. And at the firm level, the EY US AI Pulse Survey found that 96% of organizations investing in AI report some productivity gain, with 57% calling those gains significant — particularly at the $10 million-plus investment threshold, where 71% report significant gains.

These are real numbers, and the J-curve thesis deserves to be taken seriously. But two observations temper the optimism. First, macroeconomic productivity gains are dominated by a small number of sectors and firms — precisely the high performers and heavy investors that McKinsey's 6% represents. A rising aggregate number can obscure the fact that most organizations are not participating in the surge. Second, the controlled studies that show 15-55% task-level gains, catalogued in the ICLE review, consistently demonstrate those gains under experimental conditions with clear task boundaries and verification protocols — exactly the organizational design features that most enterprises lack. The technology delivers when the organization is designed to receive it. For the majority, it is not.

Implications for Leaders

Stop treating AI deployment as a technology initiative. The 6% of organizations achieving significant EBIT impact from AI differ from the rest not in their technology stack but in their willingness to redesign workflows, as McKinsey's data demonstrates. If your AI program reports to the CTO and is evaluated on model performance, you have already misclassified the problem. AI deployment is a transformation of operating models. Staff it, fund it, and govern it accordingly.

Audit your organizational capacity before expanding your AI portfolio. Deloitte found that 68% of workers lack sufficient uninterrupted time for important tasks, and 41% of work time produces no value. Adding AI tools to this environment does not create productivity. It creates work slop, shadow AI, and pilot fatigue. Before launching the next initiative, conduct what Deloitte terms a "sludge audit" — a systematic inventory of low-value work, unnecessary approvals, and duplicated processes. Clear the ground before planting the seed.

Establish explicit governance for AI quality, not just AI risk. Most AI governance frameworks focus on compliance — data privacy, bias, and security. These are necessary but insufficient. The BetterUp/Stanford workslop research shows that the absence of quality norms undermines trust, collaboration, and productivity from within. Define which outputs require human review. Specify where AI-generated content must be disclosed. Create team-level norms for what constitutes acceptable AI-assisted work — and make those norms as visible as your compliance policies.

Invest in training that changes behavior, not just awareness. Only 12% of employees receive sufficient AI training, per the EY survey, and the gap between awareness and capability is where value dies. Employees who received over 81 hours of annual AI training reported productivity gains of 14 hours per week — nearly double the median. Training must go beyond tool tutorials to include judgment frameworks: when to use AI, when to override it, and how to evaluate its output within the jagged frontier that the BCG experiment documented.

Redirect budget from pilots to scaling infrastructure. The 42% abandonment rate documented by S&P Global/451 Research reflects a structural failure at the proof-of-concept-to-production transition. Most organizations run too many pilots and invest too little in the change management, process redesign, and cross-functional coordination required to scale them. The EY data showing a $10 million investment threshold for significant gains suggests that concentrated, committed investment outperforms distributed experimentation.

The Bottom Line

The AI productivity paradox is not about patience or technology maturity. It is about most organizations' unwillingness to do the difficult, unglamorous work of redesigning how they operate. The 6% of companies capturing real value from AI are not using better models. They are running better organizations — with redesigned workflows, trained employees, clear governance, and leadership that treats AI as an operating system change rather than a software purchase.

The macro data says a productivity harvest is beginning. It is. But it is being reaped by organizations that invested in capacity, not just capability. For the rest, more spending on AI will produce more of what they already have: stalled pilots, tools that generate noise, and a busier but less productive workforce. The strategic question is no longer whether to invest in AI. It is whether you have built an organization capable of using it.

Sources

  1. McKinsey & Company. "The State of AI: Global Survey 2025." McKinsey, 2025. https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai

  2. Hill, Andrea. "Why 95% Of AI Pilots Fail, And What Business Leaders Should Do Instead." Forbes, 2025. https://www.forbes.com/sites/andreahill/2025/08/21/why-95-of-ai-pilots-fail-and-what-business-leaders-should-do-instead/

  3. S&P Global / 451 Research. "Generative AI Shows Rapid Growth but Yields Mixed Results." S&P Global Market Intelligence, 2025. https://www.spglobal.com/market-intelligence/en/news-insights/research/2025/10/generative-ai-shows-rapid-growth-but-yields-mixed-results

  4. EY. "EY Survey Reveals Companies Are Missing Out on Up to 40% of AI Productivity Gains Due to Gaps in Talent Strategy." EY Global Newsroom, 2025. https://www.ey.com/en_gl/newsroom/2025/11/ey-survey-reveals-companies-are-missing-out-on-up-to-40-percent-of-ai-productivity-gains-due-to-gaps-in-talent-strategy

  5. Deloitte. "When Work Gets in the Way of Work: Reclaiming Organizational Capacity." Deloitte Insights, 2025. https://www.deloitte.com/us/en/insights/focus/human-capital-trends/2025/reclaiming-organizational-capacity.html

  6. Niederhoffer, Kate. "AI-Generated 'Workslop' Is Destroying Productivity." Harvard Business Review, 2025. https://hbr.org/2025/09/ai-generated-workslop-is-destroying-productivity

  7. Liu, Jennifer. "AI-generated 'workslop' is here. It's killing teamwork and causing a multimillion dollar productivity problem, researchers say." CNBC, 2025. https://www.cnbc.com/2025/09/23/ai-generated-workslop-is-destroying-productivity-and-teams-researchers-say.html

  8. Ma, Jason. "One of Stanford's original AI gurus says productivity liftoff has begun." Fortune, 2026. https://fortune.com/2026/02/15/ai-productivity-liftoff-doubling-2025-jobs-report-transition-harvest-phase-j-curve/

  9. Arnon, Alexander. "The Projected Impact of Generative AI on Future Productivity Growth." Penn Wharton Budget Model, 2025. https://budgetmodel.wharton.upenn.edu/p/2025-09-08-the-projected-impact-of-generative-ai-on-future-productivity-growth/

  10. Lehmann, R.J. "AI, Productivity, and Labor Markets: A Review of the Empirical Evidence." International Center for Law and Economics, 2026. https://laweconcenter.org/resources/ai-productivity-and-labor-markets-a-review-of-the-empirical-evidence/

  11. EY. "AI-Driven Productivity Is Fueling Reinvestment over Workforce Reductions." EY US Newsroom, 2025. https://www.ey.com/en_us/newsroom/2025/12/ai-driven-productivity-is-fueling-reinvestment-over-workforce-reductions

Keep Reading