Three years after ChatGPT crossed the 100-million-user threshold, the debate about generative AI has matured from "will it work?" to "where is it working, and how much?" The NVIDIA State of AI 2026 report, surveying 3,200 enterprises, finds that 91% of businesses are now deploying AI in production, up from 72% in 2024 and 55% in 2023. But enterprise usage is not the same thing as measurable productivity gain, and not every sector is capturing the value that the technology appears to offer.
This guide is about the five sectors where the gap between hype and reality has narrowed most — where productivity improvements are real, replicable, and reflected in published peer-reviewed studies as well as company filings. It is also, honestly, about the sectors where the gap remains wide. Our goal is to give investors, executives and workers a sober picture of where the AI productivity dividend is genuinely accruing in 2026, and where the hype continues to run ahead of the evidence.
Our framework throughout this report is evidence-based. We have looked at adoption data from major surveys, peer-reviewed productivity studies, regulatory disclosures from listed companies, and operational case studies from specific deployments. We explicitly excluded vendor marketing claims, press-release-driven "pilot program" results without independent verification, and speculation about future capability. What remains is a narrower but more trustworthy picture of where AI is actually changing economic productivity in 2026. We identify five sectors where the evidence meets our threshold, five where it falls short, and a handful where evidence is still accumulating. The aim is to help readers calibrate their own expectations — whether as investors, managers, workers, or simply as citizens attempting to understand an unusually consequential technology transition.
1. The Productivity Puzzle: Why the Debate Exists
At the aggregate level, US labour productivity growth runs at 1.7% year-on-year in 2026 — elevated relative to the 1.3% post-2010 average, but not dramatically so. Against hyperscaler capex of $600 billion a year and near-universal enterprise adoption, this is a modest number. This gap between what AI should be doing and what the macro data shows is the heart of the 2026 productivity puzzle.
Three explanations for the gap are all partially true. First, diffusion takes time. Prior general-purpose technologies — electricity, computing, the internet — all required 15–25 years between adoption and measurable productivity gains in the macro data. AI is diffusing faster than any of them, but "faster than electricity" still allows for significant lags in the statistics. Second, measurement is hard. Productivity statistics are calibrated to the manufacturing era; they measure widgets per hour well and knowledge-worker output poorly. Much of AI's value appears in quality, speed, and customer satisfaction rather than in widgets-per-hour equivalents. Third, distribution is skewed. PwC finds that 74% of AI value is captured by the top 20% of companies in each industry. The aggregate number averages across all companies, including the laggards; the sector data, when examined carefully, shows large gains at the frontier.
The skill in 2026 is to move past the aggregate debate and ask: in which specific sectors, doing which specific tasks, is AI delivering measurable productivity gains right now? That question has answers, and the answers point consistently to five sectors where the evidence is substantial, peer-reviewed and replicable across organisations.
A brief historical analogy is useful. When electricity was being deployed in US manufacturing between 1890 and 1930, productivity statistics did not show meaningful gains until the 1920s — roughly 30 years after initial adoption. The reason was that firms initially used electricity simply to power existing steam-era machinery, without redesigning factories around electric motors. Only when factory architecture itself was redesigned around the new technology (single-story layouts, modular production lines, workflow optimisation) did productivity accelerate dramatically. AI in 2026 is arguably in the "electrifying existing processes" phase; the equivalent of the 1920s redesign phase — workflows rebuilt around AI from the ground up — is beginning but not complete.
This pattern suggests two things. First, current productivity statistics likely understate AI's eventual impact. Second, the firms that move earliest toward workflow redesign — not just tool adoption — will capture disproportionate gains as the statistics catch up. We see this pattern in enterprise data: firms that restructured their healthcare documentation workflow around AI tools are seeing dramatically better productivity than firms that grafted AI onto unchanged workflows. The same applies in software engineering, financial services and manufacturing.
2. How We Identify Real Productivity Gains
Our criteria for inclusion are deliberately stringent. A sector qualifies for inclusion if: (i) at least two peer-reviewed or government-research studies demonstrate measurable productivity gains exceeding 15%; (ii) gains are replicated across multiple firms, not just a single lighthouse deployment; (iii) the underlying tasks being improved are core to the sector's economic activity, not edge cases; and (iv) the gains are sustained over at least 12 months rather than novelty effects.
Many sectors currently running "AI pilot" programmes fail these tests. They have adoption, but not measurable gain. They have experimentation, but not replication. Our five sectors all pass the four tests with clear evidence.
Meeting all four criteria is a deliberately high bar. A sector with one strong study but no replication is interesting but unproven. A sector with widespread pilot activity but no published outcome metrics is speculative. A sector with strong results in an edge-case application (such as marketing copy generation, where AI clearly works but the task is not economically core) does not qualify. The five sectors discussed below each pass the full test, and therefore represent the strongest available evidence base for AI as a productivity technology.
What the tests explicitly do not require is that AI has replaced human labour. In fact, the most persuasive evidence comes from augmentation, not replacement: a doctor using AI makes fewer errors, a coder using AI ships features faster, a customer-service agent handles more inquiries per hour. The productivity gain is in the augmented worker-plus-AI unit, not from firing the worker. This framing matters because it shapes how the gains are distributed, how sectors are adapting, and how the political economy of AI is unfolding.
A second methodological choice worth naming is our focus on productivity rather than revenue or valuation. Many AI success stories discussed in financial media focus on what AI companies are selling, not on what AI is producing in the using sectors. These are different questions. A firm may sell billions of dollars of AI services without those services producing equivalent productivity gains at the customer level; conversely, AI may be producing genuine productivity gains at customer firms while AI-vendor valuations remain uncertain. The productivity lens — measuring what AI enables its users to do — is the more durable guide to where the technology is genuinely creating economic value.
A third choice is our focus on sectors rather than technologies. Much AI commentary is organised around specific technologies — large language models, computer vision, reinforcement learning, robotics. This is useful for technical understanding but less useful for economic analysis, because what matters economically is which sectors deploy which technologies to which effect. A mature computer-vision application in manufacturing may be more economically consequential than a flashy new LLM capability that has not yet found a reliable business use. Sector-first analysis keeps the focus on actual economic outcomes.
3. Sector 1: Healthcare & Medical Diagnostics
Healthcare is, by our measurement framework, the sector with the strongest evidence of productivity gains from AI in 2026. Three specific applications lead the field: medical imaging and diagnostics, clinical documentation, and drug discovery.
Why does healthcare lead? Three features of the sector combine favourably. First, healthcare generates enormous quantities of structured and unstructured data — imaging, clinical notes, lab results, vital signs — that are exactly the kind of inputs AI excels at processing. Second, healthcare productivity is persistently constrained by labour scarcity (clinicians are expensive and limited) and administrative burden (documentation alone consumes 2+ hours of a typical physician's day). Third, the cost-per-error in healthcare is high enough to justify substantial investment in tools that improve accuracy. Each of these characteristics makes healthcare an unusually good match for current-generation AI capability.
Medical imaging: the clearest evidence base
The landmark study remains the Swedish MASAI mammography trial, which followed 105,934 women across a multi-year period. The trial compared AI-assisted mammogram reading (with human radiologist review) against conventional double-blind human reading. The AI-assisted arm identified 29% more breast cancers without increasing false positives, while reducing the radiologist workload by 44%. This is a genuine productivity-and-quality improvement simultaneously — the holy grail of healthcare intervention research. The trial has been replicated in smaller studies in the UK, Germany and Israel, with broadly consistent results.
Clinical documentation is an even more visible application. The German firm Clinomic, whose Mona product is deployed in ICUs across Europe, reports a 68% reduction in documentation errors and a 33% reduction in documentation workload for intensive-care clinicians. Similar products from Abridge and Nuance/DAX in the US show workload reductions of 2–4 hours per clinician per day — time that is redirected to direct patient care. These are not laboratory findings; they are deployed products generating measurable operational improvements across hospital systems.
Drug discovery is the longest-duration application and the most speculative about future value. Current-gen AI drug-discovery platforms — from Isomorphic, Recursion, Insilico and others — are showing materially shorter timelines from target identification to candidate selection, potentially cutting 2–3 years off the preclinical phase of a typical drug programme. Whether this translates into more approved drugs (the binding constraint of the pharma pipeline) is still an open question, but the early evidence is encouraging. The first wave of AI-discovered drugs is now entering late-stage trials.
For investors, the healthcare AI trade is cleaner than most other sectors because the value accrues visibly: health systems report reduced costs, pharma companies report shortened timelines, device makers report better outcomes. Specialist health-AI firms, healthcare IT incumbents (Epic, Cerner/Oracle), and big-pharma adopters each offer exposure. The risks — regulatory approvals, insurance reimbursement, deployment complexity — are substantial but tractable.
Within hospital operations, a less-publicised AI win is patient-flow and bed-management optimisation. Large US academic medical centres — Mass General Brigham, Cleveland Clinic, UCLA Health — have deployed AI models that predict admissions, discharge timing, surgical theatre availability and ICU demand. The result is operational throughput gains of 8–15% with no additional capital investment, which in healthcare economics is extraordinary. These systems are not newsworthy in the way that diagnostic AI is, but they are likely more valuable in aggregate because they scale across every department of every deploying institution.
Primary care is a harder AI story. The volume of primary-care encounters is enormous, the clinical complexity is real, and the reimbursement structure in most healthcare systems under-rewards time-intensive preventive care. Early-stage AI tools — symptom triage, chronic-disease management, medication-adherence coaching — show promise but have not yet produced the measured productivity improvements of hospital-based applications. The next 3–5 years will probably close this gap, but the evidence base today is thin.
The global picture is uneven. North American and Western European health systems have moved fastest. Japan and South Korea are aggressive adopters, particularly in elder-care applications where demographic pressure is acute. Chinese hospital systems are deploying indigenous AI tools at remarkable pace. Most emerging markets remain early-stage, though India's Apollo Hospitals and several Brazilian, Mexican and Gulf state health systems have impressive deployments. The net is that the best healthcare AI deployments now originate from a geographically diverse set of institutions, not just the US.
4. Sector 2: Financial Services
Financial services is the sector where AI productivity gains are most skewed toward high-skilled work. The Atlanta Fed working paper 2026-04, analysing hundreds of finance-sector firms, found that AI-assisted workers in finance roles are 30–45% more productive on complex analytical tasks, with the gains concentrated among the most senior workers (senior analysts, portfolio managers, risk officers) rather than juniors. This is an inversion of the Argentina/Denmark result that applies to other sectors, and it reflects something specific about how finance work is structured.
The reason for the inverted pattern in finance is that finance work at senior levels is primarily about judgement, synthesis and communication — precisely the dimensions where AI augmentation adds the most value. A senior portfolio manager who uses AI to rapidly digest earnings calls, company filings and analyst research is producing a much larger multiple of their baseline output than an entry-level analyst whose work was already heavily structured. In this sense, finance provides a useful preview of what AI does to knowledge work generally: it amplifies expertise rather than replacing it.
Customer service is the sector where AI has most visibly changed the customer experience and where adoption has moved from pilot to production at the fastest pace. The NVIDIA report finds that 48% of enterprises have deployed agentic AI systems in customer service, up from 22% in 2024. Modern systems handle multi-turn conversations, integrate with backend systems to execute transactions, and increasingly handle complex issues that previously required human escalation.
The applications are wide. Research production — equity research, credit analysis, macro strategy — has been transformed by AI's ability to summarise filings, extract named entities from regulatory filings, and generate first-draft analyses. Compliance — KYC, AML transaction monitoring, regulatory reporting — is another high-volume area where AI now handles 60–80% of initial flagging. Portfolio management uses AI for portfolio construction, rebalancing, and factor research. Fraud detection is almost entirely AI-driven in the large card networks (Visa, Mastercard, American Express).
JPMorgan, which has been unusually transparent about its AI programme, reports 200+ production AI deployments across trading, credit, research and operations, with measurable productivity improvement across each. The bank's 2025 annual report attributed roughly $2 billion of operating efficiency to AI-related initiatives — a material number even at JPMorgan's scale. Goldman Sachs, Morgan Stanley, Bank of America and the European majors all report broadly similar trajectories.
Consumer finance is a distinct story. Robo-advisors, personalised financial planning tools, credit-decisioning platforms, and conversational banking interfaces are all being transformed. The pricing power of mass-market financial advice is compressing as AI makes previously bespoke services available to smaller account balances. For workers, the implication is that the junior financial advisor role is more at risk than the senior one; the Atlanta Fed pattern will likely hold across the cycle.
One caveat: financial services is heavily regulated, and regulators have been cautious about pure-AI decision-making in areas like lending, insurance underwriting and investment advice. The "human in the loop" requirement is increasingly binding as regulators update their frameworks. This is a healthy constraint that also limits the full automation potential in the sector.
Insurance is a sub-sector worth discussing separately. Claims processing, underwriting, actuarial analysis and fraud detection are all natural AI applications. Major insurers — Allianz, AXA, Chubb, Progressive — report 20–35% productivity gains in claims operations, with particular strength in automating the initial intake and triage of straightforward claims. Underwriting is harder because of regulatory sensitivity, but the directional trend is clear: AI-assisted underwriting is gradually replacing manual underwriting for mid-complexity cases, while humans retain authority over complex cases. Reinsurance, surprisingly, has been slower to adopt — partly because the volume of transactions is lower, and partly because each individual transaction is large enough to justify extensive human analysis.
Wealth management is at an inflection. Historically a high-touch, high-margin business requiring human advisors, wealth management is now seeing AI-augmented hybrid models gain market share. Morgan Stanley's AI-powered advisor assistant, JPMorgan's private-client AI tools, and several fintech challengers are making personalised financial advice available to smaller account balances than previously economic. The winners in this transition are the firms that combine strong human advisor teams with the best AI augmentation. Pure-AI robo-advisors have plateaued; pure-human advice without AI support is progressively uncompetitive.
Trading is the most AI-intensive financial activity, but the changes in 2026 are less dramatic than the changes of the 2000s–2010s. Algorithmic trading, high-frequency trading and systematic quant strategies were already heavily automated before the current AI wave. What generative AI adds is new capability in unstructured data processing — parsing regulatory filings, central bank statements, satellite imagery, social media sentiment. Several hedge funds report material alpha from these sources, though competitive dynamics will compress returns over time as the techniques diffuse.
5. Sector 3: Software Development
Software development is the sector where the AI productivity evidence is most unambiguous — and where the gains are most visible in the everyday experience of workers. The landmark study by Harvard Business School, tracking software engineers across multiple firms, found that those using AI-assisted coding tools produced work 25.1% faster, with 40%+ higher quality as measured by code review defect rates.
The sector lends itself to clear measurement because software development output is observable: commits merged, tickets closed, pull requests approved, features shipped, bugs fixed, test coverage improved. Engineers and their managers can directly compare productivity with and without AI assistance across teams, time periods and projects. This measurability has produced a rich evidence base that other sectors can only envy.
A related reason software development leads is cultural fit. Software engineers were among the first professionals to have both the skills and the inclination to adopt AI tools aggressively. Developer-facing tool vendors — GitHub, Anthropic, OpenAI, Cursor, Cognition — have rapidly built products specifically designed for engineering workflows, often with unusual levels of input from professional developers. The result is that AI tools for coding are, by 2026, more refined and better integrated into daily workflows than in almost any other domain.
Subsequent studies have been broadly consistent. A study at a major global bank found that AI-assisted coders cleared Jira tickets 31% faster. A GitHub-sponsored study of Copilot usage across hundreds of teams found similar results. The Denmark study by Humlum and Vestergaard found that 93% of workers in AI-equipped firms adopt the tools, with 19% reporting 1+ hours per day saved. These are reproducible, cross-firm findings.
What is interesting is the mechanism. AI coding tools do not write software autonomously; they accelerate specific subtasks — boilerplate generation, documentation, test-writing, bug diagnosis, code review, and searching large codebases. The productivity gain is therefore a composite of many small time savings, not a single dramatic automation. This has proven more durable than the more dramatic "AI replaces programmers" predictions of 2023–24, which have not materialised.
The sector-level implication is that software teams are producing more output per engineer. Firms report roughly 15–25% more features shipped per quarter than pre-AI baselines. Some firms have used this to hire less aggressively; others have reinvested the capacity in new product lines. The net labour-market effect for experienced developers has been mildly positive (wages up, employment stable), while entry-level developer hiring has softened — as AI reduces the value of tasks traditionally given to juniors.
Longer-term, the software sector is likely to be the first to see truly "agentic" AI deployed at scale: AI systems that plan, execute and verify multi-step tasks with reduced human intervention. The 2025–26 generation of agent-capable coding assistants (Cursor, Claude Code, Cognition's Devin, and similar) is approaching usable quality for discrete engineering tasks. Full-stack agentic development is probably 2–3 years away, and it will reshape software team structures when it arrives.
The distributional pattern within software development is interesting. Our conversations with engineering leaders at major technology firms suggest that AI assistance produces larger absolute gains for experienced engineers than for juniors, but larger relative gains for juniors than for seniors. That is, a senior engineer who uses AI tools produces more total additional output than a junior who does, but the junior's output rises more in percentage terms. The practical implication is that AI is both a junior-productivity accelerator and a senior-productivity multiplier, and firms that recognise this are restructuring their engineering organisations accordingly.
The tools market itself has become competitive in a way that would surprise anyone who predicted winner-take-all dynamics two years ago. GitHub Copilot remains the market share leader, but Cursor has built a loyal developer base, Claude Code has emerged as a favoured tool for specific use cases, Anthropic's and OpenAI's direct-API-to-editor integrations continue to evolve, and several specialised tools have carved out niches (Codeium for enterprise, JetBrains AI for IntelliJ users, Devin for agentic workflows). For developers, this competitive dynamic is healthy; for investors, it suggests pricing power at the tools layer is more limited than the bullish case assumes.
6. Sector 4: Manufacturing & Quality Control
Manufacturing is the oldest "productivity through technology" sector, and AI is the latest wave. The specific applications with the strongest evidence are visual quality inspection, predictive maintenance, and process optimisation.
Manufacturing's AI productivity story is distinctive because it builds on decades of earlier automation. Modern factories were already deeply instrumented — sensors, SCADA systems, MES platforms, and statistical process control — before AI arrived. The incremental productivity gain from AI is therefore layered on top of an already optimised baseline, which makes the additional percentage gains look smaller than in newer-to-technology sectors, but the absolute economic value is often larger because it applies to high-volume, thin-margin operations.
The specific mechanisms through which AI adds value to manufacturing are different from other sectors. In healthcare and software, AI augments individual knowledge workers. In manufacturing, AI optimises physical processes, detects anomalies faster than statistical methods allow, and enables ongoing learning across production runs. The result is a productivity gain that accrues not to individual workers but to the capital stock of the manufacturing facility — which has important implications for who captures the economic value.
Visual inspection has been automated with rule-based machine vision for 30 years, but pre-AI systems struggled with complex, varied or subtly defective items. Modern AI vision systems handle these cases — detecting hair-thin surface cracks, subtle colour variations, micro-defects invisible to human inspectors. Siemens' partnership with major automotive OEMs reports 30–50% defect-detection rate improvements across production lines, with proportional reduction in downstream warranty costs.
Predictive maintenance combines IoT sensor data with AI models to predict equipment failure before it happens. This is genuinely transformative for industries where unplanned downtime is expensive — chemicals, refining, semiconductor fabs, steel mills, power generation. Case studies from Ascend Performance Materials, Schlumberger and several major utilities report 20–35% reductions in unplanned downtime. At plant scale, this is worth tens of millions of dollars annually.
Process optimisation uses reinforcement learning and large models to optimise complex, multi-variable production processes. Semiconductor fabs have used these techniques for years; recent AI advances are bringing similar capability to chemicals, pharmaceuticals manufacturing, food processing, and steel-making. Gains of 2–8% in yield translate into large absolute numbers when applied to sectors with tight margins.
The labour implications in manufacturing are more about skill-mix than job count. Plants need fewer pure manual inspectors and more data-and-systems engineers. Skilled technicians who combine domain knowledge with data fluency are commanding premium wages. The pattern is familiar from earlier automation waves: aggregate employment in manufacturing is not falling faster than it already was, but the composition of jobs is shifting materially.
Supply-chain and logistics applications deserve a separate mention. Large logistics networks — Amazon, Maersk, DHL, FedEx — are deploying AI for route optimisation, demand forecasting, warehouse operations and last-mile delivery. The productivity gains here are large (10–25% in measurable efficiency metrics) and compound rapidly because the underlying networks touch so many transactions. Smaller logistics firms that cannot afford in-house AI infrastructure are increasingly dependent on third-party platforms, which is producing a quiet consolidation in the sector.
Industrial robotics, once a parallel-track technology to AI, is increasingly integrated. Modern industrial robots use AI vision systems, learn from demonstration rather than explicit programming, and adapt to varied inputs. Japanese (Fanuc, Yaskawa), European (ABB, KUKA) and US (Rockwell, Emerson) industrial automation firms each report growing AI-integrated product share. For investors, this sub-sector has the rare characteristic of combining demographic tailwinds (labour scarcity in manufacturing regions), AI technological progression, and reshoring-driven demand in a single thesis.
The geographic distribution of manufacturing AI matters. China has been an extremely aggressive adopter in its electric vehicle, solar panel and electronics manufacturing base, leveraging scale and domestic AI capability. Germany and Japan lead in complex precision manufacturing. The US is strong in semiconductor fabrication, aerospace and speciality chemicals. Each region is finding niches where it combines existing industrial strength with AI augmentation. Manufacturing is therefore not a sector in which a single geography will dominate through AI; the gains are distributed, and the competitive dynamics remain intense across multiple regions.
7. Sector 5: Customer Service & Support
The productivity gain, measured as tickets resolved per human hour, has been substantial. Major call centres report 35–55% higher human agent productivity because AI handles tier-one queries directly, routes complex issues to the right specialists, and provides real-time coaching to agents during calls. Salesforce, ServiceNow and the specialist customer-service AI firms (Cresta, Observe.AI, and others) have documented these improvements in hundreds of deployments.
The customer-facing story is more nuanced. Customer satisfaction with AI-handled interactions is roughly on par with human-handled interactions for simple transactional inquiries, and still below humans for complex emotional or ambiguous issues. The 2026 frontier is improving AI performance on the complex cases, and most of the leading systems are making measurable quarter-on-quarter progress there.
The labour-market effect has been significant. Entry-level call-centre headcount has declined in most advanced economies, with the effect particularly visible in the Philippines, India and parts of Eastern Europe where large call-centre industries exist. But the decline has been gradual rather than collapse-shaped, and the remaining human agents are higher-skilled and better-paid than before. This is consistent with the broader AI-augmentation pattern: fewer workers in the entry tier, stable-to-rising wages for the experienced tier.
B2B customer support is a particularly high-impact application. Enterprise software vendors, cloud providers and specialist B2B service firms report that AI-handled technical support now resolves 55–70% of customer inquiries without human involvement, while simultaneously producing higher customer satisfaction scores than pre-AI human-only support. The mechanism is that AI systems have perfect product knowledge, infinite patience, instant availability, and no call-centre wait times — advantages that outweigh the disadvantages in most technical-support contexts. Human support agents have been elevated to handle only the most complex or sensitive issues, which produces a more satisfying job for the remaining workforce.
Field service — dispatched technicians repairing equipment at customer locations — is also being reshaped. AI-powered diagnostic assistants running on technicians' tablets reduce the time needed to identify and solve problems by 20–30%. Remote expert systems allow a single senior technician to guide multiple juniors on different sites simultaneously via AR or video. These applications are slower to scale than pure software applications, but the productivity gains are durable and accumulating.
Interactive: The Productivity Bar Race
Enterprise AI Adoption, by Sector (2022–2026)
Press "Play" to animate enterprise AI adoption by sector over five years. Each bar is the share of firms in the sector with AI in production. Notice which sectors moved fastest — and which are still catching up.
8. The Distributional Story: Who Captures the Gains
Stepping back from sector-by-sector, a distributional story emerges that is more striking than any individual sector trend. PwC's landmark finding that 74% of AI value is captured by the top 20% of companies in each industry is visible in firm-level data. In every one of the five sectors we highlight, the productivity gains are concentrated among the largest, best-run, best-capitalised firms.
This concentration has several causes. First, data matters: AI value is proportional to the quality, scale and accessibility of an organisation's data. Large firms have more data and better data infrastructure. Second, capital matters: serious AI deployment requires substantial infrastructure spend, which smaller firms cannot match. Third, talent matters: the engineers and product managers who can deploy AI effectively are scarce, and they cluster at leading firms. Fourth, change management matters: the hardest part of enterprise AI is not the technology but the workflow redesign and training, and this is where professional management capacity — concentrated at larger firms — matters most.
The implications for investors, managers, policymakers and workers are large. For investors, the concentration implies that sector ETFs may understate the divergence — the market-cap-weighted return may capture the winners but mask the underperformance of laggards. For managers, the imperative is to be in the top 20% of one's sector, because the middle will be squeezed. For policymakers, the concentration raises legitimate concerns about competition, barriers to entry, and the political economy of AI deployment. For workers, working at top-20% firms may matter more in 2026 than it did in 2019.
The geographic distribution of value capture is also informative. US firms capture a disproportionate share of AI-related earnings growth, driven by the concentration of hyperscalers, leading software firms, and AI-native businesses in North America. European firms are generally in the second tier of adopters, with some standouts (SAP, Schneider Electric, AstraZeneca, Spotify) but an overall slower pace. Asian firms — particularly in China, Japan, South Korea and increasingly India — are catching up rapidly in specific sub-categories (manufacturing AI, consumer AI, B2B SaaS). The geographic picture is fluid, but the concentration among a relatively small set of leading firms globally is unlikely to weaken.
A subtler distributional story is within sectors. Within healthcare, academic medical centres and large integrated systems capture more than community hospitals. Within financial services, global universal banks capture more than regional or specialist firms. Within manufacturing, global-scale OEMs capture more than mid-market producers. Each of these patterns reflects the reality that AI value is proportional to operational scale, data richness, and capital intensity — all characteristics that larger firms have more of.
9. Where AI Is NOT Delivering (Yet)
A rigorous view of AI productivity requires honesty about where the gains are absent or small. Several categories of work show disappointing or ambiguous evidence.
General white-collar office work (administrative tasks, routine email management, meeting coordination) is a "small gains, wide spread" category. Most workers report some productivity gain, but the gains are modest (5–15%) and harder to measure. AI here is more like better spreadsheet software than transformative technology.
Complex creative work — strategic planning, original research, long-form writing, advanced design — shows ambiguous results. AI accelerates certain sub-tasks but creative leaps still require human judgement. Firms that have tried to replace senior creative roles with AI have typically walked those changes back within 12 months.
Education and training shows pockets of strong gain (personalised tutoring, automated grading, language learning) but the broad application has been slower than expected. Regulatory, institutional and cultural factors slow adoption in education more than in most other sectors.
Legal services outside of document review and due diligence shows more modest gains than initially predicted. Case strategy, novel argumentation and client relationship work remain human-dominated.
In-person services (restaurants, hospitality, care, construction) show limited AI productivity impact because the physical component of the work is not easily automated. This is the sector cluster most insulated from near-term AI disruption — a fact worth naming because it sometimes gets lost in the noisier discussions of AI impact.
Government and public-sector work has been notably slow to adopt AI despite clear use-case fit. Procurement constraints, data-security concerns, union contracts and political caution combine to slow deployment well below the private-sector pace. Several national governments — Singapore, UAE, UK and some Nordic countries — are exceptions, with measurable AI deployment in tax administration, benefits processing and regulatory approvals. Most other governments are 3–5 years behind equivalent private-sector institutions. This is a productivity opportunity that will eventually be realised but is currently dormant.
Agriculture shows scattered but slowly growing adoption. Precision agriculture, satellite-based crop monitoring, and AI-driven yield optimisation tools are real, but penetration remains low because of the fragmented nature of global agriculture and the capital constraints of most farmers. Large agribusiness operators and commodity traders are adopting faster, but small and medium farms — which represent the majority of global food production — are largely unaffected.
Consumer-facing retail has been an ambiguous story. Large e-commerce platforms (Amazon, Alibaba, Shopify) have deployed AI extensively in personalisation, pricing, fraud and logistics. But the in-store retail experience has been less transformed than widely predicted. Checkout-free stores, AI-powered staff assistants and personalised in-store displays all have working deployments, but adoption remains limited and customer reaction mixed. The retail productivity story remains less convincing than the five sectors we highlight.
10. The Enterprise Adoption Playbook
For organisations that want to move from the "adopting AI" camp to the "capturing productivity gains" camp, the evidence suggests a disciplined playbook. Successful deployments share common features across our five high-performing sectors.
Start with a clear use case, not with the technology. The most successful deployments begin by identifying a specific, measurable business problem, then selecting the right AI approach. Enterprises that chase "AI initiatives" without problem specification consistently under-deliver.
Invest in data infrastructure first. AI value is data-proportional. Firms that attempt to deploy AI on top of fragmented, low-quality data find the projects fail. Data quality, integration, access governance, and metadata discipline are prerequisites.
Treat this as a workflow redesign, not a tool deployment. The biggest gains come from redesigning how work flows through the organisation with AI embedded — not from sprinkling AI tools into existing processes. This requires operational management leadership, not just technology leadership.
Invest in training and change management. Workers need both technical skills (how to prompt, when to accept AI output) and judgement skills (when to override, how to combine AI output with domain knowledge). Firms that skip this investment see pilot-phase success but deployment-phase failure.
Measure outcomes, not activity. The right metrics are sector-specific but always output-focused: features shipped, diagnoses made, cases resolved, defects caught. Input metrics (AI usage, prompts per day) are leading indicators but not the goal.
Close the feedback loop. The best-performing firms systematically capture user feedback, compare AI outputs to gold-standard outputs, and iterate model selection and prompt design based on real performance. This operational discipline is the difference between a firm in the top 20% and one in the next 60%.
Common deployment failure modes
Equally instructive are the patterns that produce underperformance. Our work with enterprises suggests five failure modes account for the majority of disappointing AI deployments.
First is "pilot purgatory" — projects that succeed in proof-of-concept but never reach production because the deployment team lacks the operational authority or infrastructure to scale them. This is particularly common in large organisations where innovation teams sit outside the operating line of business.
Second is "tool sprawl without integration" — an organisation deploys multiple AI tools from different vendors without coherent data-integration architecture, producing duplicated effort and inconsistent results. The symptom is AI spend rising without measurable productivity gain.
Third is "compliance theatre" — adopting elaborate governance processes that in practice prevent any real deployment. Regulatory concerns are legitimate, but over-engineered approval pipelines can produce a net productivity loss relative to the pre-AI baseline.
Fourth is "vendor lock-in to inferior tools" — commitments to specific AI vendors made in 2023–24 when the market was less competitive now produce subpar results compared with current best-in-class alternatives. Willingness to re-evaluate tool choices regularly is important, not just contract discipline.
Fifth is "under-investment in human capability" — deploying tools without retraining the workforce to use them well. The tools may be purchased but remain unused or misused. The return on training spend is typically 5–10x the return on additional tool spend, yet training budgets are routinely cut first when projects come under cost pressure.
11. Risks, Constraints and Pushback
Even the sectors we identify as clear AI winners face meaningful risks and constraints. Three deserve explicit naming.
Regulatory risk is real. The EU AI Act is now in force; the UK has adopted a principles-based approach; China has detailed generative AI regulations; US state-level and federal regulation is fragmenting. Healthcare and financial services face the most stringent regimes, which limits the pace of deployment even where technical capability exists.
Model-risk and reliability concerns are more significant than vendor marketing suggests. AI systems continue to produce incorrect outputs ("hallucinations") at rates that, while declining, remain non-trivial. In high-stakes applications — medical, legal, financial advice — human oversight is not optional. This is a feature, not a bug, but it constrains the labour-displacement potential of AI in these sectors.
Labour-market pushback is emerging. Entry-level hiring is contracting in several AI-exposed fields, and worker unions, professional associations and some political voices are beginning to push back against rapid deployment. The 2024 Hollywood writers' strike is a preview of broader labour responses. Firms that deploy AI without genuine engagement on the labour implications will face operational friction, reputational risk and potential regulatory response.
A fourth risk worth naming is model-vendor concentration. The enterprise AI market remains dominated by a small number of foundation-model providers — Anthropic, OpenAI, Google, Meta and a few others. This concentration produces genuine commercial risks: pricing power that may shift in the vendors' favour, availability and uptime dependencies, and strategic vulnerability to vendor-specific decisions or outages. Sophisticated enterprises are increasingly adopting multi-model strategies to mitigate these risks, but single-vendor dependency remains common.
Finally, data-privacy and intellectual-property risks are real. AI systems inadvertently trained on proprietary data, customer data shared with foundation-model providers without clear governance, and the as-yet-unresolved questions about copyright in AI-generated content each present legal exposure that many firms have not fully grappled with. The litigation landscape is still forming, and early cases will establish precedents that reshape how enterprises can and cannot use AI tools. Firms that treat AI governance as a compliance afterthought rather than a core operational priority are accepting tail risks that may well materialise.
12. What Happens Next: The 2026–2030 Trajectory
Looking ahead, four trajectories are worth anticipating. First, agentic AI moves from demo to production. Systems that execute multi-step tasks with minimal human intervention are already viable in narrow domains (coding, customer service, financial trading). By 2028 they will be mainstream in the five sectors we identify, which will change the productivity math.
Second, the productivity gap widens. The 74% / 20% pattern we observe today is likely to strengthen rather than weaken, because top performers will compound their advantages. This has major implications for how economies — and firms — think about competitiveness.
Third, domain-specific AI outperforms general AI in enterprise contexts. Foundation models provide the substrate, but the winners in the specific sectors will be firms that combine general AI capabilities with deep domain knowledge and proprietary data. This favours incumbents in regulated industries (healthcare, finance, legal) who have both.
Fourth, the labour market bifurcates further. High-skilled workers using AI will continue to command premium wages. Entry-level routine work will continue to compress. The middle — middle-management, middle-skill white-collar — will experience the most disruption. This is a difficult social and political transition, and will be one of the defining labour-policy debates of the decade.
Fifth — an often-overlooked observation — the returns to infrastructure investment will compound. The current AI boom has been characterised by massive capital deployment: hyperscaler capex at historic scale, specialised chip manufacturing capacity coming online, and continent-spanning data-centre build-outs. This infrastructure takes years to construct but then operates for a decade or more. The 2026–2030 period will see the fruits of 2022–2025 investment showing up in productivity statistics, and investors who can look past the cyclical noise to the structural accumulation of AI-capable infrastructure will compound at faster rates than those focused purely on quarterly earnings.
The meta-observation for workers, executives and investors is that the AI productivity story is genuinely only at its beginning. The five sectors we highlight are where the evidence is strongest today, but the pace of change suggests that several additional sectors will reach the same evidence threshold within 2–3 years. Rather than treating the current list as fixed, readers should monitor the emergence of new evidence bases. Our working hypothesis is that insurance, legal services, government administration and selected agriculture applications are the most likely candidates to join the "clear productivity gains" list by 2028.
What enterprises should be doing now
For boards and executive teams thinking about how to position their organisations over the 2026–2030 window, three priorities stand out. Infrastructure readiness — clean data, integrated systems, secure cloud foundations — remains the prerequisite. Firms that spent the 2022–24 period building this capacity are now executing; firms that did not are still trying to catch up. Talent strategy must combine retention of senior domain experts with aggressive hiring and promotion of AI-fluent operators. The gap between firms with this combination and those without is visible in quarterly earnings calls. Portfolio rebalancing — divesting from activities where AI is commoditising margins, investing in activities where AI compounds advantages — is a strategic decision that most boards are only now engaging with seriously.
For investors, the sector analysis above translates into several actionable themes. Quality-concentrated exposure to AI leaders in each sector has outperformed broad-based exposure. Cost-efficient AI-enabled firms are gradually displacing legacy competitors in healthcare, financial services and manufacturing — the substitution is slow but durable. Infrastructure firms supporting AI deployment (data, cloud, security, networking) continue to capture a disproportionate share of value-chain economics.
For workers, the practical takeaway is unusually clear. AI fluency is becoming a baseline workplace competence, not a specialist skill. The worker who combines genuine domain expertise with pragmatic AI fluency is the single highest-leverage profile in the 2026 labour market. Investment in developing this combination pays dividends that compound over the rest of the career.
The Bottom Line
Three years after the generative AI revolution began, the productivity story is real — but concentrated. Healthcare, financial services, software development, manufacturing and customer service show measurable, replicable, peer-reviewed productivity gains. Other sectors show smaller or more contested results. For investors, executives and workers, the practical imperative is to identify where the gains are occurring, how they are distributed, and what the adoption playbook looks like at the firms that are actually capturing them. The AI hype cycle is ending; the AI productivity era is beginning.