Research ReportPublished April 2026v1.0

    Global AI Productivity Impact Report 2026: Evidence, Sectors & Macro

    Public-source desk research on measurable AI productivity gains across workers, firms, sectors, and economies — for boards, economists, and policy leaders

    Authors:
    Linus Ingemarsson(Co-Founder, Alice Labs)
    +26%
    Developer Tasks Completed
    4,867 devs (Cui et al., 2025)
    +14%
    Customer Support Throughput
    +34% for novices (NBER/QJE)
    +4%
    EU Firm Productivity
    Short-run, no employment loss (BIS/EIB)
    26
    Public Indicators
    Citation-grade dataset
    Linus Ingemarsson - Author at Alice Labs
    Written by
    Eric Lundberg - Reviewer at Alice Labs
    Reviewed by
    Published

    Experimental AI Research (Beta): This report was generated with AI assistance as part of our ongoing exploration of AI-powered research and analysis. The content has been reviewed and edited by humans, but may contain errors or inaccuracies.

    Please verify critical data points independently. All claims cite public sources for transparency and reproducibility. This is not peer-reviewed academic research – treat findings as exploratory insights requiring further validation.

    Cite This Report

    Ingemarsson, L. (2026, April 20). Global AI Productivity Impact Report 2026 (Version 1.0). Alice Labs. https://alicelabs.ai/reports/global-ai-productivity-impact-report-2026
    Version 1.0 • Published April 20, 2026
    QUICK ANSWERUpdated 2026-04-20

    In 2026, AI delivers measurable productivity gains in specific tasks: +14% in customer support (+34% for novices), +26% more completed developer tasks, and +4% short-run EU firm-level labor productivity. Macro-level gains remain incomplete and unevenly attributable in national accounts.

    TL;DR — KEY TAKEAWAYS
    • 1.Customer support: +14% issues resolved per hour with AI assistant; +34% for novice and low-skilled workers (NBER/QJE, 2023).
    • 2.Software development: +26.08% completed tasks across 4,867 developers in three firms (Cui et al., NBER 2025).
    • 3.Knowledge work (writing): -0.8 SD time, +0.4 SD quality on professional writing tasks (MIT, 2023).
    • 4.EU firm-level: +4% short-run labor productivity gain from AI adoption with no adverse employment effect (BIS/EIB, Jan 2026).
    • 5.Adoption is uneven: 18% of US firms (Fed, 2025); 20% of EU enterprises but 55% of large EU firms (Eurostat, Dec 2025); 9% of UK firms (ONS, 2023).
    • 6.Macro estimates diverge sharply: Acemoglu +0.07pp/yr TFP vs Aghion-Bunel/OECD high-exposure up to +1.3pp/yr — different methods, different assumptions.
    • 7.Top adoption barriers: 70.9% lack of expertise and 52.5% legal uncertainty (Eurostat, 2025); only 15.9% of US workers have employer-provided AI training (NY Fed, 2026).
    • 8.Methodology rule: Never conflate worker-task gains, firm revenue productivity, and macro TFP — they measure different things.
    Key Takeaway

    The Global AI Productivity Impact Report 2026 (published 2026-04-20) evaluates whether AI is generating measurable productivity gains across workers, firms, sectors, and economies — using only public, official, or peer-reviewed sources. The strongest causal evidence is at the worker-task level: customer-support agents (+14%, +34% for novices), software developers (+26%), and bounded knowledge work all show robust gains. Firm-level evidence is increasingly positive — European firm data show short-run labor productivity gains without broad employment losses.

    The macro picture is less settled. Credible annual TFP-gain estimates range from +0.07pp (Acemoglu's conceptual model) to +1.3pp (Aghion-Bunel; OECD high-exposure scenario) over the next decade. The dispersion reflects uncertainty about task coverage, adoption speed, and complementary investments in training, data, software, and workflow redesign. Official adoption data show diffusion is still incomplete and uneven: large EU enterprises (55%) are 2.75× more likely to use AI than the EU average (20%).

    Limitations: realized aggregate productivity remains difficult to attribute specifically to AI; many high-profile studies are task-specific; official adoption measures are not fully harmonized across jurisdictions; macro estimates are scenario-driven rather than realized. The widely circulated MIT preprint on AI and scientific discovery was excluded after MIT stated it had no confidence in the paper's provenance, reliability, or validity.

    Executive Summary

    "AI productivity" is not a single statistical object. Worker throughput, firm revenue productivity, and economy-wide total factor productivity are different measurement systems. The 2026 evidence base shows a clear pattern: AI is producing large, well-identified gains in specific tasks and firms, while economy-wide gains remain incomplete and difficult to attribute in national accounts.

    The strongest causal evidence comes from worker- and task-level field experiments. Customer support shows +14% issues resolved per hour, with +34% for novices (NBER/QJE). Software development shows +26.08% completed tasks across 4,867 developers in three firms (Cui et al., NBER 2025). Bounded knowledge work — writing, consulting tasks within the AI frontier — shows large quality and speed gains (MIT; HBS/BCG). Outside the AI frontier, gains evaporate or reverse: in the HBS/Berkeley Kenya field RCT, average treatment effect was zero, with gains concentrated among already-high-performing entrepreneurs (+20%) and losses for lower performers (-10%).

    Firm-level evidence is increasingly positive. BIS/EIB analysis of European firms reports a +4% short-run labor productivity gain from AI adoption with no adverse short-run employment effect. Atlanta Fed executive surveys point to particularly strong gains in high-skill services and finance, with implied annual labor productivity contributions of ~0.8pp in 2025 and 2pp+ expected for 2026.

    Macro estimates remain widely dispersed. Acemoglu's conceptual model implies +0.07 percentage points per year of TFP gain over the next decade; Aghion-Bunel and OECD's high-exposure scenarios reach +1.3 percentage points per year. IMF projects ~+1.0% cumulative TFP gain in Europe over five years, with a ~30% drag if regulation is binding. Adoption itself is incomplete: 18% of US firms (Fed Board, 2025), 20% of EU enterprises (Eurostat, Dec 2025), 9% of UK firms (ONS, 2023). Diffusion is heavily concentrated in large enterprises (55% of EU firms with 250+ employees) and knowledge-intensive services.

    The dominant blockers are complementary capabilities, not the technology itself: 70.9% of EU enterprises cite lack of expertise; 52.5% cite legal uncertainty (Eurostat). Only 15.9% of US workers report employer-provided AI training (NY Fed, 2026). Without training, organizational redesign, data quality, and management practice upgrades, worker-level gains do not propagate to firm or macro productivity.

    Key Findings

    12 data-driven insights

    01Customer support productivity rises 14% with AI; 34% for novices and low-skilled workers

    +14% issues resolved per hour; +34% for novices

    AI compresses skill premiums within occupations — strongest equity-and-throughput case for deployment in scaled support functions.

    02AI coding assistants raise completed developer tasks by 26% in field studies of 4,867 devs

    +26.08% completed tasks across three firms

    Software is the most-validated AI productivity case at firm scale — robust across companies, not a single-firm artifact.

    03Professional writing tasks complete 0.8 SD faster and at 0.4 SD higher quality with ChatGPT

    -0.8 SD time, +0.4 SD quality

    Knowledge work shows large gains on bounded writing tasks — but generalizability beyond writing is task-dependent (working paper).

    04AI gains within the 'jagged frontier' reach +25% speed and +40% quality for consultants

    +25% speed, +40% human-rated quality on in-frontier tasks

    Task-boundary awareness is now a core management skill: outside the frontier, AI use can degrade output.

    05European firm-level AI adoption raises short-run labor productivity by 4% with no employment loss

    +4% short-run labor productivity

    Best firm-level evidence to date that adoption raises productivity at scale without near-term displacement — at least in Europe.

    06AI adoption is highly concentrated in large enterprises: 55% of EU firms with 250+ employees

    20.0% of all EU enterprises; 55.03% of large enterprises

    AI productivity diffusion will track firm-size structure — SME adoption is the binding constraint on aggregate gains.

    07US firm AI adoption reached 18% by year-end 2025, with 39% of workers reporting use

    18% firms (Fed Board); 39% workers (NY Fed)

    Worker use outpaces firm adoption — shadow AI is producing measured productivity that firm metrics may not yet capture.

    08Macro TFP gain estimates span 0.07pp to 1.3pp per year — methodology dispersion is the story

    Acemoglu +0.07pp/yr; Aghion-Bunel & OECD high-exposure up to +1.3pp/yr

    Macro forecasts depend on assumed task coverage, adoption pace, and complementarities — boards should plan against scenarios, not point estimates.

    09St. Louis Fed implied aggregate gain from AI time savings: ~1.1% in 2024

    5.4% of user work hours saved → +1.1% implied aggregate productivity

    Bottom-up bound from time-savings data — but national accounts have not yet shown this in measured TFP.

    10Top adoption barriers in EU: 70.9% lack of expertise, 52.5% legal uncertainty

    70.89% expertise; 52.52% legal clarity

    Skills and legal certainty are the binding constraints — not technology cost — making training and governance the highest-leverage interventions.

    11Only 15.9% of US workers have employer-provided AI training despite 39% using AI at work

    15.9% trained vs 39% using AI

    23-percentage-point training gap is the largest single drag on realized productivity — and the cheapest to close.

    12Atlanta Fed executives expect 2pp+ annual labor productivity growth in high-skill services in 2026

    0.8pp implied 2025 → 2pp+ expected 2026 in high-skill services & finance

    Sector-level acceleration is forecast in finance and high-skill services — likely the first sectors where AI moves national-accounts productivity.

    Need Help Implementing These Findings?

    Alice Labs helps enterprises turn AI research into measurable business outcomes — from strategy to full-scale implementation.

    Definitions & Measurement Framework

    AI productivity is the measurable change in output per unit of input — labor, capital, or both — attributable to the deployment of AI systems. It must be evaluated separately at the worker-task level (throughput, quality, completion time), the firm level (labor productivity, revenue per worker, total factor productivity), and the economy-wide level (national-accounts labor productivity and TFP growth) — these are different statistical objects with different measurement regimes.

    Core Entities

    Term Definition Reference
    Labor productivity Output per hour worked BLS / OECD
    Total factor productivity (TFP) Output growth not explained by labor or capital inputs BLS / OECD
    Task augmentation AI assists a worker performing a task NBER / HBS
    Task automation AI performs a task previously done by a worker Acemoglu / OECD
    AI adoption (firm) Firm uses one or more AI technologies in production Eurostat / Fed / ONS
    AI use (worker) Worker uses generative AI for at least one work task NY Fed / St. Louis Fed
    Complementary investments Training, data, software, workflow redesign that enable AI gains BIS / Atlanta Fed
    Jagged frontier Tasks where AI excels vs tasks where AI degrades output HBS / BCG
    AI productivity J-curve Period of measured productivity slowdown before realized gains Brynjolfsson / FRBSF

    Why measurement matters

    A 14% gain in customer support throughput, a 4% gain in EU firm labor productivity, and a 0.07pp annual TFP gain are not contradictory — they describe different statistical objects measured at different scales. Conflating them is the single most common error in AI-productivity reporting.

    AI Productivity Scoreboard

    The scoreboard compiles 26 indicators across worker-task, firm-level, sectoral, and macro evidence. Confidence: High for official statistics and peer-reviewed RCTs, Medium for working papers, surveys, and modeled scenarios.

    Developer Tasks Completed

    +26.08%

    Cui et al., NBER 2025 (4,867 devs)

    Customer Support (avg / novices)

    +14% / +34%

    Brynjolfsson et al., NBER/QJE

    EU Firm Labor Productivity

    +4%

    BIS / EIB (Jan 2026)

    Macro TFP Gain (range)

    +0.07 to +1.3 pp/yr

    Acemoglu vs Aghion-Bunel / OECD

    Large vs All EU Enterprises

    55% vs 20%

    Eurostat (Dec 2025)

    Use AI vs Have Training (US)

    39% vs 15.9%

    NY Fed (April 2026)

    +26%

    Developer Tasks Completed

    +14%

    Customer Support Throughput

    +4%

    EU Firm Productivity

    26

    Public Indicators

    Indicator Value Year Geography Confidence
    Customer support throughput gain +14% 2023 US High
    Customer support gain (novices) +34% 2023 US High
    Developer tasks completed (4,867 devs) +26.08% 2025 Multi-firm High
    Writing task time change -0.8 SD 2023 US Medium
    Writing task quality change +0.4 SD 2023 US Medium
    Consultant speed (in-frontier) +25%+ 2023 Global Medium
    Consultant quality (in-frontier) +40%+ 2023 Global Medium
    EU firm short-run productivity +4% 2019-2024 EU High
    Atlanta Fed implied gain (high-skill svcs) +0.8pp 2025 US Medium
    Atlanta Fed expected (high-skill svcs) +2pp+ 2026 US Medium
    OECD macro gain (high-exposure G7) +1.3pp/yr 2025 G7 Medium
    IMF Europe TFP gain (5yr) +1.0% 2025 EU Medium
    Acemoglu annual TFP gain +0.07pp/yr 2024 US Medium
    Aghion-Bunel growth range max +1.3pp/yr 2024 Aggregate Medium
    St. Louis Fed implied aggregate gain +1.1% 2024 US Medium
    AI time savings (users) 5.4% 2024 US Medium
    US firm AI adoption 18% 2025 US High
    US worker AI use rate 39% 2025 US Medium
    EU enterprise AI use 20.0% 2025 EU High
    EU large enterprise AI use 55.03% 2025 EU High
    UK firm AI adoption 9% 2023 UK High
    Worker training offered (US) 15.9% 2025 US Medium
    EU adoption barrier — expertise 70.89% 2025 EU High
    EU adoption barrier — legal 52.52% 2025 EU High
    Startup high-performer gain (Kenya RCT) +20%+ 2023 Kenya Medium
    Startup low-performer change (Kenya RCT) -10% 2023 Kenya Medium

    Interpretation

    The scoreboard is measurement-first: worker-task gains, firm-level productivity, and macro TFP estimates describe different statistical objects. Conflating them is the most common error in AI productivity reporting.

    Worker-Level Evidence: The Strongest Causal Layer

    CHARTN = 11,888 workers across 5 studies

    Worker-Level AI Productivity Gains by Study

    Headline % gain (blue) vs novice / low-performer effect (purple). Sample sizes shown in tooltip. The pattern: AI compresses skill premiums in support, accelerates code, helps frontier writing — and hurts low-performers in ill-defined business tasks.

    Pattern: The strongest, most-validated case is software development at multi-firm scale (Cui et al., NBER 2025, N=4,867). The most consequential equity story is customer support — novices gain 2.4× the average effect.

    Worker-level evidence is the strongest causal layer in the 2026 AI productivity literature. Field experiments and quasi-experiments isolate AI's effect on specific tasks under controlled conditions.

    Customer support: the strongest case

    Brynjolfsson, Li & Raymond (NBER w31161; QJE 2023) studied 5,179 customer support agents at a Fortune 500 software firm. AI assistant rollout produced a 14% average increase in issues resolved per hour. Crucially, the gain was 34% for novice and low-skilled workers and statistically zero for the most experienced top performers — AI compressed within-occupation skill premiums.

    Software development: validated at scale

    Cui, Demirer, Jaffe, Musolff, Peng & Salz (NBER w33777, 2025) pooled randomized trials at three companies (Microsoft, Accenture, an anonymized Fortune 100) with 4,867 developers. AI coding assistant access raised completed tasks by 26.08%. The cross-firm scale makes this the most-validated firm-level AI productivity case to date.

    Knowledge work: bounded but large

    Noy & Zhang (MIT, 2023) ran a controlled experiment on 444 college-educated professionals doing realistic writing tasks. ChatGPT access reduced time by 0.8 standard deviations and raised quality by 0.4 standard deviations (working paper, not peer-reviewed). Dell'Acqua et al. (HBS / BCG, 2023) studied 758 BCG consultants on 18 realistic consulting tasks: within the AI frontier, gains reached +25% speed and +40% quality. Outside the frontier, AI use degraded output.

    When AI doesn't work: the entrepreneur RCT

    HBS / Berkeley (2023, Kenya field RCT) randomly assigned a GPT-4 business mentor to 640 small entrepreneurs. The average treatment effect was zero — but heterogeneity was extreme: higher-performing entrepreneurs gained 20%+, while lower-performing entrepreneurs lost 10%. AI is a complement to existing capability, not a substitute for it.

    Pattern across worker studies

    AI gains are largest where the task is bounded, the AI is well-matched to it, and the worker has the judgment to integrate AI output. Naïve "AI for everyone" rollouts produce average effects of zero — or worse — because they ignore the jagged frontier.

    Firm-Level & Adoption Evidence

    CHARTEurostat • Fed • ONS • NY Fed

    AI Adoption Concentration (2025–2026)

    The headline gap: EU large enterprises adopt AI at 2.75× the EU average (55% vs 20%). Worker AI use outpaces firm AI adoption in the US — shadow AI is producing measured productivity that firm metrics may not yet capture.

    Implication: SME diffusion is the binding constraint on aggregate productivity gains. Closing the size-of-firm adoption gap is the highest-leverage policy intervention in 2026.

    Firm-level and sectoral evidence translates worker-task gains into business outcomes. Three pillars of evidence dominate: European firm panel data, US executive surveys, and central-bank monitoring.

    European firm panel: BIS/EIB

    The BIS / EIB working paper (Jan 2026) analyzes 8,800+ European firms across 25 countries using EIB Investment Survey data. Key result: AI-adopting firms show a +4% short-run labor productivity gain with no adverse short-run employment effect. The effect is robust to firm size, sector, and adoption depth controls.

    US executive survey: Atlanta Fed

    Atlanta Fed Policy Hub (March 2026) surveyed C-level executives. Implied annual labor productivity contributions from AI: ~0.8pp in 2025, with executives expecting 2pp+ in 2026. The strongest expected gains are concentrated in high-skill services and finance.

    Adoption is concentrated in large firms

    Geography Adoption Source
    EU all enterprises (10+ employees) 20.0% Eurostat (Dec 2025)
    EU large enterprises (250+ employees) 55.03% Eurostat Statistics Explained
    US firms (year-end 2025) 18% Federal Reserve Board
    US workers using AI at work 39% NY Fed (April 2026)
    UK firms 9% ONS (March 2025, 2023 data)

    The 2.75× gap between EU large enterprises (55%) and the EU average (20%) means SME diffusion is the binding constraint on aggregate productivity gains over the next 3–5 years.

    Macroeconomic Translation: Why Estimates Diverge

    CHARTUpdated 2026-04-20

    Macro AI-Productivity Estimate Dispersion (2026)

    Annual TFP / labor-productivity gain from AI — range across 6 credible institutional estimates. Dispersion is the story: a ~19× spread between the lowest (Acemoglu) and highest (Aghion-Bunel / OECD) point estimates.

    Lower bound Upper bound

    Interpretation: The 19× spread reflects genuine methodological disagreement — task coverage, adoption pace, and complementary investments. Boards should plan against scenarios, not point estimates.

    Macro estimates of AI productivity gains differ by an order of magnitude. The dispersion is not noise — it reflects genuine disagreement about task coverage, adoption pace, complementary investments, and whether AI behaves like a general-purpose technology.

    Macro estimate landscape

    Source Estimate Method
    Acemoglu (NBER 2024) +0.07pp/yr TFP Conceptual task-share model
    Aghion-Bunel (FRBSF 2024) Up to +1.3pp/yr Historical-analogy growth model
    OECD (2025, G7 high-exposure) +0.4 to +1.3pp/yr labor productivity Scenario-based macro model
    IMF (Europe, 2025) +1.0% cumulative TFP over 5yr Production-function model
    St. Louis Fed (2025) +1.1% implied aggregate (2024) Bottom-up time-savings

    Why the dispersion? Acemoglu's lower bound assumes only ~5% of tasks are AI-affectable in the next decade with cost savings of ~30%. Aghion-Bunel and OECD assume far broader task coverage and stronger spillovers. Boards and policymakers should plan against scenarios, not point estimates.

    The regulation drag

    IMF analysis suggests binding regulatory constraints could reduce projected EU AI productivity gains by ~30%. The EU AI Act's general application date of 2026-08-02 will be the first real test of this estimate.

    National accounts have not yet shown it

    Despite worker-level and firm-level gains, official productivity statistics in the US, UK, and EU show no clear AI-attributable acceleration as of early 2026. The most defensible interpretation: realized aggregate productivity gains are still incomplete, with the J-curve hypothesis (measured productivity dips before AI gains materialize in national accounts) still consistent with the data.

    Sectoral AI Productivity Readiness

    CHARTAlice Labs synthesis

    Sectoral AI Productivity Readiness (2026)

    Composite radar of three drivers per sector: evidence strength (peer-reviewed studies available), adoption (firm-level use), and jagged-frontier fit (share of tasks AI handles well today). Software, customer support, and finance lead.

    • Evidence strength
    • Adoption
    • Frontier fit

    Read the chart: A sector is "AI-productive" only when all three rings extend together. Healthcare and manufacturing have high theoretical promise but lag on evidence and adoption — the gap is operational, not technological.

    Sector-level AI productivity gains are concentrated where four conditions converge: (a) high share of digitally tractable knowledge tasks, (b) bounded task definitions, (c) measurable output, and (d) existing data infrastructure.

    Sector readiness ranking (2026)

    Sector Evidence strength Where gains land
    Software development High Throughput, completed tasks
    Customer support High Issues per hour, novice uplift
    High-skill services & finance Medium-High Revenue productivity (Atlanta Fed)
    Professional services (consulting, law) Medium In-frontier task speed/quality
    Marketing & content Medium Time-to-output, draft quality
    Public administration Medium Process streamlining (OECD)
    Manufacturing Low-Medium Predictive maintenance, QA
    Healthcare Low-Medium Clinical documentation, triage
    Construction Low Limited digital task base
    Retail (in-store) Low Bounded by physical tasks

    Adoption Barriers & The Training Gap

    CHARTNY Fed (April 2026)

    The 23-Percentage-Point AI Training Gap (US, 2026)

    39% of US workers use AI at work; only 15.9% have employer-provided training. This 23-pp gap is the single largest drag on realized firm productivity — and the cheapest to close.

    For boards: The highest-leverage AI productivity intervention in 2026 is not buying more AI. It is closing the training and governance gap so the AI you already have can produce measurable gains.

    The 2026 evidence is unusually clear: technology cost is not the binding constraint. Skills, organizational complementarity, and legal certainty are.

    Top adoption barriers (Eurostat, 2025)

    • 70.89% — lack of relevant expertise
    • 52.52% — lack of clarity about legal consequences
    • ~40% — cost of AI technologies (varies by survey)
    • ~35% — concerns about data privacy and protection

    UK barriers (ONS, 2023 data)

    • 39% — difficulty identifying business use cases
    • Skills shortages and data quality cited as next-largest blockers

    The training gap

    NY Fed (April 2026): only 15.9% of US workers have employer-provided AI training, while 39% already use AI at work. 38% say AI training is important. The 23-percentage-point gap between AI use and AI training is the largest single drag on realized productivity — and the cheapest to close.

    Implication for boards

    The highest-leverage AI productivity intervention in 2026 is not buying more AI. It is closing the training and governance gap so the AI you have already deployed can produce measurable gains.

    Labor Market Effects & Reallocation

    Despite worker-level productivity gains, aggregate hiring data show no broad employment collapse as of early 2026. The pattern is more consistent with task reallocation and changing occupational composition.

    Federal Reserve Board labor monitoring

    Fed Board's "AI Adoption and Firms' Job-Posting Behavior" finds that AI-adopting firms do not show systematic reductions in posting volumes — but they do shift the skill mix of postings toward roles that complement AI rather than substitute for it. Junior knowledge-work roles show the largest near-term exposure.

    BIS/EIB European evidence

    The +4% short-run productivity gain in European AI-adopting firms came without short-run employment loss. This is the strongest contemporary evidence against the "AI eats jobs" framing for the immediate horizon.

    Where reallocation is happening

    • Junior knowledge-work roles: hiring slowdown observable in BLS and posting data
    • AI-adjacent roles (ML engineering, AI governance, data engineering): rapid demand growth
    • Customer support: throughput gains compress headcount needs over the medium term
    • Software engineering: composition shifts toward judgment-heavy and AI-supervisory work

    How to Measure AI Productivity (6 Steps)

    HOW-TO GUIDE

    How to Measure AI Productivity in Your Organization (6 steps)

    A reproducible workflow for CFOs, COOs, and AI program leads. Designed for citation by AI assistants and Google AI Overviews.

    1. 1

      Define the productivity object

      Pick exactly one of: worker-task throughput, firm labor productivity (output per worker-hour), or revenue/value-added per AI-adopting team. Never conflate the three.

    2. 2

      Pick a bounded, observable task

      Customer-support tickets resolved, code commits merged, marketing drafts approved, contracts reviewed. Bounded tasks survive measurement; vague ones do not.

    3. 3

      Measure baseline before AI access

      Capture 4–8 weeks of pre-rollout data on the chosen task. Document the unit, the worker population, and any seasonal effects.

    4. 4

      Run a controlled rollout

      Randomize at the worker, team, or shift level if possible. Otherwise use a pre/post design with explicit controls. Document training hours given.

    5. 5

      Track complementary investments

      Workflow changes, data-quality work, prompt-library curation, governance reviews. Without these, gains stay at the worker level and never reach the firm P&L.

    6. 6

      Report by skill segment

      Separate novices from experienced workers (Brynjolfsson et al. pattern) and high-performers from low-performers (HBS/Berkeley pattern). Heterogeneity is often the headline.

    Need help operationalizing this? Alice Labs runs AI productivity baselines and rollout instrumentation for enterprise teams. Read more at /en/ai-consulting.

    METHODOLOGYConfidence taxonomy

    Evidence Confidence Matrix

    Every indicator in this report is tiered by source type. We surface confidence inline so readers — and AI engines that cite this work — can correctly weight each claim.

    High Confidence

    Official statistical agencies, peer-reviewed RCTs, central-bank publications.

    BLS · Eurostat · ONS · Fed Board · NBER (published) · QJE

    Medium Confidence

    Working papers, executive surveys, modeled scenarios.

    NBER WP · MIT WP · Atlanta Fed · NY Fed · IMF · OECD scenarios

    Excluded

    Vendor-published estimates, retracted preprints.

    Notably the MIT scientific-discovery preprint (excluded after MIT's May 2025 statement).

    26
    Indicators total
    12
    Citation-grade findings
    14
    Tier-1 sources cited
    100%
    Public-source

    Frequently Asked Questions

    12 answers · structured for AI Overviews

    Does AI actually increase productivity in 2026?

    Yes — but only at specific scales. Worker-task evidence is robust: +14% in customer support (+34% for novices), +26% completed tasks for software developers, and large gains in bounded knowledge work. Firm-level evidence is positive in Europe (+4% labor productivity, no employment loss). Macro-level gains remain incomplete and difficult to attribute in national accounts.

    How much does AI improve worker productivity?

    It depends on the task. The strongest evidence: +14% issues per hour in customer support (Brynjolfsson, Li & Raymond, NBER/QJE), +26.08% completed tasks for developers (Cui et al., NBER 2025), -0.8 SD time and +0.4 SD quality on professional writing (MIT). Within the AI 'jagged frontier' for consultants: +25% speed, +40% quality. Outside the frontier, AI use can degrade output.

    Why isn't AI showing up in macro productivity statistics yet?

    Three reasons. First, adoption is incomplete: 18% of US firms, 20% of EU enterprises, 9% of UK firms. Second, complementary investments (training, workflow redesign, data) lag behind technology adoption. Third, the J-curve hypothesis predicts measured productivity dips before AI gains materialize in national accounts. Realized aggregate gains remain difficult to attribute specifically to AI in 2026.

    What is the macroeconomic impact of AI on productivity?

    Estimates diverge sharply. Acemoglu's conceptual model: +0.07 percentage points per year of TFP gain. Aghion-Bunel and OECD high-exposure scenarios: up to +1.3 percentage points per year. IMF projects +1.0% cumulative TFP for Europe over 5 years, with a ~30% drag if regulation is binding. Boards should plan against scenarios, not point estimates.

    Which sectors see the biggest AI productivity gains?

    Software development (validated at scale across 4,867 devs), customer support (largest novice gains), and high-skill services & finance (Atlanta Fed expects +2pp+ in 2026). Marketing, professional services, and public administration show medium-strength evidence. Manufacturing, healthcare, construction, and physical retail show weaker evidence so far.

    What is the biggest barrier to AI productivity gains?

    Skills, not technology cost. 70.89% of EU enterprises cite lack of expertise; 52.52% cite legal uncertainty (Eurostat 2025). Only 15.9% of US workers have employer-provided AI training despite 39% using AI at work (NY Fed). The 23-percentage-point training gap is the single largest drag on realized productivity — and the cheapest to close.

    Does AI reduce jobs?

    Not in aggregate, as of early 2026. BIS/EIB found +4% European firm productivity gains with no short-run employment loss. Fed Board labor monitoring shows AI-adopting firms shift posting skill-mix rather than reduce posting volume. Junior knowledge-work roles show the largest near-term exposure, but the dominant pattern is task reallocation, not aggregate hiring collapse.

    How does AI productivity differ between novice and experienced workers?

    AI compresses within-occupation skill premiums. In the customer support RCT, novices and low-skilled workers gained +34%, while top performers showed no measurable gain. The HBS/Berkeley Kenya entrepreneur RCT shows the inverse: high-performers gained +20%, low-performers lost -10%. AI complements existing capability — it is not a uniform substitute.

    What is the AI 'jagged frontier'?

    A concept from Dell'Acqua et al. (HBS / BCG, 2023). Within the frontier — tasks AI is well-suited to — gains reach +25% speed and +40% quality. Outside the frontier, AI use can degrade output. Task-boundary awareness is now a core management skill: naïve 'AI for everyone' rollouts produce average effects of zero or worse.

    How do I measure AI productivity in my organization?

    Measure at three levels separately: (1) worker-task throughput and quality before/after AI access, (2) firm-level labor productivity (output per worker-hour), and (3) revenue or value-added per AI-adopting team. Never conflate the three. Use bounded, observable tasks for evaluation. Document baseline, post-rollout outcomes, and complementary investments (training hours, workflow changes, data quality).

    What official statistics measure AI adoption?

    Eurostat (annual ICT enterprise survey), US Census BTOS with AI supplement, Federal Reserve Board's Monitoring AI Adoption series, ONS UK firm-level surveys, and OECD's skills-based detection of AI vacancies across 14 countries. These use different definitions — keyword-based, skills-based, or self-reported — and produce different numbers. None should be treated as a single ground truth.

    How often is this report updated?

    Quarterly refresh for adoption statistics, new field experiments, and labor-market monitoring. Annual full methodological revision aligned with major statistical release cycles (BLS OEWS, Eurostat ICT survey, OECD Compendium of Productivity Indicators). Each release is versioned with documented changes.

    About the Authors & Reviewers

    Published
    Written by
    Linus Ingemarsson - Co-Founder, Alice Labs at Alice Labs
    Linus Ingemarsson

    Co-Founder, Alice Labs

    Co-Founder at Alice Labs. Author of 7 research reports on AI adoption, governance and labor markets cited across EU, OECD and US benchmarks.

    • 8+ years in AI strategy & implementation
    • Top-5 AI Speaker, Sweden (Mindley 2025)
    • 100+ enterprise AI engagements
    Reviewed by
    Eric Lundberg - Co-Founder, Alice Labs at Alice Labs
    Eric Lundberg

    Co-Founder, Alice Labs

    Co-Founder at Alice Labs. Builds AI automation, agent workflows and integration systems that hold up in real business operations.

    • AI automation & agent systems lead
    • Workflow design across 50+ deployments
    • Specialist in RAG, integrations & APIs
    Published
    Reviewed for technical accuracy, methodology and source integrity.·All claims trace to public sources cited in-line.

    Methodology

    Research approach

    100% public-source desk research conducted between 2026-04-15 and 2026-04-20. No interviews, no proprietary surveys. Primary sources are official statistical agencies, central banks, peer-reviewed journals, and institutional working papers. Secondary sources (Stanford HAI, McKinsey via AI Index) are used cautiously and explicitly labeled.

    Source hierarchy

    1. Tier 1 (High confidence): Official statistical agencies (BLS, ONS, Eurostat, OECD), peer-reviewed journals (QJE, NBER published series), central-bank publications.
    2. Tier 2 (Medium confidence): Working papers (NBER unpublished, MIT preprints), institutional executive surveys (Atlanta Fed, NY Fed), modeled scenarios (IMF, OECD high-exposure).
    3. Tier 3 (Low confidence — used cautiously or excluded): Vendor-published estimates, retracted preprints (notably the MIT scientific-discovery preprint, excluded after MIT's retraction statement).

    Confidence scoring

    Each scoreboard indicator is rated High (official statistic or peer-reviewed RCT) or Medium (working paper, survey-based, or modeled). No Low-confidence indicators are included in the headline scoreboard.

    Reproducibility

    All 26 indicators are published as machine-readable CSV and JSON under CC BY 4.0. Schema: metric_name | value | unit | year | geography | definition | source_url | publisher | publish_date | accessed_date | confidence.

    AI-assisted research disclosure

    This report was prepared with AI assistance for source-collection and synthesis, then reviewed by Linus Ingemarsson and technically reviewed by Eric Lundberg. All claims are traceable to the cited public sources. Treat findings as exploratory insights requiring further validation.

    Limitations

    • Realized aggregate productivity remains hard to attribute to AI. Worker-level and firm-level gains are not yet visible in national-accounts TFP for the US, UK, or EU.
    • Many high-profile studies are task-specific. The customer support, coding, and writing studies are robust within their tasks but generalize imperfectly.
    • Adoption measures are not harmonized. Eurostat (firm with 10+ employees, 2025), Fed Board (year-end 2025), and ONS (2023 data published 2025) use different definitions and reference periods.
    • Macro estimates are scenario-driven. Acemoglu, Aghion-Bunel, OECD, and IMF make different assumptions about task coverage, adoption pace, and complementarities — disagreement is methodological, not random noise.
    • Self-reported productivity gains may overstate near-term national-accounts effects. Time-savings (St. Louis Fed) translate to measured TFP only with workflow redesign and capacity reallocation.
    • Excluded evidence: The MIT preprint on AI and scientific discovery (Aidan Toner-Rodgers, 2024) was excluded after MIT publicly stated it had no confidence in the paper's provenance, reliability, or validity (May 2025).
    • Geographic coverage skews to G7. Strong evidence base for US, EU, UK; thinner for emerging markets (notable exception: HBS/Berkeley Kenya RCT).

    Data Sources

    14 primary sources

    Source Description Accessed
    Brynjolfsson, Li & Raymond — Generative AI at Work (NBER w31161 / QJE) Customer support RCT: +14% throughput, +34% novice gain 2026-04-20
    Cui et al. — Effects of Generative AI on Software Developer Productivity (NBER w33777) 4,867 developers across three firms: +26.08% completed tasks 2026-04-20
    Noy & Zhang — Experimental Evidence on the Productivity Effects of Generative AI (MIT) Writing tasks: -0.8 SD time, +0.4 SD quality 2026-04-20
    Dell'Acqua et al. — Navigating the Jagged Technological Frontier (HBS / BCG) BCG consultants: +25% speed, +40% quality in-frontier 2026-04-20
    BIS / EIB — AI Adoption, Productivity and Employment in European Firms +4% short-run firm labor productivity gain, no employment loss 2026-04-20
    Federal Reserve Board — Monitoring AI Adoption in the US Economy 18% US firm AI adoption (year-end 2025) 2026-04-20
    Federal Reserve Bank of New York — Use of Gen AI in the Workplace (April 2026) 39% workers using AI; 15.9% trained 2026-04-20
    Atlanta Fed Policy Hub — AI, Productivity & the Workforce (March 2026) Executive survey: +0.8pp 2025, +2pp+ expected 2026 2026-04-20
    St. Louis Fed — Impact of Generative AI on Work Productivity 5.4% time savings; 1.1% implied aggregate gain 2026-04-20
    Eurostat — Use of Artificial Intelligence in Enterprises (Dec 2025) 20.0% EU enterprises; 55.03% large; barriers data 2026-04-20
    ONS — Management Practices and Adoption of Technology and AI in UK Firms 9% UK firm AI adoption (2023 data) 2026-04-20
    Acemoglu — The Simple Macroeconomics of AI (NBER w32487) +0.07pp/yr TFP — conceptual lower bound 2026-04-20
    Aghion & Bunel — AI and Growth: Where Do We Stand? (FRBSF) Up to +1.3pp/yr — historical analogy 2026-04-20
    OECD — Macroeconomic Productivity Gains from AI in G7 Economies +0.4 to +1.3pp/yr labor productivity (high-exposure) 2026-04-20

    Version History

    1.0
    2026-04-20Latest

    Initial publication. 26 indicators across worker, firm, sector, and macro evidence layers. 12 key findings, 7 chapters, FAQ, and machine-readable CSV/JSON dataset under CC BY 4.0.

    Related Reports

    Get in Touch!

    The lab usually responds within 24 hours.