Evidence (8066 claims)
Adoption
5586 claims
Productivity
4857 claims
Governance
4381 claims
Human-AI Collaboration
3417 claims
Labor Markets
2685 claims
Innovation
2581 claims
Org Design
2499 claims
Skills & Training
2031 claims
Inequality
1382 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 417 | 113 | 67 | 480 | 1091 |
| Governance & Regulation | 419 | 202 | 124 | 64 | 823 |
| Research Productivity | 261 | 100 | 34 | 303 | 703 |
| Organizational Efficiency | 406 | 96 | 71 | 40 | 616 |
| Technology Adoption Rate | 323 | 128 | 74 | 38 | 568 |
| Firm Productivity | 307 | 38 | 70 | 12 | 432 |
| Output Quality | 260 | 71 | 27 | 29 | 387 |
| AI Safety & Ethics | 118 | 179 | 45 | 24 | 368 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 75 | 37 | 19 | 312 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 74 | 34 | 78 | 9 | 197 |
| Skill Acquisition | 98 | 36 | 40 | 9 | 183 |
| Innovation Output | 121 | 12 | 24 | 13 | 171 |
| Firm Revenue | 98 | 35 | 24 | — | 157 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 87 | 16 | 34 | 7 | 144 |
| Inequality Measures | 25 | 76 | 32 | 5 | 138 |
| Regulatory Compliance | 54 | 61 | 13 | 3 | 131 |
| Task Completion Time | 89 | 7 | 4 | 3 | 103 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 33 | 11 | 7 | 98 |
| Wages & Compensation | 54 | 15 | 20 | 5 | 94 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 27 | 26 | 10 | 6 | 72 |
| Job Displacement | 6 | 39 | 13 | — | 58 |
| Hiring & Recruitment | 40 | 4 | 6 | 3 | 53 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 11 | 6 | 2 | 41 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 6 | 9 | — | 27 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Effectiveness and safety of AI agents require structured guardrails and human-in-the-loop designs; AI agents function as scalable cognitive infrastructure only conditional on such governance.
Synthesis of deployment experience and analysis of constraints; recommendation grounded in observed model reliability issues, governance complexity, and oversight needs from the Alfred AI experiments.
Deployment of AI agents shifts demand toward roles focused on oversight, orchestration, prompt/agent engineering, and governance, creating new types of labor that may offset some direct labor reductions.
Authors' inference based on observed need for human oversight and orchestration in deployments; not quantitatively measured in the study (no headcount or labor-share data reported).
AI‑enabled risk assessment (weather, pests, price forecasts) can improve index insurance and credit scoring for smallholders, lowering financing costs and increasing investment — but it also raises concerns about data bias and exclusion.
Pilot programs and modeling studies on index insurance and credit scoring, combined with policy analyses documenting equity and bias risks; primary empirical work is limited to pilots and simulations.
Returns to AI investments depend on complementary investments in farmer knowledge, extension services, and local institutions; AI tends to amplify returns to managerial skills and digital literacy.
Empirical studies and randomized/quasi‑experimental trials showing complementarity effects, and qualitative evidence from stakeholder interviews; cited studies report larger impacts where complementary services exist.
Impacts of technology‑ecology integration are heterogeneous: they vary by farm size, crop type, local infrastructure, and farmer skills; smallholders can benefit substantially but are more constrained by liquidity, information, and market access.
Observational econometric analyses and randomized/quasi‑experimental studies reporting heterogeneous treatment effects, supplemented by qualitative interviews and case studies documenting constraints faced by smallholders.
Data governance, platform market structure, and inclusive policy design determine whether gains from AI/IoT are widely shared or captured by large firms.
Policy review, conceptual analysis, and case studies of platform markets that document capture risks and distributional outcomes linked to data ownership and market concentration.
Innovations can reduce emissions and resource use per unit of output but risk lock‑in to input‑heavy models unless ecological principles and monitoring are integrated.
Case study and pilot evidence showing reduced input intensity or emissions intensity in some interventions; conceptual discussion and examples highlighting trade‑offs and potential for input‑intensive lock‑in absent ecological safeguards.
Labor-market consequences will involve reallocation effects: routine-task automation, rising returns to managerial and technical skills, and potential within-firm wage dispersion.
Synthesis of labor economics theory and prior empirical work on automation; book recommends matched employer-employee panel studies to trace these effects but does not report such new panel results.
AI’s effects vary by industry, task composition, and firm capabilities; high-data, standardized-task sectors see faster, deeper impacts.
Cross-sector examples and theoretical arguments about task routineness and data intensity; calls for heterogeneity-aware empirical designs (e.g., difference-in-differences with staggered adoption).
Automation of routine tasks raises demand for cognitive, interpersonal, and technical skills; firms face reskilling needs and changing task allocation between humans and machines.
Task-level analytic framework and literature review on automation effects; book recommends empirical approaches (e.g., occupation and job-task data) to quantify these changes but does not present a single large empirical estimate.
Managers shift from routine decision execution to tasks involving oversight, interpretation, strategic design, and ethical stewardship of AI systems.
Qualitative case studies and literature review of task-level research; suggested task-analytic methods rather than reporting a specific empirical task dataset.
AI complements some researcher tasks (idea generation, analysis, writing) and substitutes others (routine editing, literature searches), changing skill demand and training priorities.
Stated under Labor Market Effects. Supported conceptually and likely by task-level studies or surveys; abstract doesn't cite specific empirical evidence or measurement details.
Impacts of AI adoption are broad, affecting individual researcher productivity, team workflows, and institutional outcomes in scholarly communication and digital scholarship.
Key Points summary. Basis likely includes mixed-methods evidence (surveys/interviews at individual and team levels, case studies, platform usage data) synthesized in the paper; abstract lacks detail on scope and samples.
Investment in governance and training is a necessary cost to realize sustained returns from generative AI; these costs influence adoption timing and the distribution of benefits.
Conceptual argument from the review supported by case examples and economic reasoning about complementary investments.
There is a risk of wage polarization: increased returns to AI‑complementary skills and potential downward pressure on wages for automatable tasks.
Theoretical synthesis drawing on economic models of skill‑biased technological change and early empirical observations; no definitive causal wage studies reported.
Generative AI will drive occupational reallocation by substituting routine cognitive tasks while complementing higher‑order cognitive and monitoring skills.
Theoretical labor economics arguments synthesized with early empirical examples; no large‑scale causal labor market study provided in the review.
Routine, boilerplate, and debugging tasks are most automatable or complemented by LLMs, shifting value toward design, verification, and systems thinking.
Task-level analyses, observational studies, and synthesized findings showing larger gains on repetitive or templated tasks versus high-level design tasks.
Liability and intellectual-property ownership around AI-assisted code are unresolved practical and legal concerns.
Legal and policy analyses, practitioner reports, and qualitative interviews noting ambiguous legal frameworks and unresolved questions about ownership and liability for AI-assisted code.
Token taxes reduce some geographic tax arbitrage relative to input taxes but do not eliminate cross-border avoidance; international coordination and trade/regulatory levers are crucial.
Political-economy analysis and recommendations in the paper; no international case studies or empirical coordination outcomes provided.
The framework quantitatively captures trade-offs between public-health outcomes and economic stability across macroscopic scenarios and different LLM backends.
Quantitative analysis reported across scenarios and model variants, tracking trade-off metrics between health (infection curves) and economic outcomes (aggregate activity). The summary notes cross-backend comparisons but does not report numerical effect sizes.
When coupled with an epidemic–economic model, the LLM-PDA framework robustly generates divergent macro trajectories across scenarios.
Coupled epidemic and economic modules in simulation; experiments run across diverse macroscopic scenarios (varying transmissibility, shocks, policy regimes) with metrics tracked at macro scale (infection prevalence over time, aggregate economic indicators). (Number of scenarios and runs not specified in summary.)
A robust empirical pattern in the literature is that AI’s effects vary by skill level: displacement risk is concentrated among lower-skilled tasks while augmentation and wage gains are more likely for higher-skilled tasks.
Empirical findings and syntheses cited (Brynjolfsson et al., 2023; Chen et al., 2024) that report task- and skill-differentiated effects on employment and wages; evidence comprises cross-sectional exposure analyses and panel studies in the cited literature.
TGAIF clarifies where GenAI acts as a complement (augmenting consultant capability) versus where it risks substitution.
Conceptual distinction and mapping presented in the TGAIF derived from practitioner accounts; theoretical/qualitative, not empirically quantified across tasks.
TGAIF implies reallocation of work away from GenAI‑suitable subtasks (routine synthesis, drafting, summarization) toward tasks where human judgment and client interaction add most value.
Based on authors' inductive analysis of practitioner interviews describing which subtasks firms consider suitable for GenAI and which require human oversight; qualitative, not quantitatively tracked reallocation.
Aligning consulting tasks with generative-AI capabilities via a Task–GenAI Fit (TGAIF) framework can unlock substantial efficiency gains while containing key risks (notably hallucinations and loss of skill retention).
Inductive framework developed from qualitative, interpretive interviews with practitioners at leading German management‑consulting firms. The abstract does not report sample size, interview protocol, or quantitative validation; evidence is based on practitioner reports and the authors' synthesis.
DAR implies changes to labor and contracting: reversible AI leadership reshapes task boundaries, demand for oversight skills, and should be reflected in contracts and procurement with explicit authority-reversal rules and audit obligations.
Theoretical/ normative argument in implications section; no empirical labor or contract data included.
AI substitutes for routine coding tasks but complements higher-order tasks such as system architecture, integration, and orchestration.
Interpretation from qualitative evidence at Netlight where practitioners used AI for routine chores while retaining control of higher-order design tasks; no quantitative task-time displacement data presented.
Human roles are shifting toward oversight, curation, specification, and orchestration of multiple AI components and tools.
Synthesized from practitioner descriptions and changing task allocations observed in the Netlight fieldwork (interviews/observations); no longitudinal measurement of role changes reported.
Short-run consumer gains from faster, cheaper service can be undermined by trust losses from hallucinations or perceived deception, reducing long-term consumer surplus.
Conceptual welfare analysis and cited case examples in the literature; no longitudinal consumer-surplus measurement provided in this review.
Conventional productivity metrics (e.g., handle time) may misstate value because they do not capture multi-dimensional impacts like quality and trust.
Conceptual critique and synthesis of measurement challenges discussed in the literature; no empirical measurement study presented in this review.
There is potential for substantial cost savings and throughput gains in repetitive, high-volume interactions, but these are offset by costs for integration, monitoring, and error remediation.
Industry case examples and conceptual cost–benefit reasoning aggregated in the review; the paper contains no new quantitative cost estimates or sample-based measurements.
Generative AI will substitute for routine service tasks while complementing skilled workers for escalations and complex problem solving, shifting labor demand toward supervisory and relationship-focused roles.
Economic and labor-market analyses synthesized in the review; projections are inferential and based on heterogeneous secondary sources, not primary labor-market experiments.
Full automation of customer service is suboptimal because persistent risks (hallucinations, contextual errors, lack of genuine empathy, integration complexity) remain; hybrid human–AI systems achieve the best outcomes.
Synthesis of documented failure modes and practitioner case examples from the literature; no primary experimental data or controlled trials in this review. Inference based on heterogeneous empirical reports and conceptual analyses.
Welfare effects of democratized access to AI-assisted ideation are ambiguous: access could democratize innovation but also amplify low-quality outputs and misinformation absent proper curation.
Theoretical discussion and empirical examples of misinformation/low-quality outputs from LLMs cited in the review; no comprehensive welfare accounting provided.
Net gains in innovation from increased idea volume depend on complementary human capacity for curation and development; raw increases in ideas do not automatically translate into higher-quality innovation.
Synthesis noting studies where idea quantity rose but downstream quality or successful development did not necessarily increase; review highlights heterogeneity across workflows and dependence on human integration.
The most effective deployment model is a 'cognitive co-pilot' in which AI expands and challenges the idea space while humans provide curation, strategic evaluation, and experiential judgment.
Prescriptive conclusion drawn from synthesis of studies where human-AI collaboration (human curation/selection) produced better downstream outcomes than AI-alone outputs; evidence heterogenous and largely short-term.
Generative AI functions as a dual-purpose cognitive tool: a high-volume catalyst for divergent idea generation and a structured assistant for decomposing complex problems.
Nano-review / synthesis of existing empirical literature on LLM-assisted creativity and problem-solving, drawing on experimental ideation tasks, design/ideation studies, and applied case evidence; no original dataset or new experiments in this paper.
Net value from generative AI is contingent: gains are largest where breadth of ideas and rapid iteration matter, and smaller or riskier where deep domain expertise, tacit knowledge, or high-stakes judgments are required.
Synthesis of heterogeneous empirical results showing task-dependent benefits; argument grounded in observed differences across lab and field contexts and documented limitations in domain-specific performance.
Generative AI raises measurable productivity (lower marginal cost per interaction) but introduces quality and trust externalities; optimal deployment balances these trade-offs.
Pilot cost analyses and operational reports showing lower marginal costs per interaction alongside documented quality/trust issues; primarily observational and model-based reasoning.
Full automation produces trade-offs unfavorable to complex service quality and trust; hybrid models with human-in-the-loop control are preferable.
Synthesis of case studies, pilot results, and conceptual reasoning comparing fully automated routing to hybrid/human-in-the-loop deployments; limited randomized comparisons.
Generative AI can materially improve customer service productivity through 24/7 automation, scalable personalization, and agent augmentation — but is not a substitute for humans.
Synthesis of deployments, pilot studies, vendor reports, and some experimental A/B tests described in the paper; no pooled sample size provided and much evidence is short-run or observational.
Data-driven HRM reinforces skill-biased technological change: routine HR tasks are being substituted by automation while demand rises for analytical and interpersonal skills.
Theoretical implication and synthesis across studies in the review noting automation of routine tasks and increased demand for analytic/interpersonal skills.
Adoption will be heterogeneous and distributional effects will follow: organizational readiness, regulatory environments, and industry structure will drive uneven adoption and competitive impacts.
Review finds varying adoption patterns in empirical and practitioner literature and synthesizes theoretical reasons for heterogeneity; empirical causal estimates are noted as scarce.
One-off AI features typically produce limited returns unless organizations build complementary human and process capabilities and adapt governance and incentives.
Interpretive synthesis of case studies and practitioner guidance showing short-lived or limited benefits from isolated feature deployments without complementary investments.
Blockchain and decentralized fintech tools could increase transparency and access to alternative assets for women, but practical adoption barriers remain.
Qualitative assessment of blockchain capabilities and uptake surveys / case studies cited in the article (product analyses and early adoption data; no large‑scale causal evidence).
Governance reduces downside risk (compliance fines, outages) but raises implementation costs; economic assessments must weigh risk-adjusted returns.
Conceptual economic argument in the paper; supported by reasoning and practitioner experience but not by empirical cost–benefit studies within the article.
When evaluating GenAI investments, firms should treat prompt-fraud controls and monitoring as persistent operating costs rather than one-time setup costs.
Practical recommendation informed by conceptual cost and governance analysis; not supported by longitudinal cost studies in the paper.
Smaller firms or departments using shadow AI may realize productivity gains but face outsized fraud exposure due to weaker controls.
Theoretical trade-off analysis in the implications section; no empirical firm-level comparisons or experiments presented.
Safer scaling of automation may increase substitution of routine ERP/CRM tasks while governance and oversight roles create complementary high-skill positions (e.g., compliance engineers, auditors, prompt engineers).
Labor-market implications presented as theoretical reasoning based on how governance and automation interact; informed by practitioner observation but not empirically tested in the paper.
Overall, secure and resilient cloud infrastructure supported by SECaaS facilitates broader and safer diffusion of AI but creates economic trade-offs (market concentration, externalities, liability) that require empirical study and policy responses.
Synthesis of the chapter's literature review, case studies, and theoretical arguments; calls for empirical methods (regressions, event studies, structural models) to quantify effects.