Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
Predictive governance can change fiscal timing (earlier interventions) and alter uncertainty profiles for public budgets, requiring economists to model dynamic fiscal impacts and risks from algorithmic failure or bias.
Implication drawn in the review from case studies and economic reasoning present in the literature; recommendation for fiscal modeling based on synthesized evidence across the 103 items.
Interoperability and ethical‑by‑design requirements influence vendor lock‑in, competition, and the emergence of platform providers in markets for public‑sector AI solutions.
Policy and market analyses within the reviewed literature that link technical standards and ethical design requirements to market structure and vendor dynamics (synthesized from the 103 items).
Predictive analytics and AI enable anticipatory policy design (early intervention, forecasting), but they raise normative and governance questions about acceptable levels of prediction‑driven intervention.
Thematic findings from the review's mapping of predictive analytics use cases and accompanying ethical/governance discussions across the 103‑item corpus.
Human–AI interaction issues—such as automation bias and shifting public servant roles—affect decision quality and legitimacy, creating a need for human‑in‑the‑loop processes.
Multiple empirical and theoretical contributions in the reviewed literature identified automation bias and role shifts; recommendation for human‑in‑the‑loop emerges from synthesis of these studies.
Legal frameworks like the EU GDPR provide a useful normative benchmark, but their protections do not automatically translate across jurisdictions; cross‑border research encounters gaps and asymmetries in enforcement and rights.
Normative and legal analysis contrasting GDPR principles with the Chilean/regional regulatory context and observed cross‑border data flow practices in the case study.
State-level divergence in AI-related regulation will create geographic heterogeneity in adoption costs and labor protections, potentially inducing firm and worker sorting across states and making national inference about AI’s effects more difficult.
Comparative policy review across states described in the commentary; inferential claim without presented empirical migration or firm-location data.
Regulatory uncertainty (rollbacks and a patchwork of rules) can raise compliance and political risk costs, causing some firms to accelerate private governance and self-regulation while causing others to delay investment or relocate activities.
Theoretical and policy reasoning based on review of regulatory signals and firm behavior literature; no empirical firm-level study or sample provided in the commentary.
Regulatory volatility and fragmentation will shape firms’ AI investment decisions, firms’ workplace practices (surveillance, task allocation), and the distributional consequences of AI for wages, employment and bargaining power.
Analytic synthesis linking observed policy instability and jurisdictional patchwork to likely firm responses and labor-market outcomes; conceptual inference rather than causal empirical evidence.
Economic outcomes of healthcare AI depend critically on governance design: policies and technical architectures (e.g., federated learning, certification standards, tiered risk management) will determine whether mixed open/proprietary ecosystems yield broad welfare gains or entrench inequities and concentrated market power.
High-level economic reasoning and synthesis of empirical and theoretical literature on governance, market structure, and technology adoption; prescriptive conclusion based on aggregated evidence rather than causal testing within the paper.
Reliable, well-integrated AI may raise clinical productivity and shift labor toward higher-value tasks, but misaligned deployments risk increased administrative burden (e.g., appeals, oversight).
Mixed evidence from pilot studies, observational reports, and stakeholder feedback synthesized in the paper; heterogeneity across settings and limited long-term outcome data noted.
Proprietary models concentrate costs into vendor payments and can potentially lower internal operational burden for providers.
Industry reports and economic synthesis comparing vendor-managed proprietary offerings with self-managed alternatives; based on reported vendor pricing models and operational roles.
Open-source lowers licensing fees but can shift costs toward in-house engineering, governance, and validation.
Cost-structure analyses and industry reports aggregated in the synthesis comparing licensing vs. internal operational costs across deployment models.
Open-source models show narrow but growing parity with proprietary models on some diagnostic tasks.
Synthesis of peer-reviewed comparative studies and benchmark reports indicating comparable diagnostic accuracy in limited tasks; authors note heterogeneity across studies and lack of long-term clinical trials.
Automation displaces some routine jobs but creates demand for roles in programming, data science, system maintenance, and higher‑order cognitive tasks.
Synthesis of labor‑market literature and sectoral case studies summarized in the review; relies on secondary empirical studies rather than new microdata analysis; sample sizes and study designs vary by referenced work.
Potential policy levers include mandatory provenance metadata, liability rules, taxes/subsidies to internalize harms, antitrust actions to limit concentration, and funding for public verification tools; each policy choice will shape incentives, innovation rates and market outcomes.
Policy options and scenario analysis summarized from legal/policy literature; presented as hypothetical levers rather than empirically tested interventions.
Economic returns may shift toward owners of data, model capacity and verification technology, while traditional creators may demand new compensation mechanisms (e.g., data-use royalties, collective licensing).
Conceptual economic analysis and synthesis of stakeholder- and rights-based literature in the narrative review.
Abundant synthetic media may erode the signaling value of standard digital content and create demand for authentication services, certification markets and premium 'human-made' labels.
Conceptual analysis grounded in signaling and market-for-authenticity literature reviewed in the paper (no primary WTP studies included).
Large productivity gains in content production could reduce marginal costs and compress prices for many creative goods, potentially displacing some human labor while raising demand for high-skill oversight, curation and novel creative inputs.
Economic reasoning and literature review on automation/productivity effects; no new empirical estimates presented (narrative inference).
Social acceptance is uncertain: some studies find people may rate AI-generated content equal or superior to human-created content, while proliferation of artificial media could also spur distrust or rejection of digital media.
Cited empirical studies on content perception and trust summarized in the narrative review (no primary data; exact sample sizes and studies vary by citation).
If consumers prefer AI-generated content, demand shifts could lower prices and increase consumption volume for certain media types; alternatively, trust erosion could reduce overall demand for digital content.
Reference to empirical studies with mixed results (paper notes 'some studies show higher ratings for AI content') and economic scenario modeling in the discussion; the paper does not report sample sizes or meta-analytic statistics.
Ambiguities in copyright and dataset licensing will affect value capture (original creators versus model operators) and may create new rent opportunities from provenance/authentication services or certified 'human-made' labels.
Legal and economic literature synthesized in the review, plus policy discussion; no empirical royalty or rent-share data provided.
Generative audiovisual models pose displacement risk for creative and production roles, but also create demand for new skills (prompt engineering, curation, verification) and complementarities in oversight and post-production.
Economic argumentation and citations to labor-impact literature and case examples in the review; no original labor-market empirical study or sample statistics provided.
Adoption frictions—integration costs, data access, reliability, and regulatory compliance—may slow diffusion of AI agents and create heterogeneity in economic value across firms and sectors.
Theoretical implication supported by observed orchestration and governance challenges in deployments; recommendation/interpretation rather than direct causal measurement.
Implementation heterogeneity (how guardrails, human oversight, and orchestration are configured) likely drives outcome variation across deployments.
Observed heterogeneity in Alfred AI deployments and stated limitation that configuration differences affect outcomes; based on deployment comparisons and qualitative analysis (sample size/configurations unspecified).
Net productivity gains may be smaller once indirect costs—governance, monitoring, error-correction, orchestration—are accounted for; standard productivity accounting should include these costs.
Conceptual argument supported by observational documentation of governance and monitoring burdens in deployments; no precise cost accounting reported in summary.
Autonomous agents are likely to substitute for routine, structured cognitive tasks while complementing higher-level managerial and strategic tasks, accelerating task reallocation within firms.
Synthesis of prior literature (generative AI productivity findings) and observational deployment patterns from Alfred AI indicating substitution of routine tasks and continued human involvement in oversight/strategy.
Realized productivity gains from AI agents are materially constrained by governance complexity, model reliability limits (errors, hallucinations, edge cases), orchestration challenges across tools/data/human teams, and continued need for human-in-the-loop oversight.
Qualitative operational impacts and deployment observations from Alfred AI implementations, documented frictions in policies, safety constraints, error handling, and orchestration; evidence drawn from observational deployments and operational logs.
Effectiveness and safety of AI agents require structured guardrails and human-in-the-loop designs; AI agents function as scalable cognitive infrastructure only conditional on such governance.
Synthesis of deployment experience and analysis of constraints; recommendation grounded in observed model reliability issues, governance complexity, and oversight needs from the Alfred AI experiments.
Deployment of AI agents shifts demand toward roles focused on oversight, orchestration, prompt/agent engineering, and governance, creating new types of labor that may offset some direct labor reductions.
Authors' inference based on observed need for human oversight and orchestration in deployments; not quantitatively measured in the study (no headcount or labor-share data reported).
Labor-market consequences will involve reallocation effects: routine-task automation, rising returns to managerial and technical skills, and potential within-firm wage dispersion.
Synthesis of labor economics theory and prior empirical work on automation; book recommends matched employer-employee panel studies to trace these effects but does not report such new panel results.
AI’s effects vary by industry, task composition, and firm capabilities; high-data, standardized-task sectors see faster, deeper impacts.
Cross-sector examples and theoretical arguments about task routineness and data intensity; calls for heterogeneity-aware empirical designs (e.g., difference-in-differences with staggered adoption).
Automation of routine tasks raises demand for cognitive, interpersonal, and technical skills; firms face reskilling needs and changing task allocation between humans and machines.
Task-level analytic framework and literature review on automation effects; book recommends empirical approaches (e.g., occupation and job-task data) to quantify these changes but does not present a single large empirical estimate.
Managers shift from routine decision execution to tasks involving oversight, interpretation, strategic design, and ethical stewardship of AI systems.
Qualitative case studies and literature review of task-level research; suggested task-analytic methods rather than reporting a specific empirical task dataset.
AI complements some researcher tasks (idea generation, analysis, writing) and substitutes others (routine editing, literature searches), changing skill demand and training priorities.
Stated under Labor Market Effects. Supported conceptually and likely by task-level studies or surveys; abstract doesn't cite specific empirical evidence or measurement details.
Impacts of AI adoption are broad, affecting individual researcher productivity, team workflows, and institutional outcomes in scholarly communication and digital scholarship.
Key Points summary. Basis likely includes mixed-methods evidence (surveys/interviews at individual and team levels, case studies, platform usage data) synthesized in the paper; abstract lacks detail on scope and samples.
Investment in governance and training is a necessary cost to realize sustained returns from generative AI; these costs influence adoption timing and the distribution of benefits.
Conceptual argument from the review supported by case examples and economic reasoning about complementary investments.
There is a risk of wage polarization: increased returns to AI‑complementary skills and potential downward pressure on wages for automatable tasks.
Theoretical synthesis drawing on economic models of skill‑biased technological change and early empirical observations; no definitive causal wage studies reported.
Generative AI will drive occupational reallocation by substituting routine cognitive tasks while complementing higher‑order cognitive and monitoring skills.
Theoretical labor economics arguments synthesized with early empirical examples; no large‑scale causal labor market study provided in the review.
TGAIF clarifies where GenAI acts as a complement (augmenting consultant capability) versus where it risks substitution.
Conceptual distinction and mapping presented in the TGAIF derived from practitioner accounts; theoretical/qualitative, not empirically quantified across tasks.
TGAIF implies reallocation of work away from GenAI‑suitable subtasks (routine synthesis, drafting, summarization) toward tasks where human judgment and client interaction add most value.
Based on authors' inductive analysis of practitioner interviews describing which subtasks firms consider suitable for GenAI and which require human oversight; qualitative, not quantitatively tracked reallocation.
Aligning consulting tasks with generative-AI capabilities via a Task–GenAI Fit (TGAIF) framework can unlock substantial efficiency gains while containing key risks (notably hallucinations and loss of skill retention).
Inductive framework developed from qualitative, interpretive interviews with practitioners at leading German management‑consulting firms. The abstract does not report sample size, interview protocol, or quantitative validation; evidence is based on practitioner reports and the authors' synthesis.
DAR implies changes to labor and contracting: reversible AI leadership reshapes task boundaries, demand for oversight skills, and should be reflected in contracts and procurement with explicit authority-reversal rules and audit obligations.
Theoretical/ normative argument in implications section; no empirical labor or contract data included.
AI substitutes for routine coding tasks but complements higher-order tasks such as system architecture, integration, and orchestration.
Interpretation from qualitative evidence at Netlight where practitioners used AI for routine chores while retaining control of higher-order design tasks; no quantitative task-time displacement data presented.
Human roles are shifting toward oversight, curation, specification, and orchestration of multiple AI components and tools.
Synthesized from practitioner descriptions and changing task allocations observed in the Netlight fieldwork (interviews/observations); no longitudinal measurement of role changes reported.
Short-run consumer gains from faster, cheaper service can be undermined by trust losses from hallucinations or perceived deception, reducing long-term consumer surplus.
Conceptual welfare analysis and cited case examples in the literature; no longitudinal consumer-surplus measurement provided in this review.
Conventional productivity metrics (e.g., handle time) may misstate value because they do not capture multi-dimensional impacts like quality and trust.
Conceptual critique and synthesis of measurement challenges discussed in the literature; no empirical measurement study presented in this review.
There is potential for substantial cost savings and throughput gains in repetitive, high-volume interactions, but these are offset by costs for integration, monitoring, and error remediation.
Industry case examples and conceptual cost–benefit reasoning aggregated in the review; the paper contains no new quantitative cost estimates or sample-based measurements.
Generative AI will substitute for routine service tasks while complementing skilled workers for escalations and complex problem solving, shifting labor demand toward supervisory and relationship-focused roles.
Economic and labor-market analyses synthesized in the review; projections are inferential and based on heterogeneous secondary sources, not primary labor-market experiments.
Full automation of customer service is suboptimal because persistent risks (hallucinations, contextual errors, lack of genuine empathy, integration complexity) remain; hybrid human–AI systems achieve the best outcomes.
Synthesis of documented failure modes and practitioner case examples from the literature; no primary experimental data or controlled trials in this review. Inference based on heterogeneous empirical reports and conceptual analyses.
Welfare effects of democratized access to AI-assisted ideation are ambiguous: access could democratize innovation but also amplify low-quality outputs and misinformation absent proper curation.
Theoretical discussion and empirical examples of misinformation/low-quality outputs from LLMs cited in the review; no comprehensive welfare accounting provided.