Evidence (2432 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Labor Markets
Remove filter
Agent-based models (ABMs) are needed to simulate micro-to-macro dynamics of token taxes because standard representative-agent or DSGE models may miss heterogeneity, network effects, and path dependence.
Methodological argument in the paper advocating ABMs; no ABM results included (proposal only).
Black-box token verification (tamper-evident consumption tokens or receipts tied to API calls) can prove taxable consumption without full model inspection.
Technical proposal for cryptographic/ledgered receipts described in the paper; no prototype, security analysis, or empirical tests provided.
A staged audit pipeline—black-box token verification, norm-based tax rates, then white-box audits—provides a feasible path to design and evaluate token taxes.
Proposed enforcement architecture described in the paper (conceptual design); no deployment or simulation results presented.
Token taxes can be enforced using existing compute-governance and commercial billing infrastructure (API billing, cloud metering, hardware telemetry, attestation).
Technical architecture discussion proposing use of existing billing and telemetry systems; no implementation or pilot data provided.
Compared with robot- or FLOP-based taxes, token taxes better capture where AI-generated value is realized.
Analytic comparison in the paper arguing tokens map to user-facing consumption while FLOP/robot taxes map to inputs; conceptual reasoning rather than empirical test.
Task‑based, dynamic exposure measures and real‑time data enable earlier detection of displacement risks and reallocation needs than static, occupation‑level extrapolations.
Conceptual argument and proposed architecture; no empirical timing comparison or lead-time statistics provided.
LLMs can be used to score task automation/augmentation plausibility and to detect emergent tasks.
Methodological proposal describing use of LLMs for semantic mapping/scoring of tasks; no empirical validation or accuracy metrics for LLM task scoring provided in the paper.
Modeling nonlinearity (threshold adoption, network spillovers, complementarities) and path dependence in adoption dynamics is necessary rather than relying on linear extrapolation.
Theoretical argument and model suggestions (S‑curve diffusion, agent-based models) in the paper; no empirical comparison demonstrating superior performance provided.
Applying causal inference methods (difference‑in‑differences, synthetic controls, instrumental variables, structural counterfactuals) can distinguish automation (task substitution) from augmentation (productivity/role change) and estimate net employment effects.
Methodological recommendation with examples of applicable identification strategies; no specific empirical applications or results reported in the paper.
Integrating multiple data streams (CPS, LEHD/LODES, UI wage records, administrative microdata, job ads, occupational manuals, enterprise adoption surveys) yields richer gross‑flows and skills measurement than using single data sources.
Proposed data-integration strategy and references to candidate datasets; no empirical demonstration or quantified improvement in measurement presented.
A dynamic Occupational AI Exposure Score (OAIES) can quantify exposure at the task level using LLMs, job‑task matrices (e.g., O*NET), and real‑time job ad / workplace data to capture evolving capability of AI systems.
Methodological description of OAIES construction (mapping tasks to occupations, LLM scoring, weighting by time use/criticality); no empirical implementation or validation data presented in the paper.
Measurement and forecasting should move away from occupation-level forecasts toward task-level, continuously updated indicators linked to real-world adoption measures (firm purchases, API usage, procurement).
Recommendation in the paper motivated by rapid changes in AI capabilities and limitations of static indices; evidence basis is methodological argument and examples of richer adoption measures rather than a quantified evaluation of forecast improvements.
Policy should prioritise flexible reskilling and retraining programs targeted at high-risk tasks and low-skilled workers, informed by task-level exposure maps.
Policy implication recommended by the paper drawing on distributional findings (higher displacement risk for low-skilled tasks) and the availability of task-level exposure indices; evidence basis combines empirical pattern synthesis and normative recommendation rather than an RCT or program evaluation.
Think tanks and international organisations are emphasising scenario planning with differing adoption initial conditions to inform reskilling and labour-market policy.
References to policy and scenario work by organisations named in the paper (TBI, IPPR, IMF, TBI 2024; IPPR 2024; Korinek 2023); evidence basis is published scenario reports and policy papers rather than experimental data.
Labor complementarities with agentic AI will shift resources toward oversight, interpretation, and coordination roles rather than routine task execution.
Economic and organizational reasoning; literature synthesis on skill complementarities; no empirical labor-market data analyzed in the paper.
Principal–agent contracting frameworks must be extended to account for evolving agent objectives and open-ended action spaces; contracts should be dynamic and include continuous renegotiation and monitoring.
Theoretical extension and recommendations based on economic reasoning; proposed formal models for future work.
Projection congruence — alignment of forecasts/plans across heterogeneous agents — becomes a central metric for assessing alignment in agentic human–AI teams.
Conceptual modeling and proposal in the paper; introduced as a new measurable construct (projection congruence indices) for future empirical work.
Continuous human-in-the-loop oversight, monitoring, and retraining are required to maintain quality and prevent model drift.
Practitioner reports and conceptual literature synthesized in the review advocating monitoring and retraining; no longitudinal empirical study provided here.
Transparent disclosure to customers about AI involvement helps preserve trust.
Conceptual analyses and referenced empirical/regulatory discussions in the literature aggregated by the review; this paper presents no new experimental evidence on disclosure effects.
Hybrid designs that automate low-risk, high-volume tasks while routing complex, judgment-sensitive cases to humans produce the best operational outcomes.
Inferred best-practice from aggregated empirical studies, industry examples, and conceptual reasoning; no controlled comparative trials presented in this review.
Agent augmentation via suggested responses, summarization, and information retrieval improves agent productivity.
Aggregated evidence from prior empirical research and practitioner reports cited in the review; no new measurements or sample sizes presented here.
Generative AI enables personalization at scale through automated tailoring of messaging and recommendations.
Qualitative synthesis of empirical studies and industry reports showing automated personalization use-cases; no systematic effect-size estimates or new quantitative data in this review.
Generative AI provides 24/7 availability and cost-effective scaling of routine interactions.
Industry case examples and prior empirical studies aggregated in the review; no original data or quantified sample sizes provided in this paper.
Generative AI can materially transform customer service and strategic communication by enabling continuous automation, scalable hyper-personalization, and effective agent augmentation.
Nano review: qualitative aggregation and synthesis of existing empirical studies, industry case examples, and conceptual analyses. No novel primary data or sample size; conclusion drawn from heterogeneous secondary sources and practitioner reports (not a systematic meta-analysis).
There is a need for standards around evaluation, bias mitigation, provenance, and accountability in AI-assisted ideation and design.
Policy recommendation motivated by documented biases, errors, and provenance issues in the reviewed studies; grounded in the synthesis's critique of existing practice.
There will likely be complementarity-driven increases in demand for evaluative, integrative, and domain-expert roles (curators, synthesizers, implementation experts).
Inference from task-level studies and economic reasoning about complementarities between AI generative capability and human evaluative skills; empirical labor-market evidence is limited in the reviewed literature.
Lower search and idea-generation costs enabled by LLMs may speed early-stage R&D and increase the gross flow of candidate innovations.
Theoretical economic interpretation supported by empirical findings of increased idea volumes in experimental/field studies summarized in the review; no long-run causal firm-level evidence presented.
Generative AI accelerates early-stage hypothesis and prototype development by providing scaffolded prompts and procedural suggestions.
Applied case evidence and experimental studies summarized in the review showing reduced time or increased productivity in early-stage experimental/design tasks when using LLM assistance; no pooled effect size presented.
Empirical studies document that AI-assisted tools can help break cognitive fixation and generate cross-domain analogies.
Cited experimental tasks and lab studies in the literature showing higher incidence of analogical or cross-domain suggestions from LLMs and improvements on fixation-related task metrics; heterogeneity across tasks and measures.
Generative AI provides scaffolded, structured support that aids systematic hypothesis formation, prototyping steps, and decomposition of complex problems.
Review of design/ideation studies and applied case evidence where LLMs produced stepwise plans, decomposition prompts, or hypothesis scaffolds; evidence drawn from multiple short-term experimental and applied studies, sample sizes and exact designs vary by study.
Generative models rapidly produce many candidate ideas, analogies, and associative prompts that help overcome cognitive fixation.
Synthesis of experimental ideation and design studies reporting increases in number of ideas and examples of reduced fixation when participants used LLM outputs; heterogeneous sample sizes across cited studies (not reported in review).
Generative AI can raise per-worker productivity for tasks involving brainstorming, drafting, and prototyping, but realized gains depend on downstream filtering and implementation costs.
User studies showing higher output on specific tasks (brainstorming/drafting), combined with qualitative reports of filtering/implementation effort; many studies measure immediate task output but not net realized productivity after implementation.
Generative AI can increase creative output in both lab and field tasks as judged by external raters.
Controlled experiments and field studies reporting higher judged creativity/novelty scores for AI-assisted outputs versus controls; judged creativity/novelty is typically assessed by human raters using rubric-based scoring.
AI assistance helps people overcome fixation and produces cross-domain analogies that they might not generate alone.
Experimental studies and qualitative analyses documenting reductions in fixation effects and increases in cross-domain analogical suggestions when participants use generative models.
Generative AI supports systematic problem breakdown and early-stage prototyping, accelerating hypothesis generation and prototype development.
Field case studies of AI-supported prototyping and lab/user studies reporting reduced time-to-prototype and generated hypotheses; measures include time-to-prototype and user-reported usefulness.
Generative AI boosts ideational fluency—the quantity and diversity of ideas produced in brainstorming tasks.
Controlled experiments and user studies measuring number and diversity of ideas with and without AI assistance; typical study designs compare participant idea counts/uniqueness across conditions (note: many studies use small or convenience samples).
When used as a 'cognitive co-pilot' that expands the solution space and challenges assumptions while humans curate and evaluate, generative AI generates economic value.
Inferred from experimental and field findings showing increased idea quantity/diversity and faster prototyping combined with qualitative studies showing human curation is needed; economic interpretation drawn from the review rather than direct macroeconomic measurement.
Generative AI serves a dual cognitive role: (1) a high-volume catalyst for divergent idea generation and cross-domain analogy-making, and (2) a structured assistant for deconstructing complex problems and scaffolding hypotheses and prototypes.
Synthesis of controlled experiments, lab studies, field case studies, and qualitative analyses summarized in the review; evidence includes measures of idea fluency/diversity, examples of analogy production, and observations of AI-assisted problem decomposition in prototyping tasks. (Note: underlying studies are heterogeneous and often short-term or convenience samples.)
Agent augmentation (drafting replies, summarizing histories, suggesting actions) raises frontline productivity and can improve response consistency.
Pilot deployments and internal A/B tests cited that measure time saved by agents and improvements in draft quality/consistency; mostly short-run and firm-specific reports.
Hyper-personalization at scale can increase relevance of responses and customer engagement when fed high-quality signals.
Case studies and pilot deployments that applied personalization signals (customer history, behavioral data) and reported improved relevance/engagement metrics; evidence conditional on availability and quality of signals and largely non-randomized.
24/7 automation reduces routine handling time and operational costs for simple, repetitive queries.
Operational deployments and pilot studies reporting reduced handling times and cost-per-interaction for routine queries; some vendor-supplied before/after or A/B comparisons, but heterogeneous measurements and limited randomized evidence.
With appropriate policies and ecosystem building, AI offers strategic opportunities for 'leapfrogging' in service delivery (for example, healthcare diagnostics and precision agriculture) that can raise productivity and welfare.
Synthesis of case studies and prior empirical work showing promising AI applications; the assertion remains inferential and the paper calls for pilots and empirical validation.
Investing in human capital—technical skills, digital literacy, and institutional capacity—is critical for African actors to capture value from AI and to design culturally aligned systems.
Policy and academic literature synthesis linking human capital investment to technology adoption and innovation; no primary training program evaluation in the paper.
Context‑sensitive interventions—stronger governance, capacity building, multi‑stakeholder collaboration, and locally tailored strategies—are necessary to steer AI toward inclusive outcomes in Africa.
Policy and literature synthesis recommending interventions; recommendations are normative and inferential without empirical pilots in this paper.
AI adoption in Africa is already transforming multiple sectors (healthcare, finance, agriculture, education, industry, governance) and has the potential to improve productivity, service delivery, and decision-making.
Desk-based literature synthesis of prior empirical studies, policy reports and case studies; no primary data or field experiments reported in this paper.
Audit cycles and inter-rater reliability studies should be used to improve assessment validity.
Suggested under Evaluation/Research Designs and Implementation Artifacts: the paper recommends systematic audits and inter-rater reliability studies as validity checks. This is a recommended practice, not an empirically validated result within the paper.
Better competency mapping and standardized, machine-readable program outputs facilitate automated matching platforms and reduce search/matching costs in AI labour markets.
Stated in Implications for AI Economics: the paper links machine-readable competency outputs to improved labour-market matching. This is a theoretical implication; no empirical matching-cost estimates are presented.
The approach increases traceability and compliance readiness, facilitating audits and regulatory verification.
Paper cites audit-ready documentation, systematic audits, and versioned curriculum artifacts as outputs and recommends audit cycles and inter-rater reliability studies. This is an asserted benefit without reported empirical testing.
IT integration is necessary for documentation, traceability, and continuous monitoring of curriculum artifacts.
Listed among core components and implementation artifacts (version-controlled documentation, traceability logs, IT-backed traceability). Support is prescriptive and conceptual rather than empirical.
Logical modelling tools (logigrams and algorigrams) support lesson planning and audits by formalising decision rules and automated workflows.
Described as a core component and implementation artifact; paper explains process modelling using logigrams/algorigrams to formalise instructional algorithms and audit workflows. No empirical validation provided.