Evidence (1902 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Skills Training
Remove filter
Principal barriers to DT adoption include data quality and continuity problems at handover.
Thematic synthesis across reviewed literature reporting frequent issues with data quality and handover continuity between project phases.
Principal barriers to DT adoption include interoperability gaps and lack of standards.
Thematic findings from qualitative synthesis of the 160 reviewed studies (recurring theme across conceptual papers, case studies and pilots).
Unequal access to high-quality AI tools creates demand-side market failures and vendor concentration risks, justifying public intervention (subsidies, procurement tied to privacy/audit requirements).
Economic reasoning supported by literature on market failures and vendor dynamics; policy recommendations drawn from comparative analysis. No empirical market-share data provided.
Traditional signals (test scores, credentials) may lose reliability as AI assistance becomes widespread, which will alter estimates of skill endowments and returns to education.
Conceptual economic analysis and literature synthesis arguing how AI augmentation can change signaling and measurement; no empirical quantification presented in the paper.
Teachers currently lack sufficient preparedness (training, time, tools) to integrate AI into formative assessment and to interpret AI-informed evidence; addressing this is necessary for successful transition.
Review of education policy documents, literature on teacher professional development, and comparative case descriptions highlighting teacher-focused policies; no primary survey data reported.
Unequal access to AI amplifies existing achievement gaps and biases assessment outcomes, making equity a primary concern for AI-compatible assessment.
Conceptual and economic analysis drawing on literature about digital divides and policy documents; illustrated through comparative country cases showing variation in access and resources.
AI changes the production of student work (e.g., generative content, altered authorship), undermining traditional notions of student-authored artifacts used in assessment.
Conceptual analysis plus secondary literature on generative AI usage in education and observed capabilities of tools; case studies reference policy responses but no primary measurement of prevalence.
Standardized summative tests were designed for an environment without routine, external AI assistance; those design assumptions are breaking down.
Literature review and synthesis of assessment frameworks contrasted with descriptions of contemporary AI capabilities; conceptual argument rather than empirical test.
Conventional standardized, summative assessment is becoming increasingly misaligned with classroom reality because widespread student access to AI tools changes what, how, and where learning occurs.
Conceptual and policy analysis drawing on established assessment theory and literature on educational technology and AI; supported by comparative case studies of four countries using publicly available policy texts and secondary literature. No primary empirical/causal data or sample size reported.
Students raised concerns about ChatGPT producing factual errors, the risk of overreliance that could reduce independent thinking, and functional constraints of free ChatGPT versions.
Qualitative analysis of open-ended student survey responses; concerns consistently reported across responses in the sample of 254 students.
At the organizational scale, AI adoption is constrained and shaped by compliance requirements, formal policies, and prevailing norms.
Participants' accounts in workshops (n=15) noting compliance and policy considerations; thematic analysis classified these as organizational-level constraints.
The sphere + dislodgement-threshold material approximation may not capture all real-world mechanical and adhesive properties, limiting generalization.
Authors note/modeling limitation: summary explicitly states the material physics are approximated and may not capture all real-world properties; this is presented as a limitation rather than an empirical result.
Key technical and organizational risks include model brittleness, privacy and IP concerns in code generation (training-data provenance), and increased governance and QA burdens.
Literature review highlighting known risks and survey responses reporting practitioner concerns; no quantified incident rates provided.
Practitioners report barriers to adoption including integration costs, lack of trust/explainability, poor data quality, and skills gaps.
Thematic analysis / coding of open-ended survey responses and literature review identifying common adoption barriers; survey sample size not specified.
Expect diminishing returns from AI investments if parallel investments in organizational change and data governance are not made.
Synthesis of case evidence and theoretical argument: instances where additional AI investment produced limited marginal benefit absent organizational complements.
Legacy systems and siloed organizational structures produce persistent forecasting inaccuracies, operational disconnects, and constrained responsiveness.
Cross-case interview narratives documenting continued forecasting issues and operational misalignment in firms with legacy IT and functional silos.
Adoption complementarities (AI tools + developer skill + organizational processes) favor larger incumbents and well‑funded firms, possibly increasing concentration in tech sectors.
Theoretical argument about complementarities and returns to scale; illustrative examples; lacks firm‑level empirical testing.
In the near term, displacement risks concentrate on junior or highly routine roles; mobility and retraining will determine realized unemployment impacts.
Task automatability mapping indicating routine tasks more automatable and qualitative reasoning on labor mobility; no empirical unemployment projections.
Adoption will be heterogeneous: larger firms and well‑resourced teams will capture more gains earlier, producing competitive advantages.
Theoretical argument about adoption complementarities (AI tools + developer skill + organizational processes) and illustrative examples; no cross‑firm empirical analysis.
Extractive industries often deliver limited local employment and mainly generate rents rather than broad employment or skill spillovers.
Review of empirical studies and case evidence showing extractive FDI tends toward enclave production with low local hiring and limited upstream/downstream linkages; coverage varies by country and project.
FDI may increase within‑country wage inequality, especially when concentrated in extractive sectors or low‑skill activities.
Cross-study empirical results and theoretical arguments summarized in the review showing wage premia accruing to skilled workers and enclave effects in extractives; underlying studies vary in location, methods, and samples.
FDI may deepen labor market dualism: creating formal, higher‑paying jobs for a minority while many remain in precarious, low‑pay informal work.
Literature synthesis pointing to patterns where foreign investment produces enclave formal jobs while broader labor markets remain informal or precarious; evidence drawn from firm- and sector-level studies cited in the review.
A one standard-deviation increase in AI adoption lowers wages in the middle income quintile by 1.4%.
Panel of 38 OECD countries, 2019–2025; wage outcomes by income quintile using the AI Adoption Index and IV estimation; robustness checks reported.
Uneven inclusion in digital/AI deployments risks exacerbating digital divides and creating distributional harms.
Descriptive and case-based studies report differential access and uptake among demographic groups; limited causal quantification and varying measurement approaches across studies.
Limited auditability and explainability of AI systems increase trust and legitimacy risks.
Technical governance literature and case reports show challenges in model explainability and external audit; evidence is technical and illustrative rather than based on large-sample causal studies.
Inadequate regulatory frameworks raise privacy, accountability, and fairness concerns for AI in government.
Governance reviews and risk assessments documented in the literature highlight regulatory gaps and associated incidents/risks; empirical incident counts are not comprehensively tabulated in the review.
Procurement, budgeting rules, and siloed incentives discourage cross-cutting transformation and modular iterative deployments.
Policy and institutional analyses in the reviewed literature point to rigid procurement cycles, capital budgeting practices, and siloed funding as obstacles; examples and case narratives are provided but systematic quantification is limited.
Organizational resistance and fragmented coordination block integrated rollouts of cross-cutting digital reforms.
Qualitative case studies and governance analyses repeatedly identify intra-governmental silos, conflicting incentives, and change-resistance as implementation barriers; evidence is primarily descriptive.
Skills shortages (technical, managerial, data literacy) impede adoption and maintenance of digital and AI systems.
Multiple surveys, policy briefs and qualitative studies cited in the review report workforce capacity gaps; often based on targeted assessments or organizational audits rather than representative sampling.
Infrastructure deficits (connectivity, legacy systems) limit scale and reliability of digital/AI initiatives.
Recurring barrier documented across governance analyses and case studies; evidence includes reports of downtime, integration failures, and limited geographic reach; no unified cross-study sample provided.
Robust, locally appropriate data governance (privacy, interoperability, standards) is a public good that underpins trust and data-driven markets; weak governance raises risks of exclusion and foreign dependency.
Governance and policy literature synthesized in the review; conceptual arguments supported by examples but limited empirical evaluation in LMIC SME contexts.
Platform effects and supplier ecosystems associated with AI may create winner-takes-most market dynamics, so policy should monitor market concentration and enable competitive access to core AI services.
Literature on platforms and market structure combined with case examples; review notes potential for concentration but lacks broad causal studies quantifying effects in LMIC SME markets.
Fragmented or weak data governance (privacy rules, standards, interoperability, and trust) reduces SMEs’ ability to participate in data-driven markets and adopt AI.
Policy analyses and governance-focused studies in the review highlighting data governance weaknesses in LMICs and associated risks for SMEs; examples discussed rather than quantified nationally.
Failing to retrain health workers for AI will produce structural labor-market mismatches, slow adoption, and reduce realized economic benefits.
Labor-market analysis and workforce readiness findings from the narrative synthesis and Delphi inputs; argument is inferential based on observed skill gaps and adoption barriers in the reviewed literature.
Indonesia risks technological dependency on foreign vendors if domestic capability, data governance, and procurement are not strengthened.
Market and policy assessment from the review, including procurement analyses and discussion in supplementary national reports and Delphi studies; based on observed market structures and procurement practices identified in the literature.
Approximately 58.7% of the relevant Indonesian health workforce lacks the AI competence or literacy needed for safe, scalable adoption.
Workforce readiness estimate derived from reviewed workforce assessments, Delphi consensus studies, and national reports included in the narrative synthesis; the summary does not specify sample frames or exact survey instruments that produced the 58.7% figure.
Indonesia’s AI healthcare maturity score is approximately 52/100, trailing regional peers (example comparators: Singapore ≈ 92, Malaysia ≈ 78).
Benchmarking performed in the review against regional maturity catalogues and international standards (EU AI Act, Singapore, Australia); maturity scoring method referenced in the paper but detailed scoring rubric and underlying metrics not fully reproduced in the summary.
Widespread adoption of LLMs without adequate verification increases systemic cybersecurity risks with potential economic spillovers.
Synthesis of security incident case studies and risk analyses revealing vulnerabilities in generated code and potential downstream impacts.
Models lack deep contextual reasoning and may fail on tasks requiring long-term design thinking or deep domain knowledge.
Benchmark failures and user studies in the reviewed literature demonstrating degraded performance on complex architectural/design tasks and domain-specific reasoning problems.
Use of these tools can mask gaps in foundational computational skills among novices.
Pedagogical case studies and assessments indicating reliance on AI can produce superficial solutions and lower demonstrated understanding of core concepts.
This generation–verification mismatch produces a chronic bottleneck in development processes.
Analytic diagnosis and behavioral reasoning in the paper (design principles and system analysis); no empirical testing or simulation results provided.
AI-assisted software development creates a persistent structural imbalance: generation throughput (machine-produced code, tests, docs) outpaces human verification capacity.
Conceptual/theoretical argument and systems/architectural modeling in the paper; no empirical measurement, no sample size, no field data reported.
Overreliance on generative AI risks eroding worker critical thinking and loss of tacit expertise.
Conceptual arguments supported by observational reports and theoretical concerns in the literature synthesis; limited empirical evidence cited.
Security vulnerabilities and IP leakage create negative externalities; absent internalization, social costs (breaches, legal disputes) may rise.
Security analyses, documented incidents, and economic externality reasoning synthesized from the literature; empirical quantification of social cost is limited.
Generated code may incidentally reproduce copyrighted or licensed snippets from training data.
Analyses detecting verbatim or near-verbatim reproductions of licensed/copyrighted code in model outputs in selected tests and audits; evidence heterogeneous and depends on prompts and model/data.
Outputs often lack deep, project-level contextual reasoning (e.g., design tradeoffs, architecture constraints).
Qualitative failure-mode analyses, user studies, and benchmark tasks showing limitations in system-level reasoning and context-aware design decisions; evidence from short-horizon labs and case studies.
There is a risk of shallow learning if learners over-rely on AI outputs without understanding fundamentals.
Educational studies and observational analyses indicating reduced engagement with underlying concepts for some learners using AI assistance, plus qualitative reports from instructors; studies often short-term.
Existing extrapolation‑based projection systems understate AI’s nonlinear, spillover, and augmentation effects and miss differential impacts across occupations, industries, regions, and demographic groups.
Theoretical argument and literature-based reasoning in the paper; no quantitative demonstration comparing extrapolation systems to the proposed approach.
Traditional BLS projection methods are insufficient for forecasting labor market changes driven by rapid AI adoption.
Conceptual critique and argumentation in the paper; no empirical evaluation or comparative forecast error statistics provided.
Rapid post-2020 advances in AI (LLMs and multimodal models) have already rendered some pre-2020 profession-level conclusions obsolete by 2025.
Argument based on observed acceleration in AI capabilities after 2020 (LLMs, multimodal systems) discussed in the paper; evidence is temporal comparison of the state of capabilities and the applicability of older exposure indices rather than a single empirical re-test of all prior predictions.