Evidence (1286 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Inequality
Remove filter
Practical barriers—software access, available datasets, and lab time—limit experiential learning that builds AI competency.
Survey items listing barriers to AI learning and training; thematic coding of open responses highlighting software, dataset, and scheduling constraints.
Respondents cite limited opportunities for applied, project-based learning with AI tools; where AI appears in curricula, coverage is more theory-oriented than hands-on.
Quantitative items and open-ended responses about types of training and curricular integration; thematic analysis of qualitative data indicating prevalence of theory-focused instruction versus applied opportunities.
Many institutions lack clear, consistent, or context-sensitive policies for AI use in learning, assessment, and academic integrity.
Survey questions about the presence and clarity of institutional AI policies and thematic coding of open-ended responses reporting policy gaps; descriptive summaries across respondents.
Educators frequently report lower confidence in teaching AI-relevant skills than students report in using AI tools, reducing instructional capacity.
Survey items measuring self-reported competency/confidence for educators (teaching) and students (using); comparative descriptive analysis across roles within the >600 participant sample.
Rural digital divides mean AI benefits will be unevenly distributed; models trained on digitally-rich urban records could bias resource allocation away from rural trainees.
Analytical/risk assessment in the paper noting distributional risks; no empirical bias measurement presented.
Key disadvantages and barriers to the proposed digital modernization are administrative backlogs, rural infrastructure deficits, and qualification fragmentation.
Identified limitations in the paper's diagnostic section; based on conceptual review and sector knowledge rather than quantified barrier assessment.
Rural constraints (limited electricity, limited ICT access, and fewer training centers) reduce inclusion of rural trainees in vocational-to-engineering pathways.
Qualitative discussion and domain knowledge within the paper; no field survey or representative sample quantifying the rural access gap.
Fragmentation and overlap across vocational and technical qualifications create discontinuities that impede career progression.
Conceptual analysis of qualification frameworks and mapping of vocational/technical curricula; no empirical measurement of career outcomes or frequencies of pathway breakdowns.
Administrative irregularities and backlogs exist in SAQA/NATED ratification processes, including suspension or deregistration actions carried out without due process.
Institutional review and diagnostic claims in the paper; assertions drawn from document/process analysis rather than audited data or quantified case series (no sample size provided).
Misalignment between hands-on technical training (artisan-level skills) and formal institutional certification (SAQA/NATED/NCV/SETA) is blocking vocational-to-engineering career progression.
Qualitative institutional review and conceptual systems analysis presented in the paper; no empirical dataset, no sample size, argumentation based on policy/process review and domain knowledge.
Trade policy (trade openness) should be modeled as a moderating factor when estimating technology-driven urban outcomes because openness can dampen local price effects of digital trade.
Inference based on the reported negative moderation effect of trade openness on the digital-trade → house-price relationship from interaction regressions.
Greater trade openness weakens (attenuates) the positive effect of digital trade on city-level house prices.
Interaction terms between digital trade and a measure of trade openness in the panel regressions; reported negative moderation effect (exact openness measure and sample details not provided).
Policy and regulatory vacuum (data governance, interoperability, safeguards) limits scale and inclusive diffusion of AI in agriculture.
Authors' thematic finding from reviewed literature and institutional reports noting weak policy frameworks and governance gaps.
Limited digital literacy and human capacity among smallholders is a key barrier to adoption and effective use of AI solutions.
Multiple studies and reports in the review documenting low digital literacy, limited extension capacity, and training needs among target users.
Scalable adoption of AI in developing-country agriculture is constrained by infrastructure gaps (connectivity, power, data platforms).
Thematic synthesis across reviewed studies and reports identifying recurring infrastructure constraints limiting deployment and scale-up.
Unequal access to high-quality AI tools creates demand-side market failures and vendor concentration risks, justifying public intervention (subsidies, procurement tied to privacy/audit requirements).
Economic reasoning supported by literature on market failures and vendor dynamics; policy recommendations drawn from comparative analysis. No empirical market-share data provided.
Traditional signals (test scores, credentials) may lose reliability as AI assistance becomes widespread, which will alter estimates of skill endowments and returns to education.
Conceptual economic analysis and literature synthesis arguing how AI augmentation can change signaling and measurement; no empirical quantification presented in the paper.
Teachers currently lack sufficient preparedness (training, time, tools) to integrate AI into formative assessment and to interpret AI-informed evidence; addressing this is necessary for successful transition.
Review of education policy documents, literature on teacher professional development, and comparative case descriptions highlighting teacher-focused policies; no primary survey data reported.
Unequal access to AI amplifies existing achievement gaps and biases assessment outcomes, making equity a primary concern for AI-compatible assessment.
Conceptual and economic analysis drawing on literature about digital divides and policy documents; illustrated through comparative country cases showing variation in access and resources.
AI changes the production of student work (e.g., generative content, altered authorship), undermining traditional notions of student-authored artifacts used in assessment.
Conceptual analysis plus secondary literature on generative AI usage in education and observed capabilities of tools; case studies reference policy responses but no primary measurement of prevalence.
Standardized summative tests were designed for an environment without routine, external AI assistance; those design assumptions are breaking down.
Literature review and synthesis of assessment frameworks contrasted with descriptions of contemporary AI capabilities; conceptual argument rather than empirical test.
Conventional standardized, summative assessment is becoming increasingly misaligned with classroom reality because widespread student access to AI tools changes what, how, and where learning occurs.
Conceptual and policy analysis drawing on established assessment theory and literature on educational technology and AI; supported by comparative case studies of four countries using publicly available policy texts and secondary literature. No primary empirical/causal data or sample size reported.
Students raised concerns about ChatGPT producing factual errors, the risk of overreliance that could reduce independent thinking, and functional constraints of free ChatGPT versions.
Qualitative analysis of open-ended student survey responses; concerns consistently reported across responses in the sample of 254 students.
Biased or unrepresentative AI outputs produce negative externalities, including maladaptation and inefficient investments in vulnerable regions.
Conceptual analysis and illustrative cases linking misleading model outputs to maladaptive decisions; the paper notes risks rather than providing quantified incidence or cost estimates.
Returns to scale in compute and data favor incumbents; without intervention this dynamic can entrench inequality in the global climate-information market.
Economic theory of returns to scale combined with observed compute concentration; no empirical elasticity or returns-to-scale estimates provided.
Concentration of compute and model development creates market power for Northern institutions and companies, likely leading to unequal pricing, control over standards, and capture of high-value climate services.
Descriptive mapping of concentration plus economic analysis of market structure and returns to scale; illustrative rather than quantitatively proven across markets.
Rapid AI adoption without a shift from model-centric to data- and equity-centric development risks producing systematically worse performance and misleading recommendations for the most climate-vulnerable, data-sparse regions.
Synthesis of domain-specific case studies (weather/climate, impact models, LLMs) and conceptual causal tracing demonstrating how infrastructure asymmetry can degrade outputs in vulnerable regions; evidence illustrative rather than causal-estimate based.
Large language models (LLMs) that rely on dominant, textualized climate knowledge tend to foreground Northern epistemologies and marginalize local or indigenous knowledge, reinforcing biases in climate narratives and recommendations.
Case studies and analysis of training-corpus composition and output examples illustrating the dominance of Northern textual sources and examples of sidelining local knowledge; no large-scale audit results provided.
In climate impact modelling, sparse and unrepresentative exposure and vulnerability data combined with inadequate validation generate high uncertainty and risk of misleading interventions and maladaptation in vulnerable locales.
Targeted case studies and literature synthesis showing gaps in exposure/vulnerability datasets and validation failures; argument is illustrated rather than quantified across all systems.
In weather and climate modelling, historically and spatially biased observational data produce systematic performance gaps in under-observed tropical and low-income regions, reducing forecast fidelity where adaptive capacity is lowest.
Comparative, domain-specific case studies and literature review documenting observational data sparsity and illustrative empirical performance gaps; no single cross-system statistical estimate provided.
The geographic concentration of compute and model development creates path dependence: model design, training datasets, and validation reflect Northern priorities and contexts.
Conceptual analysis supported by cross-disciplinary synthesis and illustrative case studies showing dataset selection, validation practices, and model design choices aligned with Northern contexts rather than global representativeness.
Numerical simulations using calibrated parameter sets produce phase diagrams and time-paths that show when gradual adjustment transitions into explosive demand collapse and financial stress under different combinations of capability growth, diffusion speed, and reinstatement rate.
Calibrated numerical simulation experiments described in the methods and results sections, using FRED, BLS, and occupational AI-exposure inputs and varying key model parameters.
Because consumption is concentrated and top incomes have high AI exposure, shocks to top-income labor/income disproportionately affect aggregate consumption and thereby threaten private credit and mortgage markets — the paper maps plausible exposures to roughly $2.5 trillion of global private credit and about $13 trillion of mortgages.
Calibration exercise linking household-level demand shocks (based on concentration and AI-exposure mapping) to aggregate credit and mortgage aggregates; reported dollar-amount mappings in the paper's scenarios.
Top-quintile households are also the cohort with the highest measured AI exposure (i.e., incomes/occupations most exposed to AI substitution), increasing the concentration of AI-driven demand risk.
Mapping occupation-level AI-exposure indices to household income quantiles using BLS occupation employment and wage data; used in calibration and scenario analysis.
Intermediation collapse: AI agents reduce information frictions and automate advice/coordination tasks, compressing intermediary margins toward logistics/execution costs and repricing business models across SaaS, payments, consulting, insurance, and financial advisory, with knock-on effects for firm valuations and collateral values that underpin credit markets.
Modeling of intermediary margins and information rents within the macro-financial framework; calibrated scenarios and sectoral discussion mapping margin compression to valuation and collateral effects.
Ghost GDP: AI output that replaces labor-intensive output can create a wedge between measured GDP (which may rise) and consumption-relevant income (which can fall) because a declining labor share reduces monetary velocity absent proportionate transfers — producing hidden demand shortfalls.
Formalization in the paper linking labor share to monetary velocity and thus to consumption-relevant income; calibration using FRED macro time series and monetary-aggregate/velocity proxies.
When firms rationally substitute AI for labor, aggregate labor income can fall and lower demand, which accelerates further AI substitution — a 'displacement spiral' whose net feedback is either self-limiting (convergent) or explosive (runaway adoption + demand collapse) depending on AI capability growth rate, diffusion speed across firms/sectors, and the reinstatement rate (rate at which new paid human roles or demand reappear).
Formal model derivations that identify key parameters and inequalities separating convergent vs explosive regimes; calibrated simulations that vary capability growth, diffusivity, and reinstatement elasticity to produce different phase outcomes.
Rapid AI adoption can create a macro-financial stress scenario not primarily through productivity collapse or existential risk but via a distribution-and-contract mismatch: AI-generated abundance reduces the need for human cognitive labor while institutions (wage contracts, credit, consumption patterns, financial intermediation) remain anchored to the scarcity of human cognition, producing a self-reinforcing downward spiral in labor income, demand, and intermediary margins that can tip into an explosive crisis unless offset by sufficiently fast reinstatement of human-paid demand or deliberate policy/market responses.
Analytical macro-financial model coupling firm-level substitution decisions, aggregate demand mapping, and financial-sector balance-sheet propagation; calibrated numerical simulations using U.S. macro time series (FRED), BLS occupation-level employment and wages, and published occupation-level AI-exposure indices; phase diagrams and scenario time-paths reported in the paper.
Introducing ‘agent capital’ (AI that lowers coordination costs) reduces coordination costs inside firms (coordination compression).
Definition and central assumption of the paper's formal task-based model; analytical setup assumes agent capital parametrically reduces coordination frictions.
Extractive industries often deliver limited local employment and mainly generate rents rather than broad employment or skill spillovers.
Review of empirical studies and case evidence showing extractive FDI tends toward enclave production with low local hiring and limited upstream/downstream linkages; coverage varies by country and project.
FDI may increase within‑country wage inequality, especially when concentrated in extractive sectors or low‑skill activities.
Cross-study empirical results and theoretical arguments summarized in the review showing wage premia accruing to skilled workers and enclave effects in extractives; underlying studies vary in location, methods, and samples.
FDI may deepen labor market dualism: creating formal, higher‑paying jobs for a minority while many remain in precarious, low‑pay informal work.
Literature synthesis pointing to patterns where foreign investment produces enclave formal jobs while broader labor markets remain informal or precarious; evidence drawn from firm- and sector-level studies cited in the review.
A one standard-deviation increase in AI adoption lowers wages in the middle income quintile by 1.4%.
Panel of 38 OECD countries, 2019–2025; wage outcomes by income quintile using the AI Adoption Index and IV estimation; robustness checks reported.
Loss of control over research data impedes local capture of value (knowledge, IP, downstream services) and can create externalities when data are repurposed or commercialized without equitable benefit sharing.
Conceptual argument grounded in case observations about data flows and provider practices; no quantitative measures of value capture provided.
Dominant AI/cloud providers become de facto gatekeepers of data processing and storage; researchers and institutions, particularly in lower‑capacity jurisdictions, have limited bargaining power to enforce data‑sovereignty or transparency terms.
Mapping of third‑party dependencies and interview/observational evidence of institutional procurement constraints in the Chile case; normative discussion of market power implications.
Algorithmic opacity and cross‑border regulatory fragmentation raise monitoring, compliance, and contractual costs for collaborative research, effectively increasing the transaction costs of data‑intensive science.
Analytical inference from qualitative findings (opacity, legal fragmentation) and normative economic reasoning presented in the implications section; no quantitative transaction‑cost measurement reported.
Inequalities in infrastructure (local compute, storage, institutional procurement power) amplify these problems: researchers in weaker jurisdictions face higher risks and fewer mitigation options.
Case study observations about local infrastructure capacity, procurement practices, and institutional constraints in Chile; qualitative reports of limited mitigation choices.
Rather than shifting liability away from researchers, AI systems increase researchers' ethical responsibilities: researchers must assess third‑party tools, negotiate data flows, and manage risks despite having limited contractual leverage.
Qualitative interviews and institutional observations reporting researchers' roles in assessing tools and managing data flows; normative analysis of accountability responsibilities in the case study.
Algorithmic opacity (hidden models, undocumented data flows, proprietary cloud stacks) reduces researchers' ability to control or even know how participant data are used, transferred, or monetized.
Interview data and mapping of third‑party dependencies showing opaque provider practices and limited transparency about model/data flows in the Chile case study.
Everyday AI services used in research introduce new, diffuse points of data capture and processing that complicate informed consent and privacy management.
Observations and documented mappings of tool use and data flows (e.g., transcription services, cloud platforms, meeting assistants) reported in the case study; supported by qualitative interviews with researchers/administrators.