Evidence (1902 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Skills Training
Remove filter
Existing research largely focuses on general computer literacy and lacks precise measurement of the economic returns to specific vocational digital skills.
Paper's literature review and motivating statements (qualitative assessment of prior studies; no quantitative meta-analysis reported in the excerpt).
We did not observe significant differences between using Gemini (free or paid) and not using Gemini in terms of secure software development.
Statistical comparison of code-security outcomes across the three experimental groups (no AI, free Gemini, paid Gemini) in the n = 159 participant sample; the paper reports no statistically significant group differences.
Collaborative ability is distinct from individual problem-solving ability.
Model-based estimates from the Bayesian IRT framework that separately parameterize collaborative ability and individual problem-solving ability, with results indicating they are separable constructs (analysis on n = 667 benchmark data).
Early evidence from nationally representative datasets shows limited aggregate wage and employment changes following GenAI's emergence.
Empirical analyses referenced in the paper that use nationally representative population-level datasets (specific datasets and sample sizes not provided in the excerpt).
This study empirically tests a theoretically acknowledged but rarely tested relationship (AI adoption → performance conditional on structural constraints) in an emerging-economy setting.
Literature gap claim supported by the authors' review and execution of an empirical test using survey data from 280 Tunisian SMEs and PLS-SEM.
Institutional conditions do not exert a significant moderating influence on the relationship between AI adoption and firm performance in this sample.
PLS-SEM moderation tests on the 280 Tunisian SMEs found the institutional-environment moderator to be non-significant.
Logistics efficiency does not mediate (fails to fulfill) the anticipated role in transmitting AI's effects to supply chain stability.
Mechanism/mediation tests in the DML analysis on the 45 Chinese listed SEs (2012–2023) indicate no significant mediation via logistics efficiency.
The Photo Big 5 is only weakly correlated with cognitive measures such as test scores.
Correlation/associational analysis between Photo Big 5 trait scores and cognitive measures (e.g., test scores) reported for the MBA graduate sample.
The study presents an advanced systematic ranking of I4.0 adoption barriers in the Thai automotive industry.
Paper outputs a ranked list of barriers produced by the integrated Fuzzy BWM-PROMETHEE II-DEMATEL framework; full ranked list and quantitative ranks not included in the supplied summary.
The study explores the influence of AI on HRM practice specifically within top IT companies.
Scope statement in the paper: empirical study involved HR professionals from various (described as top) IT firms. The summary does not supply the list of companies or sampling criteria.
In the sentiment-analysis task, those individual differences do not produce human–AI complementarity: the joint performance of humans and AI did not exceed that of either alone.
Empirical finding reported from the preregistered sentiment-analysis experiment showing no complementarity effect (joint human-AI performance ≤ best individual performance). (Statistical tests and sample size not included in the excerpt.)
Self-generated (model-authored) Skills provide no average benefit.
Comparison of three evaluation conditions (no Skills, curated Skills, self-authored Skills) across SkillsBench. Averaged pass-rate deltas show that model-authored Skills do not increase average pass rate relative to baseline; analysis used 7,308 trajectories over 86 tasks and 7 agent–model configurations.
AI will not cause permanent mass unemployment at the aggregate level.
Analytical argument and literature synthesis using labor-economics theory (Skill-Biased Technological Change and structural transformation). No primary microdata, no stated empirical identification strategy or sample size in the paper (methodology appears to be theoretical and sectoral synthesis).
Empirical evaluation is needed on how AI-induced productivity gains translate into aggregate demand and labor absorption.
Identified research priority in the paper, based on theoretical uncertainty about demand-side labor absorption and lack of conclusive empirical evidence.
AI will not mechanically cause permanent mass unemployment at the aggregate level.
Theoretical framing and synthesis of existing empirical findings across task-based and macro studies; no single new dataset provided (paper draws on literature and conceptual models).
Occupation-level analyses (e.g., BLS OEWS cross-occupation wage regressions) risk misleading conclusions about AI’s distributional effects because they aggregate over the task- and firm-level heterogeneity that drives the mechanism.
Theoretical argument and empirical illustration in the paper showing how aggregation masks within-task compression and firm-level rent capture; example regressions on OEWS used to demonstrate the limitation.
Testing the model requires within-occupation, within-task panel data on task-level performance and wages linked to firm-level AI adoption, ownership of complementary assets, and measures of rent-sharing; such data are not available at scale.
Author statement about data requirements and current data limitations; empirical illustration and discussion note absence of large-scale linked microdata meeting these criteria.
Occupation-level regressions using BLS OEWS (2019–2023) are insufficient for testing the model’s task-level predictions because aggregation across tasks and firms hides the mechanism.
Empirical illustration in the paper using occupation-level regressions on BLS OEWS 2019–2023 showing that such aggregates do not reveal within-occupation, within-task dispersion or firm-level rent concentration effects; paper argues this is a data-adequacy limitation.
A sensitivity decomposition shows five of the moments (the non‑ΔGini moments) identify internal mechanism rates (how AI changes task production, education responses, screening intensity) but do not determine the aggregate sign of inequality change.
Local identification / sensitivity decomposition performed on the calibrated model; decomposition results reported in the paper attribute mechanism-rate identification to five moments and show they leave the sign of ΔGini indeterminate.
AI did not significantly moderate the relationship between workplace stress and job performance.
Moderation test in PLS-SEM (SmartPLS 4.0) on N = 350; reported non-significant AI × Stress → Performance moderator (paper reports no significant moderating effect).
Use of AI raises needs for traceability, explainability, and continuous validation to maintain compliance and avoid error propagation in curricular decisions.
Paper's AI governance recommendations (prescriptive), referencing general AI risk principles rather than empirical study.
Realising DT value requires upfront investment in sensors, integration, standards, and skills; economic viability depends on contract structures and how gains are allocated between investors, owners, contractors, and operators.
Synthesis of cost/benefit discussions and case descriptions in the reviewed literature; policy and procurement examples referenced.
Results are robust across alternative AI index specifications, occupational classifications, and standard controls (country and year fixed effects, macroeconomic covariates).
Paper reports robustness checks across different index constructions and occupational taxonomies, with standard controls included in regressions.
Research priorities include causal studies on productivity gains from AI, firm‑level adoption dynamics, sectoral labor reallocation, long‑run general equilibrium effects, and heterogeneous impacts across regions and demographic groups.
Set of empirical research recommendations drawn from gaps identified in the literature review and limitations section; not an empirical claim but a prioritized research agenda based on secondary evidence.
Growth‑accounting frameworks and measurement approaches must be updated to capture AI/robotics as intangible and embodied capital, including quality improvements and spillovers.
Methodological argument grounded in literature on measurement challenges and examples of intangible capital; no new measurement exercise or empirical re‑estimation is provided in the paper.
Backtesting the proposed models against historical technological transitions (e.g., ATMs, robotics) and recent AI adoption episodes can validate model performance.
Recommended validation strategy; paper does not report backtest results but prescribes holdout/pseudo‑counterfactual experiments and calibration with administrative outcomes.
Scenario modelling in the reviewed literature typically uses counterfactual simulations with different adoption speeds, policy responses, and initial conditions to bound possible employment, wage, and productivity trajectories.
Description and citations of scenario-modelling practices by think tanks and organisations (TBI, IPPR, IMF) and academic work referenced; evidence is methodological and report-based.
NLP/LLM pipelines are used to extract tasks and skills from free-text job ads and to map those tasks to AI capabilities.
Described methods and citations (Xu et al., 2025; Hampole et al., 2025); evidence is methodological application of transformer-based models to job-ad text in recent studies.
Methods increasingly apply advanced NLP and large language models (BERT, LSTM, GPT-4) to parse job descriptions, map skills/tasks, and predict automation risk.
Cited methodological examples in the paper (Xu et al., 2025; Hampole et al., 2025) and discussion of common pipelines using transformer-based models to extract tasks from free-text job ads and to map tasks to AI capabilities; evidence is methodological and based on recent studies rather than a single benchmarked dataset.
Providing optional LLM access without training did not increase average exam scores versus no LLM access.
Intent-to-treat comparisons across randomized arms reported in the study: comparison of optional-access-without-training arm to no-access arm showed no average score improvement (sample n = 164).
The benefits of AI-enabled e-commerce and automated warehousing are conditional on complementary policies (competition policy, data governance, workforce reskilling, automation oversight) to manage concentration, privacy, distributional effects, and safety.
Policy-analysis synthesis supported by sensitivity checks in scenario analyses and discussion of governance risks; recommendations informed by observed distributional and market-concentration patterns in the case material.
AI’s net impact on employment to date is modest — no clear evidence of mass unemployment.
Systematic literature review/meta-synthesis of 17 peer‑reviewed publications (published 2020–2025). Aggregate assessment across those studies found no consistent empirical support for large-scale, economy-wide unemployment attributable to AI to date.
Drawing on analysis of agentic investment firm operational models demonstrating 50-70% cost reductions while maintaining fiduciary standards.
Internal analysis/modeling of agentic investment firm operational models reported by the authors; paper states the 50–70% cost reduction result but provides no sample size or detailed empirical validation in the provided text.
Fostering digital transformation alongside workforce reskilling and innovation-ecosystem development is essential for sustainable industrial growth and strengthening Kazakhstan’s global economic position.
Policy and strategic recommendations based on the study's empirical results, case studies, and macro-level index comparisons.
Digital transformation combined with workforce retraining optimizes labor costs and enhances productivity.
Synthesis of enterprise-level case examples and aggregated regression/correlation findings at industry and national levels that link digitalization and retraining programs to labor-cost and productivity indicators.
These findings provide quantitative foundations for AI capability-threshold governance.
Synthesis/interpretation of model results and empirical validation described in the paper (recommendation/implication).
Training humans to develop teamwork competencies, independent from task training, can enhance collaboration and performance in human-agent teams (HATs).
Overall experimental findings in KeyWe: task-independent teamwork training (<30 min) was associated with higher delegation, more strategy-based assignment, and better performance under difficulty for trained teams compared to controls.
Trained teams demonstrated resilience by achieving higher task performance when the game difficulty increased.
Performance comparison under increased difficulty in the KeyWe game between teams with trained humans and teams without training; task performance measured (score or completion metric) showed trained teams performed better under harder conditions.
The clearest added value of AI over structured self-reflection lies in increasing felt accountability.
Based on RCT comparisons showing no significant AI advantage over the written-reflection questionnaire on overall goal progress, but showing higher perceived social accountability in the AI condition and a significant mediation of the AI effect on progress via perceived accountability (indirect effect = 0.15, 95% CI [0.04, 0.31]).
AI-assisted goal setting can improve short-term (two-week) goal progress.
Aggregate interpretation based on the RCT finding that the AI condition outperformed the no-support control on two-week goal progress (d = 0.33, p = .016); two-week follow-up window specified in study.
The AI increased perceived social accountability relative to the written-reflection questionnaire.
Reported comparison from the RCT showing higher perceived social accountability in the AI condition versus the written-reflection condition; measured via self-report scales at follow-up (exact scale and statistics reported in paper).
A hybrid strategic–computational framework, supported by governance mechanisms (human-in-the-loop checkpoints, escalation paths, accountability structures), is motivated to manage tensions and ensure responsible decision-making in AI-rich managerial contexts.
Synthesis-driven prescriptive framework produced by cross-framework analysis; conceptual recommendation rather than implementation evidence.
Roles oriented to information processing, optimisation, and operational precision (monitor, disseminator, resource allocator) are substantially enhanced by computational thinking (automation, optimisation, algorithmic decision-support).
Theoretical mapping of computational capabilities onto Mintzberg’s information-processing roles; conceptual reasoning without empirical validation.
AI adoption will shift fact-checking tasks (more monitoring, less rote verification), creating a need for reskilling and new roles (AI tool operators, analysts); donor and public investments should fund capacity building for local organizations.
Workforce implications inferred from interview reports about changing task mixes and the study's interpretive recommendations.
Investments should prioritize hybrid models where automation provides scale and humans handle contextual, adversarial, and legally sensitive judgments.
Recommendation based on interview findings about AI benefits and limitations and the study's interpretive synthesis.
The study distills context-sensitive best practices for fact-checking in restrictive environments, including safety protocols, local partnerships, and hybrid verification workflows.
Synthesis of findings from document analysis and interviews producing a set of recommended practices documented in the study's outputs.
AI can lower verification costs and scale reach by automating tasks such as classification, clustering, alerting, and translation.
Interview reports from platform staff and interpretive analysis identifying AI-assisted use cases for prioritization, monitoring, and translation.
Community reporting and audience-focused formats are used to improve engagement.
Platform outputs and staff interviews describing deployment of community-reporting mechanisms and tailored audience formats.
Platforms form partnerships with media outlets, academic institutions, and civil-society actors to amplify reach and secure data.
Interview accounts and organizational documents describing cross-sector partnerships and collaboration arrangements.
Transparent workflows and clear labeling are used to build credibility with audiences.
Document analysis of platform outputs and guidelines showing explicit workflow transparency and labeling practices, supported by interview statements.