Evidence (7395 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 736 | 1615 |
| Governance & Regulation | 664 | 329 | 160 | 99 | 1273 |
| Organizational Efficiency | 624 | 143 | 105 | 70 | 949 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 348 | 109 | 48 | 322 | 836 |
| Output Quality | 391 | 120 | 44 | 40 | 595 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 275 | 143 | 62 | 34 | 521 |
| AI Safety & Ethics | 183 | 241 | 59 | 30 | 517 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 105 | 40 | 6 | 187 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 78 | 8 | 1 | 151 |
| Regulatory Compliance | 69 | 64 | 14 | 3 | 150 |
| Training Effectiveness | 81 | 15 | 13 | 18 | 129 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
Adoption
Remove filter
Alignment with evolving regulatory expectations (evidence standards, auditing, liability) is necessary to translate AI capabilities into products and reduce adoption risk.
Policy-focused argument referencing regulatory uncertainty; no empirical measures of regulatory impact included.
Realized, sustained impact ('democratized discovery') from AI depends on non-technological enablers: high-quality interoperable data, rigorous validation, transparency/auditability, workforce upskilling, ethical oversight, and regulatory alignment.
Synthesis and prescriptive argument in editorial grounded in observed constraints; no empirical testing of causal dependence provided.
The review synthesizes cross-domain evidence on the use of AI across the continuum from target identification to regulatory integration and critically evaluates existing limitations including data bias, interpretability discrepancy, and regulatory ambiguity.
Statement about the scope and content of the review (literature synthesis and critical evaluation). This is a description of the paper's methods/content rather than an empirical finding; the excerpt indicates these topics are discussed.
The study investigates the benefits and drawbacks associated with the incorporation of innovative artificial intelligence technologies into industrial policies.
Author-stated research objective reported in the text; evidence claimed to come from literature review (novel studies and existing literature), but no specific studies, sample sizes, or empirical measures are provided in the excerpt.
Model output can be treated as evidence for studying human behavior, but there are important epistemic limits to interpreting model-generated text as direct evidence of human beliefs or social facts.
Epistemic analysis and methodological critique in the paper (discussion of limits of treating model outputs as evidence); no single empirical test cited in the provided text.
The paper constructs three policy-contingent labor market scenarios for 2025–2035: (1) an Augmented Services Economy with inclusive productivity gains, (2) a Dual-Speed Labor Market characterized by polarization and uneven adjustment, and (3) a Disruptive Automation Shock involving significant displacement and social strain.
Prognostic, scenario-based approach integrating the three evidence bases (task-level capability mapping, occupational exposure/complementarity analysis, and firm- and worker-level adoption evidence). The scenarios are developed and described in the paper for the 2025–2035 horizon.
The validity of human–AI decision-making studies hinges on participants' behaviours; effective incentives can potentially affect these behaviours.
Conclusion from the authors' thematic review and theoretical rationale linking incentive design to participant behaviour and study validity (no quantitative effect sizes provided in excerpt).
The study's counterfactual analytical model links HR indicators (training intensity, absenteeism, labor productivity, turnover rates, workforce allocation) to organizational performance outcomes using regression-based simulations and predictive estimation.
Methodological claim explicitly stated: model construction from an industrial firm dataset using regression-based simulations and predictive techniques. (Specific sample size, variable operationalizations, and time frame not reported in the description.)
Only one study reported a modest improvement in predicting endoscopic intervention needs (AUC: 0.68).
Single-study result cited in the review reporting AUC = 0.68 for prediction of need for endoscopic intervention.
The review synthesizes findings across five thematic areas: AI‑driven task automation and decision support; digital literacy and capacity building; gender‑sensitive employment patterns; infrastructural and policy challenges; and sustainable development outcomes.
Thematic synthesis of the 55 included articles as described in the paper; themes explicitly listed by the authors.
Prevalence and risk factors for poverty differ by gender, as does the nature of vulnerability.
Stated as a general empirical claim in the introduction, drawing on broader literature (no specific study, method, or sample size provided in the excerpt).
Major actors such as the United States, China, and the European Union pursue distinct models of AI development and regulation.
Comparative policy analysis and qualitative document review of national/regional AI strategies and regulatory proposals for the United States, China, and the EU (specific documents and sample size not specified).
The study identifies the emergence of three competing governance paradigms: the innovation-driven liberal model, the ethics-oriented regulatory model, and the state-controlled authoritarian model.
Finding from the paper's comparative policy analysis and qualitative review of policy documents across major actors (United States, European Union, China); underlying document sources referenced qualitatively but not enumerated as a quantitative sample.
Distinct AI features (recommendation engines, chatbots, and comparison tools) influence consumer outcomes when modeled as latent constructs.
Methodological claim: the study modeled three AI features as latent constructs and analyzed their relationships with dependent variables using SEM (quantitative questionnaire data).
We develop a theoretical framework - the productivity funnel - that traces how technological potential narrows through successive stages, from access and digital infrastructure, through organizational absorption and human capital adaptation, to ultimate value capture.
Conceptual/theoretical development presented in the paper; no empirical sample needed (framework-building).
Effects of curated Skills are highly heterogeneous across domains (e.g., +4.5 pp in Software Engineering vs. +51.9 pp in Healthcare).
Per-domain pass-rate deltas reported in the paper (SkillsBench per-domain analysis). The example domain deltas (+4.5 pp and +51.9 pp) are taken from the reported per-domain results.
Institutional factors (education systems, active labor market policies, mobility, industrial policy, social protection) shape net employment outcomes from AI.
Theoretical and policy-focused synthesis; cross-country comparisons in literature highlight institutional mediation though no single new cross-country empirical estimate is provided.
Net employment effects depend on the balance of substitution and complementarity, sectoral exposure, and institutional responses.
Conceptual labor-economics framework (task-based, skill-biased change) and comparative review of cross-country/sectoral evidence emphasizing institutional mediation.
AI will substantially restructure labor markets.
Task-based theoretical approach and cross-sectoral synthesis of empirical studies showing task substitution and complementarity effects across occupations and sectors.
The pandemic produced a 1.5% increase in people identifying as potential entrepreneurs but a 2.3% contraction in emerging entrepreneurs, indicating a breakdown in converting aspiration into formal entrepreneurial activity (pipeline disruption).
Reported percentage changes in pipeline stages (potential entrepreneurs and emerging entrepreneurs) measured in the survey before/after (or during) the pandemic within the >27,000 respondent sample; comparison of identification and transition rates along the entrepreneurial pipeline.
Long-run integration (degree of long-run association) between core AI and AI-enhanced robotics differs systematically across national innovation systems.
Country-level decomposition of patent filing series and time-series econometric tests for long-run relationships / cointegration between core AI and AI-enhanced robotics patent series for each country/region (China, U.S., Europe, Japan, South Korea).
Core AI, traditional robotics, and AI-enhanced robotics follow distinct historical trajectories over 1980–2019 and do not move together uniformly.
Time-series analysis using annual patent filing counts (1980–2019) for each domain; tests for common long-run relationships / co-movement across the three patent series (as reported in the paper). Country-aggregated and domain-specific patent time series were analyzed; exact sample size (total patents) not specified in the summary.
Kondratieff, Schumpeter, and Mandel each highlight different drivers of capitalist long waves: Kondratieff emphasizes regular technological-driven renewal, Schumpeter emphasizes entrepreneurship and innovation-led creative destruction, and Mandel emphasizes class relations and production structures.
Comparative theoretical analysis and literature synthesis across the three schools; conceptual summary of canonical positions (no original dataset; qualitative interpretation).
XChronos reframes transhumanist technology evaluation in experiential terms, creating both market opportunities and measurement/regulatory challenges for AI economics.
Synthesis and concluding argument in the paper summarizing proposed implications; conceptual reasoning without empirical tests.
The methodological landscape of the evidence base is heterogeneous, consisting of cross-sectional surveys, case studies, quasi-experimental designs, and a limited number of longitudinal analyses.
Study design information was extracted from the 145 included studies revealing a mix of designs and relatively few longitudinal or experimental studies.
Human factors (training, trust calibration, workflows) determine whether clinicians accept, override, or ignore GenAI suggestions.
Qualitative and quantitative human-AI interaction studies and pilot deployments discussed in the paper; specific sample sizes and effect sizes are not reported in the paper.
Safety and net benefit of GenAI CDS hinge on deployment details: user interface, real-time feedback, uncertainty quantification, calibration, and how recommendations are presented (strong vs. suggestive).
Human factors and implementation studies referenced; early A/B tests and human-AI interaction research suggest interface and presentation affect acceptance and error rates; no large-scale standardized implementation trial data cited.
Reimbursement models (fee-for-service vs. capitation) will influence whether cost savings from GenAI are realized or offset by increased service volume.
Economic incentive framework and prior health-economics literature cited; the paper does not provide direct empirical tests but references plausible incentive channels.
RL and adaptive methods are good for real-time adaptation but can be myopic, require large amounts of interaction data, and struggle to incorporate long-term preference structure and ethical constraints.
Surveyed properties of reinforcement learning and adaptive methods in HRI/RS literature; no new empirical evaluation in this paper.
Key tradeoffs in contemporary financing models include speed/flexibility versus regulatory coverage and long‑term cost, and data reliance versus privacy/fairness.
Multi‑criteria comparative evaluation and conceptual analysis across financing models; synthesis draws on regulatory context and observed product features rather than primary quantitative tradeoff estimation.
Performance of structure prediction models scales with data, model size, and compute; there are tradeoffs between accuracy and inference speed/simplicity.
Paper explicitly states scaling behavior and tradeoffs in 'Compute and training' and 'Representative models' sections; no precise scaling curves or thresholds are provided in the text.
Important tradeoffs exist (privacy vs. utility; centralized vs. federated data architectures; automated moderation vs. freedom of expression; cost/complexity of secure hardware) that must be balanced in VR security design.
Comparative evaluation across the reviewed corpus (31 studies) identifying recurring ethical and technical tradeoffs; authors discuss these qualitatively.
Across the EU, Algeria, and Pakistan there is convergent recognition of dual‑use risks, increasing use of export controls, and interest in developing domestic AI capacity.
Cross‑jurisdictional synthesis of national/supranational legal texts, export‑control policies, and policy documents showing discussion of dual‑use issues and capacity building.
The community knowledge functions both as practical how-to guidance and as collective experimentation with platform rules and revenue mechanisms.
Observed dual nature in the 377-video corpus: instructional workflows alongside demonstrations/testing of platform-tailored monetization tactics and workarounds.
Typical practices emphasized by creators include rapid mass production of content, productizing prompt engineering, repurposing existing material via synthesis/localization, and packaging AI outputs as sellable creative services or assets.
Recurring practices surfaced through qualitative coding of workflows, tools, and pipelines described in the 377 videos.
Across the 377 videos, creators converge on a set of repeatable use cases and platform‑tailored monetization tactics.
Thematic coding of 377 videos produced a catalog of recurring use cases and tactics; the paper reports convergence across that sample.
YouTube creators have collectively constructed and circulated a practical knowledge repository about how to monetize GenAI-driven creative work.
Systematic qualitative content analysis (thematic coding) of 377 publicly available YouTube videos in which creators promote GenAI workflows and monetization strategies.
Citation counts across repeated samples follow a power-law (heavy-tailed) distribution: a few domains are cited often while many domains are cited rarely.
Empirical distributional analysis of citation counts from repeated samples collected across the three platforms and three topics (multi-day and high-frequency regimes); observed heavy-tailed / power-law fit to citation-count distribution.
The topology of service-dependency graphs (modelled as DAGs of compute stages) is a first-order determinant of whether decentralised, price-based resource allocation will be stable and scalable.
Systematic ablation study using simulation: 1,620 runs total across six experiment types, sweeping graph topology (hierarchical vs cross-cutting), load, hybrid integrator presence, and governance constraints; metrics included price convergence/volatility and allocation throughput/quality. Effect sizes reported in the paper show topology had the largest impact on price stability and scalability.
Choice of scaffold materially affects outcomes: an open-source scaffold outperformed vendor-provided scaffolds by up to approximately 5 percentage points.
Comparative experiments across three scaffolding approaches (vendor scaffolds and at least one open-source scaffold) showing up to ~5 percentage point differences in measured outcomes.
Absence of irreducibility, positive recurrence, or aperiodicity in the state dynamics can produce non-ergodic reward behavior.
Theoretical argument and examples in the paper illustrating how breakdowns of these chain conditions lead to multiple invariant measures or absorbing regimes; analysis-based evidence.
Standard Markov chain ergodicity conditions (irreducibility, positive recurrence, aperiodicity) imply ergodic reward processes when rewards depend only on the chain state.
Formal mapping in the paper between Markov-chain ergodicity properties and reward-process ergodicity; theoretical derivation (no empirical sample).
Non-ergodic processes admit path-dependent long-run behavior (e.g., absorbing sets, multiple invariant measures, path-dependent reinforcement), so different runs with the same policy can have different long-run averages.
Analytic discussion of Markov-chain examples and theory plus the paper's illustrative constructed example showing path-dependent locking into regimes; theoretical and example-driven evidence.
Ergodic reward processes are those where time averages along almost every long trajectory converge to the same value as the ensemble average.
Formal definition and discussion in the paper mapping ergodicity concepts from stochastic processes to reward processes; theoretical exposition.
The model explicitly separates competition into two stages: discovery (first-passage to resource patches) and monopolization (local takeover and stabilization).
Model specification in the paper: stochastic, spatially-structured population model with distinct discovery and monopolization dynamics; this is a modeling assumption/structure rather than empirical measurement.
Two qualitatively distinct mechanisms underlie observed dominance: (1) extreme-event-mediated lucky discovery (transient), and (2) mechanistic asymmetries (non-reciprocal biases) that convert lucky discovery into permanent dominance.
Conceptual separation in the model structure (discovery vs monopolization phases), analytic results on first-passage extreme events, and absorbing-state analysis showing necessity of asymmetry for permanence; supported by simulations demonstrating the two-stage behavior. The claim is theoretical.
Explanations change workflows, shift responsibilities between humans and machines, and can reshape power dynamics—creating both opportunities (better oversight) and risks (over-reliance, gaming).
Qualitative and conceptual studies synthesized in the review, including socio-technical analyses and case studies reporting observed or theorized workflow and responsibility shifts; no meta-analytic causal estimate.
Explanations increase user trust principally when they are understandable, actionable, and aligned with users’ domain knowledge; opaque or overly technical explanations can fail to build trust or even decrease it.
Thematic synthesis of empirical and conceptual studies in the reviewed literature reporting conditional effects of explanation form and comprehensibility on trust; review notes heterogeneity in study designs and contexts.
Explainability improves perceived legitimacy, user trust, and organizational accountability only when technical transparency is paired with human-centered explanation design and governance mechanisms.
Synthesis of studies from the reviewed literature showing conditional effects of algorithmic interpretability combined with explanation design and governance; derived via thematic coding across technical and social-science sources (no new primary experimental data reported).
Explainability is a necessary but not sufficient condition for trustworthy AI in high-stakes domains.
Systematic literature review (thematic coding and synthesis) of interdisciplinary scholarship (peer-reviewed research, technical reports, policy documents); the paper synthesizes conceptual and empirical studies rather than presenting new primary data. Emphasis on high-stakes domains (healthcare, finance, public sector).