Evidence (4137 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Governance
Remove filter
Paired-game design (baseline and matched decoy-enabled game per interaction) enables direct, causal measurement of deception impact.
Methodological design described in the paper: each interaction modeled as a paired-game enabling direct comparison of equilibrium outcomes (theoretical/method section).
Equilibrium outcomes are linked to an information-theoretic uncertainty construct (entropy-like) that captures residual attacker uncertainty after observation.
Theoretical construction and formal connection drawn in the paper between equilibrium utilities and an entropy-style measure (analytical derivation).
Defender-optimal deception allocations are characterized analytically (closed-form/structural characterization of optimal resource allocation under constraints).
Analytical derivation/proofs in the paper producing defender-optimal strategy characterizations under resource/budget constraints.
The paper introduces two operational metrics: (1) value of deception (change in defender equilibrium utility attributable to deception relative to baseline) and (2) price of transparency (marginal loss in deception value induced by increased observability).
Formal definitions and mathematical expressions in the theoretical model section of the paper (analytical definitions/proofs).
The paper provides a principled, game-theoretic framework to measure and compare the operational value of cyber deception relative to a matched non-deceptive baseline.
Analytical/modeling contribution: paired strategic-game construction (baseline vs deception) with formal definitions and equilibrium analysis presented in the paper (theoretical derivation/proofs).
Policy recommendations include: invest in open metadata standards; fund pilot programs to evaluate ROI (earnings, placement, employer satisfaction); require model governance and periodic external audits for AI-assisted curriculum tools; and support smaller providers via shared infrastructure or accreditation hubs.
Explicit policy recommendations in paper (prescriptive).
Careful attention is needed to validity/reliability of assessments and to selection bias in employment outcome measurement.
Paper's methodological caveat (prescriptive); no empirical bias analysis provided.
Suggested evaluation metrics include placement rates, wage premiums, competency attainment, compliance scores, cost per qualification, and update latency.
Paper's recommended evaluation metrics (prescriptive).
Implementation requires integration with information systems for documentation, versioning, metadata, and audit trails, and benefits from continuous monitoring dashboards.
Paper's technical implementation recommendations (prescriptive).
Recommended analysis methods are qualitative (semi-structured interviews, focus groups, document review) and quantitative (surveys, competency mapping, statistical analysis of outcomes), plus systematic audit methods including traceability checks.
Paper's methods section (methodological specification).
Data inputs for the framework should include competency taxonomies, labor-market signals, regulatory requirements, learner assessment results, and stakeholder interviews.
Paper's data-input specification (descriptive).
Management principles emphasised are transparency, traceability of outcomes, IT integration for documentation, and continuous monitoring/evaluation.
Explicit management principles in paper (prescriptive).
Research and audit should emphasise validity, reliability, and compliance using mixed methods (qualitative interviews/focus groups; quantitative surveys/statistics) and systematic curriculum audits.
Recommended research & audit approach in paper (methodological guidance).
Tools recommended include logigrams (visual decision/compliance flows) and algorigram (algorithmic step-flows for planning, assessment, audit).
Tool definitions and recommendations in paper (descriptive).
Core components of the framework are inputs (learner needs, industry requirements, regulatory standards), processes (curriculum mapping, competency alignment, career assessment), and outputs (structured lesson plans, compliance-ready frameworks, career-path documentation).
Framework component list provided in paper (descriptive).
Scope of the program includes curriculum design, organisational management, career-alignment, and audit/compliance processes.
Explicit scope statement in paper (descriptive).
The framework foregrounds logical modelling (logigrams, algorigrams) and mixed-methods data analysis to support design, auditability, and alignment with industry and regulatory standards.
Paper's methodological design and tool recommendations (conceptual). No empirical implementation data reported.
The program offers a comprehensive curriculum-engineering framework linking organizational orientation, management systems, lesson planning, and career assessment into traceable, compliance-ready curriculum products.
Paper's program description and framework specification (conceptual); no empirical evaluation or sample size reported.
The paper calls for subsequent quantitative validation (using task-based, matched employer-employee, and provider-level panel data) to estimate causal impacts on productivity, health outcomes, wages, and employment composition across the three interaction levels.
Stated research agenda and measurement recommendations in the paper's discussion section.
The study is qualitative and small-sample (four case) and therefore interpretive and illustrative rather than statistically generalizable.
Explicit methodological statement in the paper: design = qualitative multiple case study, sample = four AI healthcare applications.
The study identifies a three-level taxonomy of human–AI interaction in healthcare: AI-assisted, AI-augmented, and AI-automated.
Conceptual taxonomy derived from multiple qualitative case studies (n=4) using cross-case comparison and Bolton et al. (2018)'s three-dimensional service-innovation framework.
Non-probability sampling and self-reported measures limit claims about prevalence and causality; cross-sectional design cannot capture dynamics of skill acquisition over time.
Study limitations explicitly reported by authors: non-probability sampling, self-reported measures, and cross-sectional design.
There are few large-scale randomized controlled trials (RCTs) showing direct patient outcome improvements from GenAI CDS; high-quality real-world and longitudinal studies are limited but essential.
Evidence-maturity statement in the paper summarizing the literature; the paper explicitly notes scarcity of large RCTs and longitudinal evaluations.
Randomized or quasi-experimental evaluations of digital-ID rollouts, subsidy programs for fintech adoption, or sandboxed regulatory innovations can identify causal impacts on inclusion and growth.
Methodological recommendation proposing experimental and quasi-experimental designs to obtain causal inference; no implementation results reported in the paper summary.
AI economists should prioritize measuring how AI-driven services affect access, default rates, transaction costs, and market structure, disaggregated across income groups and regions.
Methodological recommendation in the 'Implications for AI Economics' section; suggested measurement priorities rather than an empirical finding.
There is a need for economic analysis of data governance regimes, model transparency requirements, algorithmic auditability, and incentives for responsible AI adoption in finance.
Methodological and policy recommendation based on identified gaps in the literature and regulatory practice; this is a stated research/policy need in the paper rather than an empirical claim requiring sample evidence.
Typical evaluation metrics reported are accuracy, precision, recall, F1-score, AUC, detection rate, false positive rate, latency, and computational cost.
Survey of evaluation practices in reviewed papers listing the metrics authors commonly report.
Emerging approaches in the literature include federated learning, online/streaming learning, and transfer learning for cross-device generalization.
Trend analysis across recent papers indicating adoption of federated and continual learning paradigms and transfer-learning techniques.
Unsupervised and semi-supervised methods (clustering, one-class classifiers, autoencoder-based anomaly detectors) are commonly employed to handle unlabeled/anomalous IoT traffic.
Synthesis of studies using anomaly-detection paradigms and unsupervised techniques reported in the reviewed papers.
Deep learning approaches used include CNNs, RNNs/LSTMs for sequence/traffic analysis, and autoencoders for anomaly detection.
Surveyed literature and taxonomy noting multiple studies that apply convolutional and recurrent architectures and autoencoders to network/traffic data.
Common ML approaches reported for IoT IDS include supervised models (random forest, SVM, gradient boosting, neural networks).
Taxonomy and literature synthesis showing frequent use of classical supervised classifiers in surveyed papers and experiments.
This work is a conceptual framework and design proposal synthesizing methods from recommender systems and HRI rather than a report of novel empirical experiments.
Explicit statement in the Data & Methods section of the paper.
The review followed PRISMA guidelines and included 30 scholarly articles retrieved from Scopus, published between 2020 and 2025, selected using pre-specified inclusion criteria.
Methods section of the paper reporting the SLR protocol, database, time window, and number of included studies.
The study is primarily diagnostic and prescriptive rather than empirical: no explicit empirical dataset, causal identification strategy, or statistical estimation is reported.
Methods section of the paper explicitly characterizes the work as conceptual, systems-oriented, and not reporting empirical evaluation data.
The study's empirical identification relies on longitudinal variation with city fixed effects and time effects, plus non-linear/threshold identification via polynomial (DE^2) terms and threshold-regression using green-technology-innovation as the threshold variable.
Description of empirical strategy in the paper: panel fixed-effects models (controlling for time-invariant city heterogeneity and common time shocks), mediating-effect models for channel tests, and threshold-regression models for regime-dependent effects, applied to the 278-city 2011–2022 panel.
Research recommendation: invest in longer-run, rigorous impact evaluations (RCTs, panel studies) and system-level assessments to capture spillovers and sustainability outcomes.
Authors' stated research agenda based on identified methodological gaps (limited long-term and system-level evidence) in the review.
There is variation in study design and quality in the evidence base (RCTs, quasi-experimental studies, observational case studies, pilots).
Methodological caveats noted by the authors summarizing the diversity of designs reported across reviewed studies.
The review used a structured literature review with thematic synthesis and a comparative effect-size analysis to quantify ranges for yield, cost, and efficiency outcomes.
Authors' description of review approach and analytical methods in the Data & Methods section.
The evidence base reviewed comprises more than 60 peer-reviewed articles and institutional reports from 2020–2025, primarily focusing on Sub-Saharan Africa.
Statement in the paper's Data & Methods section describing the scope and composition of the review sample.
Effect sizes and impacts vary substantially across contexts—by crop, farm size, and institutional setting.
Comparative synthesis across studies showing heterogeneity in reported outcomes and authors' methodological caveats highlighting context dependence.
Technologies assessed in the review include predictive analytics, digital advisory systems, smart irrigation, pest/disease detection, and precision fertilization.
Descriptive synthesis of the types of AI and digital technologies evaluated across the >60 reviewed articles and reports (2020–2025).
These quantitative performance figures come from case‑level, high‑performer pilots and should not be treated as typical industry benchmarks.
Authors' caveat based on the composition of evidence in the review (skew towards pilots and selected advanced implementations; limited longitudinal/multi‑project empirical studies).
Inter‑rater reliability for the study selection/encoding was Cohen’s κ = 0.83 (substantial agreement).
Reported inter‑rater reliability statistic from the review's quality control step (Cohen's kappa = 0.83).
The review screened 463 Scopus records (2018–2026) and selected 160 peer‑reviewed studies using a PRISMA‑guided process.
Systematic literature review described in paper: Scopus search (2018–2026), PRISMA screening and eligibility filtering; initial n=463, final n=160.
Kebumen UNESCO Global Geopark is used as a practical context to ground the framework; its ecological/cultural assets and emergent digital presence make it a suitable case for studying emerging destinations balancing innovation with authenticity.
Paper provides Kebumen Geopark as the illustrative case study/context for the conceptual framework; no systematic case-study data reported.
Operationalization suggestions: social proof via ratings, reviews, UGC volume and valence; behavioral proxies include bookings and inquiries as outcomes.
Paper explicitly lists social-proof indicators and behavioral proxies as part of recommended empirical approaches (digital-trace and platform data).
Operationalization suggestions: sustainability communication via message clarity, perceived authenticity, and specificity of eco-actions.
Operationalization guidance in the paper for measuring sustainability messaging in experiments/surveys.
Operationalization suggestions: AI personalization via perceived relevance, transparency, and perceived fairness of recommendations.
Operationalization guidance in the paper; proposed as latent construct indicators for future SEM or experiments.
Operationalization suggestions: digital experience quality via usability, information richness, responsiveness, multi-channel integration.
Operationalization guidance provided in the paper's methods suggestions; intended for future empirical measurement.
Recommended empirical follow-ups include Structural Equation Modeling (SEM), experimental tests (lab/field/online), quasi-experimental causal-inference methods (DiD, IVs, RD), comparative/regional designs, and analysis of digital-trace/platform data (clickstreams, recommendation logs, bookings, UGC).
Methodological recommendations explicitly listed in the Data & Methods and Research Agenda sections of the paper; no primary empirical work conducted.