Evidence (5539 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Adoption
Remove filter
RARRL increases robustness to resource constraints compared with fixed or heuristic policies (i.e., lower variance or better outcomes when compute/time budgets are constrained).
Paper reports robustness measures (variation in outcomes under constrained resources) and shows RARRL outperforming baselines and ablations across varied resource budgets in simulations with realistic latency modeling.
RARRL reduces total execution latency compared with fixed or heuristic reasoning policies.
Experimental comparisons using ALFRED-derived latency profiles report that RARRL yields lower execution latency than baseline strategies; total execution latency is listed as a primary metric.
RARRL improves task success rates compared with fixed or heuristic reasoning strategies in embodied robotic tasks (evaluated using ALFRED-derived latency profiles).
Empirical experiments reported in the paper compare RARRL to baselines (fixed strategies and heuristic triggers) using an embodied task suite based on ALFRED and empirical LLM latency profiles; results claimed to show higher task success across extensive experiments.
Policy instruments that can support shorter workweeks include tax incentives for firms that maintain pay while reducing hours, regulatory transition frameworks, and conditionality on AI subsidies or public procurement tied to job-preservation or reduced hours.
Policy-analytic argument drawing on standard policy toolkits and selected prior examples; no new policy pilot results presented.
Shorter workweeks help sustain consumer purchasing power by reducing aggregate labor supply and thereby distributing automation gains more equitably.
Theoretical labour-supply reasoning plus historical case studies of work-time reductions; argumentual and normative rather than demonstrated with new macroeconomic empirical tests in AI-rich settings.
A gradual, policy-driven reduction in the standard workweek can absorb labor displaced by automation, help maintain employment levels, and preserve wages per hour.
Synthesis of prior empirical findings on work-hour reductions and historical precedents (e.g., six-day to five-day transition); no new randomized or large-scale contemporary trials presented.
Firms use layoffs strategically to signal efficiency and boost short-term stock prices, even when automation is not fully substitutive.
Organizational- and finance-literature synthesis on signaling and market reactions to cost-cutting; historical/case examples referenced rather than new econometric estimates.
Policy implication: prioritize large-scale, targeted reskilling and lifelong learning programs to enable workforce adaptability and capture AI complementarity gains.
Policy recommendations derived from the paper's findings (association between AI adoption and skill shifts, heterogeneous sectoral impacts) and the literature synthesis that links reskilling interventions to better labor outcomes; recommendation is prescriptive rather than empirically tested within the study.
The paper provides empirical support for the complementarity hypothesis: AI tends to reconfigure jobs and create hybrid roles rather than eliminate employment wholesale.
Convergence of simulated sectoral employment patterns (some sectors showing net gains and hybrid-role growth), the strong correlation between AI adoption and skill shifts (r = 0.71), and corroborating studies from the literature synthesis emphasizing augmentation and hybridization mechanisms.
Institutional reskilling programs and governance frameworks markedly moderate labor-market outcomes: better frameworks correlate with more complementarities and lower net job loss.
Integration of literature-derived mechanisms with simulated empirical patterns; paper reports correlations/moderation-style comparisons across simulated sector-year cases incorporating policy/institutional variables (described in methods), supported by studies in the systematic review linking policy interventions to labor outcomes.
Healthcare and IT Services experienced net employment gains consistent with AI complementarity (augmented tasks and creation of new hybrid roles).
Simulated sectoral employment trends and net-change metrics for Healthcare and IT Services (2020–2024) presented in the paper, supported by literature synthesis examples showing human–AI complementarities in these sectors.
The largest rises in hybrid jobs occurred in IT Services and Healthcare.
Sectoral decomposition of hybrid job share trends in the simulated dataset across the seven industries (2020–2024) and supporting qualitative/quantitative findings from the literature synthesis focused on IT Services and Healthcare.
Hybrid human–AI jobs increased substantially across all seven analyzed sectors between 2020 and 2024.
Descriptive trend analysis of the simulated dataset's hybrid job share metric (fraction of roles reclassified as human–AI hybrid) for the seven industries over 2020–2024, combined with corroborating examples from the literature synthesis (selected ACM/IEEE/Springer studies 2020–2024).
A matching/ranking algorithm that scores candidate-job pairs by skill fit and predicted remuneration (and proximity) improves the alignment of workers to short-term gigs.
System incorporates a ranking algorithm combining inferred-skill fit, predicted wages, and proximity constraints; pilot comparison reported improved matches, but quantitative algorithmic performance metrics are not provided in the summary.
ML models can continuously derive available gigs and demand signals from marketplace activity, producing up-to-date opportunity lists and predicted wages.
Implemented ML models ingest real-time market activity/platform signals in the pilot to generate opportunity lists and wage predictions; no reported out-of-sample accuracy or prediction error metrics in the summary.
Skills can be inferred from multiple nontraditional inputs—self-reported information, short-term work histories, and community recommendations—creating richer profiles beyond formal work experience.
System design uses NLP to normalize and extract skills from profiles, short-term work records, and community recommendations; claim is supported by the implemented data integration approach rather than by quantified external validation in the summary.
The pilot implementation produced higher reported wages for youth matched through the system relative to baseline informal methods.
Pilot comparison reported higher reported wages for matched youth; summary lacks sample size, measurement protocol, and statistical inference.
The pilot implementation led to higher correct matches compared to existing informal search methods.
Pilot deployment compared matching accuracy versus baseline informal job-search approaches; the paper summary reports a 'marked increase' but provides no numerical details, sample size, or significance levels.
AI-driven NLP and ML can substantially reduce search frictions in Nairobi’s informal and gig economies by dynamically deriving individual skills and real-time market opportunities, then algorithmically matching youth to short-term work.
Pilot implementation of an end-to-end system combining NLP, ML and a matching algorithm deployed in Nairobi and compared qualitatively/aggregately against baseline informal search methods; paper summary does not report sample size, statistical tests, or numerical effect sizes.
Firms should pair strong-performing ensemble/deep models with explainability tools (e.g., feature-importance, SHAP) and fairness audits, and prefer pilot human-in-the-loop implementations to validate economic impacts and reduce operational risks.
Authors' practical recommendations based on empirical model performance, interpretability analyses, and noted limitations; presented as guidance rather than empirically validated interventions.
Variable-contribution analyses (feature importance / model explanation techniques) clarified which inputs drive predictions, making results actionable for HR decision-making.
The paper reports use of feature-importance and model-explanation methods to quantify variable contributions and interpretable outputs intended for HR practitioners.
Employee engagement/participation levels, learning agility (pace of acquiring new skills), tenure in current role, and perceived workload/manageability are consistently among the most important predictors of job performance in the datasets examined.
Feature-importance and model-explanation analyses (e.g., feature importance, SHAP-style approaches) applied across multiple publicly available workforce datasets produced consistently high importance scores for these variables.
The models' superior performance hinges on their ability to capture complex, non-linear patterns in features (e.g., engagement, learning agility, tenure, workload perception).
Inference from comparative model performance: non-linear models (ensembles, DNNs) outperform linear baselines; feature engineering captured engagement dynamics and learning trends; variable-contribution analyses highlighted these feature types as influential.
These predictive gains persist when models are applied to different company datasets, indicating better generalization of AI methods.
Cross-company tests described in the paper: models trained/tuned on one dataset and evaluated on others (holdout across organizations) with reported performance metrics demonstrating persistent improvements for AI methods.
Responsible implementation requires legal/liability clarity, continuous monitoring for performance drift and distributional shifts, usable explanations, baseline AI literacy for clinicians, and co-design with frontline radiology teams.
Synthesis of governance literature, implementation best-practice reports, and recommendations from usability and deployment studies.
Triage and automation can shorten time-to-diagnosis, increase throughput, and reduce time spent on repetitive tasks.
Observational deployment reports and simulation studies that measured time-to-report or throughput improvements in pilot settings (evidence heterogeneous and context-dependent).
Integration points for AI across the imaging pathway include acquisition (image quality/protocol selection), triage (prioritization), interpretation/reporting (detection, quantification, report pre-population), and post-interpretation (teaching, QA, model improvement loops).
Descriptive synthesis of reported implementations and proposed use cases in the literature and deployment reports across multiple institutions.
Human-AI collaboration can produce synergistic gains (diagnostic complementarity) when errors are uncorrelated and tasks are allocated to leverage comparative strengths.
Theoretical/analytical models of error complementarity and empirical reader studies showing instances where combined readings outperform either agent alone (evidence drawn from multiple small-to-moderate reader studies and simulations).
AI in radiology has clear potential to improve diagnostic performance and workflow efficiency.
Narrative synthesis of laboratory evaluation studies, reader/comparison studies, and a limited number of observational deployment reports showing improved algorithm accuracy and some improvements in measured throughput or time-to-review in pilots (study sizes and settings heterogeneous; few large-scale RCTs).
Cognitive Shadow supports real-time model updates based on immediate user feedback, enabling iterative improvement and continuous alignment with human decision patterns.
Described human-in-the-loop interaction loop where CS captures human decisions, provides recommendations, receives immediate feedback, and updates models dynamically in the simulation environment (implementation detail).
HACL/CS reduces omission rates (missed detections) in the simulated scenarios.
Omission/error rates were tracked and compared between conditions in the simulated testbed; summary claims reduction in omissions with HACL assistance but does not report numeric effect sizes or significance.
HACL/CS reduces time-to-decision in the simulated maritime surveillance tasks.
Measured time-to-classify in simulation under human-alone vs HACL-assisted conditions; summary indicates reductions in time-to-decision but lacks detailed statistics in the provided description.
In the simulated Canadian Arctic maritime surveillance domain, HACL/CS shows promise for improving classification accuracy.
Performance comparison between human-alone and HACL-assisted conditions in the maritime surveillance simulation measuring classification accuracy; summary reports improvement but does not provide sample size or significance levels.
Adjustable autonomy via self-confidence thresholds enables the system to act autonomously on high-certainty predictions and defer to humans on low-certainty cases.
System design feature of Cognitive Shadow implemented in simulation: autonomy decision rule based on meta-model confidence thresholds; behavior demonstrated in human-in-the-loop scenarios.
The Cognitive Shadow toolkit quantifies AI reliability with an empirical (0–1) confidence metric produced by a recursive meta-model.
Design and implementation detail: primary supervised models are paired with a recursive meta-model that predicts the primary model's reliability per situation and outputs a 0–1 empirical confidence score; applied in the simulated testbed.
Implementing an adaptive command-and-control process augmented by AI metacognition (the Cognitive Shadow toolkit) aligns AI judgments with expert human decision patterns.
Cognitive Shadow (CS) implemented as supervised ML models trained to mimic expert human decisions in the simulated maritime scenarios; alignment assessed by comparing model outputs to human expert decisions during human-in-the-loop interaction (implementation validated in simulation).
Human-AI co-learning (HACL) improves human-autonomy teaming (HAT) effectiveness.
Evaluated in a simulated Canadian Arctic maritime surveillance testbed using human-in-the-loop experiments comparing human-alone vs HACL-assisted conditions; exact participant sample size and statistical details not provided in the summary.
Clinical-interface validation with real physicians on mobile devices confirmed the practical viability and usability of the HADT system and interface.
Paper reports an interface test with real doctors using the mobile interface as part of validation (number of physicians, tasks, and quantitative usability metrics not specified in the summary).
On public datasets HADT achieves superior accuracy/human-effort trade-offs compared to baselines (fully human, fully automated, and simpler assignment strategies).
Comparative evaluations reported in the paper on public medical-consultation datasets (baselines listed broadly; exact baseline implementations, dataset names, and quantitative comparisons not included in the provided summary).
The execution machine uses masked hierarchical reinforcement learning with bottom-up training to ask informative symptom questions from a large symptom space.
Methodological description and reported training procedure in the paper (bottom-up training applied to the execution module; claimed to improve question selection; experimental validation referenced but details not provided in the summary).
A two-layer hierarchical reinforcement learning system—an assignment 'master' and an execution 'machine' (plus human doctors)—effectively balances accuracy and human cost.
Architectural design and experimental evaluation in the paper demonstrating trade-offs between diagnostic accuracy and human involvement using the hierarchical RL setup (experiments run on public datasets; exact sample sizes not given).
The Human-AI Diagnostic Team (HADT) framework can deliver near-expert-level online symptom inquiry and diagnosis while using very little human labor.
Performance reported on public datasets and clinical-interface tests with real physicians; described comparisons to expert-level performance in the paper's main finding (specific comparative statistics to experts not provided in the summary).
HADT reached up to 89.4% diagnostic accuracy while requiring only 10.9% human effort.
Empirical results reported in the paper from experiments on public online medical-consultation datasets and clinical-interface validation with real doctors (dataset names and sample sizes not specified in the provided summary).
AI-mediated collaboration will create new organizational roles and governance structures, such as AI mediators and verification/oversight roles.
Conceptual discussion of organizational implications and illustrative role examples; no organizational case studies with sample sizes reported.
Autonomous AI agents can automate routine coordination tasks, follow-up, and some task execution, thereby reducing human coordination overhead.
Paper uses conceptual mapping of agent capabilities to coordination/execution functions and provides illustrative case scenarios; no experimental or field data presented.
Multimodal systems (integrating text, speech, images, video) broaden communication channels and thus can improve the range and fidelity of mediated communication.
Conceptual argument and illustrative examples in the paper describing how multimodal integration maps to communication functions; no empirical validation reported.
Multilingual language models reduce language barriers by translating and normalizing meaning across languages.
Conceptual synthesis of capabilities (multilingual LMs) and mapping to coordination function (translation/normalization); supported in paper by illustrative examples rather than empirical testing.
Trust in AI should be conceptualized as a socio-technical, team-level mechanism (trust calibration) that mediates between AI design/enablers and downstream collaboration and performance, rather than an individual-level stable attitude.
Theoretical synthesis combining findings from the thematic analysis of 40 interviews with socio-technical systems theory (STS) and adaptive structuration theory (AST) to propose an initial and revised conceptual model linking enablers → trust-calibration practices → collaboration dynamics → performance.
Five enablers support effective trust calibration: transparency/explainability, clear role definitions, good user experience (UX), supportive cultural norms, and timely system feedback.
Synthesized from recurring themes in the interview data (N=40) where respondents identified these factors as facilitating appropriate reliance on AI in project settings; coded and aggregated through thematic analysis.
Performance and reward structures must be redesigned to value oversight, hypothesis testing, escalation and governance behaviours that mitigate model risk but may not immediately increase output.
Managerial recommendation derived from the framework and organizational reward literature; no empirical evaluation provided.