Evidence (3029 claims)
Adoption
5200 claims
Productivity
4485 claims
Governance
4082 claims
Human-AI Collaboration
3029 claims
Labor Markets
2450 claims
Org Design
2305 claims
Innovation
2290 claims
Skills & Training
1920 claims
Inequality
1299 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 373 | 105 | 59 | 439 | 984 |
| Governance & Regulation | 366 | 172 | 114 | 55 | 717 |
| Research Productivity | 237 | 95 | 34 | 294 | 664 |
| Organizational Efficiency | 364 | 82 | 62 | 34 | 545 |
| Technology Adoption Rate | 292 | 115 | 66 | 27 | 504 |
| Firm Productivity | 274 | 33 | 68 | 10 | 390 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Output Quality | 231 | 61 | 23 | 25 | 340 |
| Market Structure | 107 | 121 | 85 | 14 | 332 |
| Decision Quality | 158 | 68 | 33 | 17 | 279 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 88 | 31 | 38 | 9 | 166 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 105 | 12 | 21 | 11 | 150 |
| Consumer Welfare | 67 | 29 | 35 | 7 | 138 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 67 | 31 | 4 | 126 |
| Task Allocation | 70 | 9 | 29 | 6 | 114 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 11 | 16 | 94 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 76 | 5 | 4 | 2 | 87 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 15 | 9 | 5 | 47 |
| Job Displacement | 5 | 29 | 12 | — | 46 |
| Developer Productivity | 27 | 2 | 3 | 1 | 33 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 8 | 4 | 9 | — | 21 |
Human Ai Collab
Remove filter
Proposition 1: Earlier, decentralised access to training reduces information asymmetry and dependence on intermediaries.
Presented as a testable proposition derived from corridor process mapping and conceptual analysis; recommended for randomized or quasi-experimental evaluation but not empirically tested in this paper.
Redesigning pre-departure training along four axes—standards, timing, delivery architecture, and recognition/portability—can reduce information asymmetries, lower dependence on brokers, and better connect migration to labour‑market value without waiting for slower permit/enforcement reforms.
Argument derived from conceptual reframing and corridor process mapping; supported by desk review and governance gap analysis. Presented as a policy proposition rather than empirically tested causal claim.
Economically, there will be demand for 'temporal-quality' products: neurotech and AI services that explicitly measure, preserve, or enhance experienced temporality (presence, flow, meaning), representing a distinct market segment.
Speculative market implication derived from conceptual argument and literature on consumer preferences; no market data or empirical demand studies provided.
Recommended priorities include funding longer, practice‑embedded programs, developing standardized competency frameworks and validated assessments, and conducting studies that link training to organizational and patient outcomes (to enable level‑4 evidence and economic evaluation).
Authors' practical and policy recommendations based on synthesis of findings (limited depth/duration of current programs and lack of level‑4 outcomes) described in the paper.
Interpretive claim: AI interventions (upskilling and AI-guided workflows) raise worker confidence and job satisfaction and help tailor stress-management approaches, which can support retention under stress.
Authors' interpretive summary (not tied to a specific reported coefficient); described as a mechanism for the observed AI moderation on retention. Instrument/scale details and direct measurement of confidence/job satisfaction not provided in the summary.
Observed higher short-term performance and the positive correlation with iterative engagement imply that GenAI can augment short-term academic productivity and that benefits depend partly on active, skillful user interaction (complementarity).
Synthesis in implications drawing on the experimental finding of higher scores for allowed-use groups and the positive correlation between number of edits and performance; this interpretive claim is inferential and not directly tested as a structural complementarity in the study.
The dataset and model are bilingual and cover varied acquisition settings, which the authors claim increases heterogeneity and clinical realism and should improve generalizability across care settings.
Paper statement about dataset being bilingual and covering a range of acquisition settings; authors argue this increases heterogeneity and realism. (Languages, sites, and formal external validation results across healthcare systems are not provided in the summary.)
Policymakers and firms should prioritize upskilling, standards for model provenance and IP, liability frameworks for AI-generated code, and improved measurement to track AI-driven productivity changes.
Policy recommendations derived from identified risks, barriers, and implications in the literature review and practitioner survey; not an empirically tested intervention.
By better controlling tail risk and rare catastrophic harms, RAD can reduce expected social costs, liability exposure, and insurance premiums associated with high-impact AI failures.
Economic implications and argumentation in the paper that link reduced tail risk (from RAD) to lower social costs and liabilities; this is an extrapolation from method-level safety improvements rather than a direct empirical measurement of economic outcomes.
The framework formalizes complementarities between AI and managerial/human capital (e.g., exception handling, trust-driven adoption), suggesting empirical work should measure task reallocation rather than simple displacement.
Conceptual claim and research agenda recommendations in the paper (no empirical measurement provided).
Staged, practice-oriented workflows lower upfront adoption costs and implementation risk for SMEs, increasing marginal adoption likelihood when organizational readiness and governance are explicit.
Theoretical/economic implication derived from the framework and pilot rationale; not directly validated by large-scale empirical evidence in the paper (asserted implication).
High accuracy and reproducibility have been demonstrated on narrowly scoped tasks such as image interpretation, lesion measurement, triage ranking, documentation support, and drafting written communication.
Synthesized empirical evaluations of CNNs in imaging (diagnosis, lesion measurement, triage) and benchmarking/medical assessment studies of LLMs for documentation and drafting; multiple cited empirical studies and benchmarks included in the narrative review (no pooled quantitative estimate).
Effective policy should be comprehensive and sequenced: unlock data (clear ownership, safe-sharing frameworks), provide targeted investment incentives (matching grants, procurement commitments), run human-capital programs (upskilling, industry–university links), and build core infrastructure (sensors, connectivity, local compute).
Policy synthesis derived from the institutional analysis and identification of interacting bottlenecks; recommendations based on theoretical best-practices rather than causal evaluation.
Policymakers may need to mandate minimum verification standards or standardize audit trails/provenance metadata in safety-critical domains to reduce information asymmetries and monitoring costs.
Policy recommendation derived from risk- and externality-focused analysis; no policy impact evaluation or legal analysis presented.
Cognitive interlocks (e.g., mandatory proof artifacts, enforced testing gates, provenance/audit trails, verification quotas) make the verification burden explicit and non-bypassable, restoring the appropriate burden of proof.
Architectural design proposal with illustrative usage scenarios; no implementation, field trials, or quantitative evaluation in the paper.
The Overton Framework — an architectural model embedding 'cognitive interlocks' into development environments — can align throughput and verification by enforcing verification boundaries and restore system integrity.
Framework proposed and described conceptually; includes design principles and example interlocks but no empirical prototypes, experiments, or effectiveness evaluations reported.
Demand for AI tools, data infrastructure, and related services will grow; markets for research-focused AI products and scholarly-data platforms may expand.
Market implication noted in the paper. Based on projected trends and market signals rather than empirical market-sizing within the paper's abstract.
AI acts as a productivity multiplier that could raise the marginal returns to research inputs (time, funding), altering cost–benefit calculations for universities and funders.
Presented as an implication in the Implications for AI Economics section. This is a theoretical/economic projection rather than an empirically tested claim within the abstract; no empirical estimates or sample-based tests are provided.
Policy responses (standards for verification, disclosure rules, worker‑training subsidies) could mitigate negative labor and consumer outcomes while preserving productivity benefits.
Authors' policy recommendations based on interpretive analysis of risks and benefits reported by practitioners; normative suggestion, not empirically tested within the study.
The AR-MLLM prompt/design framework is adaptable to other industrial machine-operation scenarios.
Authors state generalizability as an argument based on the architecture and iterative prompt design; the empirical evaluation in the paper is limited to the CMM case study (no cross-domain experiments reported in the provided summary).
The Reversal Register is an auditable institutional artifact that records for each decision the prevailing authority state, trigger conditions causing transitions, and justificatory explanations, thereby supporting auditability and research.
Design specification and instrumentation proposal in the paper; description of required metadata fields and intended uses. No implemented dataset presented.
Firms that build effective orchestration layers and integrate AI across pipelines may capture outsized gains, increasing winner-take-all dynamics and concentration.
Authors' argument extrapolated from observed coordination benefits/frictions at Netlight and theory about returns to scale in platformized toolchains; no empirical market concentration analysis provided.
Policy and firm responses should emphasize human-in-the-loop governance, training in evaluative/domain skills, data stewardship, and regulatory attention to IP, liability, competition, and robustness standards.
Normative recommendations drawn from the review's synthesis of empirical benefits and limitations; based on identified failure modes (bias, hallucination, variable quality) and economic risks (concentration, mismeasurement).
Policy and regulation should emphasize transparency, auditability, and model-validation standards in finance to reduce systemic risks from misplaced trust or opaque algorithms.
Authors' normative recommendation based on empirical identification of risks (misplaced trust, overreliance) from survey/interview/operational data; recommendation is prescriptive and not an empirical test within the study.
Public goods investments—digital infrastructure, interoperable local data ecosystems, and multilingual language technologies—are prerequisites for inclusive economic benefits from AI.
Conceptual and policy literature review arguing for infrastructure and public data ecosystems; paper does not provide original infrastructure impact analysis.
A culturally grounded responsible‑AI governance framework based on Afro‑communitarianism (Ubuntu) and stakeholder theory—emphasizing collective well‑being and participatory governance—can help align AI deployment with inclusive and sustainable economic outcomes.
Theoretical integration and framework development based on normative literature in ethics, Afro‑communitarian thought, and stakeholder governance; framework is conceptual and not empirically validated in this paper.
Firms with large, integrated datasets and standardized processes can gain disproportionate returns, creating potential scale economies and winner-take-most dynamics.
Resource-based theoretical interpretation and illustrative patterns in the reviewed literature; the paper notes empirical evidence is limited and calls for further study.
Explainable EEG tools can shift clinician workflows by enabling faster decision-making and reducing the requirement for specialized interpretation, with implications for training, staffing, and productivity.
Projected operational impacts discussed as implications of improved explainability; no longitudinal workflow study provided in the reviewed literature.
Policy and managerial implication suggested: investing in short, targeted onboarding/training for GenAI tools (rather than only providing access) may deliver measurable performance gains and increase voluntary adoption.
Authors derive this implication from the randomized trial results showing increased adoption and improved scores with brief training (n = 164); this is an extrapolation from the trial findings.