Evidence (2469 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Org Design
Remove filter
Policy recommendations implied include: reinforce worker voice via required worker representation in AI impact assessments and protection of collective bargaining around technology use; mandate disclosure and standardized impact reporting of AI systems used for hiring/monitoring/promotion/termination; and implement targeted sector- or task-specific enforceable regulations.
Normative policy prescriptions derived from the commentary’s analysis of governance gaps and risks; not empirically tested within the paper.
Adoption of generative neural-network audiovisual tools is effectively inevitable.
Narrative synthesis of technological trends and literature in the review; no original longitudinal adoption model or empirical adoption rates provided (qualitative projection based on cited trends).
Policymakers may need to mandate minimum verification standards or standardize audit trails/provenance metadata in safety-critical domains to reduce information asymmetries and monitoring costs.
Policy recommendation derived from risk- and externality-focused analysis; no policy impact evaluation or legal analysis presented.
Cognitive interlocks (e.g., mandatory proof artifacts, enforced testing gates, provenance/audit trails, verification quotas) make the verification burden explicit and non-bypassable, restoring the appropriate burden of proof.
Architectural design proposal with illustrative usage scenarios; no implementation, field trials, or quantitative evaluation in the paper.
The Overton Framework — an architectural model embedding 'cognitive interlocks' into development environments — can align throughput and verification by enforcing verification boundaries and restore system integrity.
Framework proposed and described conceptually; includes design principles and example interlocks but no empirical prototypes, experiments, or effectiveness evaluations reported.
Demand for AI tools, data infrastructure, and related services will grow; markets for research-focused AI products and scholarly-data platforms may expand.
Market implication noted in the paper. Based on projected trends and market signals rather than empirical market-sizing within the paper's abstract.
AI acts as a productivity multiplier that could raise the marginal returns to research inputs (time, funding), altering cost–benefit calculations for universities and funders.
Presented as an implication in the Implications for AI Economics section. This is a theoretical/economic projection rather than an empirically tested claim within the abstract; no empirical estimates or sample-based tests are provided.
Policy responses (standards for verification, disclosure rules, worker‑training subsidies) could mitigate negative labor and consumer outcomes while preserving productivity benefits.
Authors' policy recommendations based on interpretive analysis of risks and benefits reported by practitioners; normative suggestion, not empirically tested within the study.
The AR-MLLM prompt/design framework is adaptable to other industrial machine-operation scenarios.
Authors state generalizability as an argument based on the architecture and iterative prompt design; the empirical evaluation in the paper is limited to the CMM case study (no cross-domain experiments reported in the provided summary).
The Reversal Register is an auditable institutional artifact that records for each decision the prevailing authority state, trigger conditions causing transitions, and justificatory explanations, thereby supporting auditability and research.
Design specification and instrumentation proposal in the paper; description of required metadata fields and intended uses. No implemented dataset presented.
Firms that build effective orchestration layers and integrate AI across pipelines may capture outsized gains, increasing winner-take-all dynamics and concentration.
Authors' argument extrapolated from observed coordination benefits/frictions at Netlight and theory about returns to scale in platformized toolchains; no empirical market concentration analysis provided.
Policy and firm responses should emphasize human-in-the-loop governance, training in evaluative/domain skills, data stewardship, and regulatory attention to IP, liability, competition, and robustness standards.
Normative recommendations drawn from the review's synthesis of empirical benefits and limitations; based on identified failure modes (bias, hallucination, variable quality) and economic risks (concentration, mismeasurement).
Policy and regulation should emphasize transparency, auditability, and model-validation standards in finance to reduce systemic risks from misplaced trust or opaque algorithms.
Authors' normative recommendation based on empirical identification of risks (misplaced trust, overreliance) from survey/interview/operational data; recommendation is prescriptive and not an empirical test within the study.
Firms with large, integrated datasets and standardized processes can gain disproportionate returns, creating potential scale economies and winner-take-most dynamics.
Resource-based theoretical interpretation and illustrative patterns in the reviewed literature; the paper notes empirical evidence is limited and calls for further study.
Better-governed automations can reduce firms’ systemic operational risk and may lower insurance premiums or capital charges; insurers and lenders will value documented governance when pricing risk.
Hypothesized consequence grounded in risk-transfer logic and suggested interaction with insurance/lending markets; presented as implication rather than demonstrated outcome; no insurer data provided.
Policy implication (inference from results): prioritizing digital infrastructure investment to pass critical thresholds will unlock stronger productivity and environmental gains than focusing solely on advanced digital services.
Inference drawn from panel threshold findings (infrastructure threshold) and observed complementarities; this is a policy recommendation rather than a direct empirical test.
The positive AGTFP gains from digital rural development are geographically heterogeneous and are concentrated in eastern provinces.
Regional heterogeneity analysis / sub-sample regressions across provinces showing larger estimated digitalization effects in eastern provinces compared with other regions.
Digital infrastructure exhibits a threshold effect: its positive impact on AGTFP becomes stronger once digital infrastructure passes a critical level.
Panel threshold model applied to the provincial panel (2012–2022) that identifies a statistically significant threshold in the infrastructure sub-index where marginal effects increase above that value.
Authors recommend promoting a shift from single-link outsourcing (PAPM) toward whole-process integrated service provision (WAPM) as a policy implication of the findings.
Discussion/policy-implication section of the paper drawing on empirical results (TWFE and robustness checks) from the CLDS 2014–2018 analysis.