Evidence (7328 claims)
Adoption
5227 claims
Productivity
4503 claims
Governance
4100 claims
Human-AI Collaboration
3062 claims
Labor Markets
2480 claims
Innovation
2320 claims
Org Design
2305 claims
Skills & Training
1920 claims
Inequality
1311 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 373 | 105 | 59 | 439 | 984 |
| Governance & Regulation | 366 | 172 | 115 | 55 | 718 |
| Research Productivity | 237 | 95 | 34 | 294 | 664 |
| Organizational Efficiency | 364 | 82 | 62 | 34 | 545 |
| Technology Adoption Rate | 293 | 118 | 66 | 30 | 511 |
| Firm Productivity | 274 | 33 | 68 | 10 | 390 |
| AI Safety & Ethics | 117 | 178 | 44 | 24 | 365 |
| Output Quality | 231 | 61 | 23 | 25 | 340 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 158 | 68 | 33 | 17 | 279 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 88 | 31 | 38 | 9 | 166 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 105 | 12 | 21 | 11 | 150 |
| Consumer Welfare | 68 | 29 | 35 | 7 | 139 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 71 | 10 | 29 | 6 | 116 |
| Worker Satisfaction | 46 | 38 | 12 | 9 | 105 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 11 | 16 | 94 |
| Task Completion Time | 76 | 5 | 4 | 2 | 87 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 16 | 9 | 5 | 48 |
| Job Displacement | 5 | 29 | 12 | — | 46 |
| Social Protection | 19 | 8 | 6 | 1 | 34 |
| Developer Productivity | 27 | 2 | 3 | 1 | 33 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 8 | 4 | 9 | — | 21 |
Improved matching from predictive tools can shorten vacancy durations and improve reallocation dynamics in labor markets.
Implication from the review citing reported improvements in candidate screening and matching in some included studies; identified as a mechanism for labor-market effects.
The framework supports innovation via logical modelling and data analysis.
Listed as an advantage: logical modelling and data analysis enable innovation in instructional design. Support is conceptual; no empirical evidence presented.
Implementing the proposed framework will reduce 'brain waste' by improving recognition and cross-border mobility of DRC-trained technical personnel.
Theoretical claim supported by operations-research logic and labor-market allocation arguments in the paper; no empirical causal evaluation, sample, or longitudinal labor-market outcome data provided.
k-QREM and its estimator provide useful behavioral primitives for applied AI-economics tasks (platform design, auctions, simulations), enabling richer modeling of boundedly rational agents and within-level heterogeneity.
Discussion and proposed applications section in the paper: authors illustrate potential uses and argue suitability based on the model's expressive structure and improved performance in numerical tests; no field experimental validation reported.
A standardized governance pattern lowers coordination and compliance costs across business units, potentially increasing adoption and accelerating diffusion of advanced automation.
Theoretical claim supported by case-level practitioner observations and economic reasoning; no empirical diffusion or adoption-rate data provided.
The reference pattern yields benefits including faster, safer scaling of automation across business units, reduced compliance incidents and data-exposure risk, and better accountability and traceability of automated decisions.
Claimed benefits supported by practitioner anecdotes and multi-sector implementation descriptions; no large-sample quantitative estimates or causal inference reported.
Embedding compliance features into automation can reduce regulatory fines and litigation risk, thereby affecting firm risk profiles and cost of capital.
Theoretical implication drawn from aligning governance with compliance objectives; no empirical evidence linking the proposed pattern to reduced fines or changes in cost of capital in the paper.
The framework is applicable across multiple sectors and aligns with industry best practices; it is presented as a deployable pattern rather than a one-size-fits-all product.
Authors' assertion based on multi-sector practitioner examples and alignment with documented industry practices (qualitative). Details on sector coverage and case selection are limited.
The proposed governed hyperautomation pattern yields benefits including faster scaling of automation, reduced operational risk, maintained regulatory compliance, and preserved long-term system integrity.
Claim grounded in conceptual argument and practitioner case-based illustrations; no large-scale quantitative evaluation or causal inference provided in the paper.
Technical mitigations such as prompt/response attestation, watermarking, model output provenance, access controls, differential-design of prompts (few-shot safety), and monitoring tools can help detect or prevent prompt fraud.
Proposed technical controls and rationale derived from threat modeling and prior literature on provenance/watermarking; proposals are not empirically validated in the paper.
Targeted subsidies or support for SMEs to access SECaaS could accelerate secure AI adoption where scale barriers exist.
Economic rationale and proposed field-experiment designs; no empirical trial results presented in the chapter.
Clarifying liability and the shared responsibility model will better align incentives between providers and customers and improve security outcomes.
Policy and legal analysis; case studies of incidents where unclear responsibilities hampered response; recommended as an intervention rather than proven by causal evidence.
Promoting interoperable standards and certification can reduce lock-in and lower search costs for buyers, fostering competition in SECaaS markets.
Policy recommendation grounded in market-design theory and analogies to other standardization efforts; supporting case studies from other technology markets suggested but not empirically established here.
Open, linked phenomic–genomic datasets could inform policy and conservation markets (e.g., biodiversity credits) by improving monitoring and trait-based risk assessment models.
Policy implication advanced in the discussion; presented as potential application rather than demonstrated outcome.
Paired phenome–genome data increases the scientific and commercial value of the dataset for models predicting phenotype from genotype and vice versa.
Analytical argument in the implications section; no empirical demonstrations in the paper of improved model performance using these pairings.
Open, standardized 3D phenomic datasets reduce the need for individual labs/companies to finance expensive scanning campaigns and democratize access for academic groups and startups.
Argument in the paper's implications section based on the public release of a large standardized dataset; not an empirically tested economic outcome in the study.
Demand would grow for liability insurance tailored to EdTech, third‑party audits, fairness certifications, and specialized legal advisory services; these markets would affect costs and differential competitiveness.
Predictive market analysis and policy reasoning (no survey or market data presented).
Stricter legal exposure may slow some risky experimentation but encourage investment in fairness testing, robust evaluation, and explainability tools — potentially increasing the quality and trustworthiness of deployed AI in education.
Normative economic argumentation about incentives for R&D and testing; no empirical measurement of innovation rates provided.
Faster iterative experimental cycles enabled by LLM orchestration may increase returns to experimental R&D and change the optimal allocation between computation, instrumentation, and labor.
Economic argumentation about iterative cycles and returns to capital/labor; proposed rather than empirically demonstrated.
The method can identify frontier topics and cross-field convergence (e.g., methods migrating from NLP to vision) to inform assessments of comparative advantage and specialization across institutions/countries.
Proposed implication: using topic maps and cluster dynamics to detect frontier topics and cross-field migration; no concrete empirical examples or validation presented in summary beyond general mapping claim on ICML/ACL abstracts.
The approach is scalable and model-agnostic: different LLMs and embedding models can be swapped into the pipeline without changing the overall method.
Claimed design property in the paper summary (asserted ability to substitute different LLMs/embedding models). No detailed cross-model robustness experiments or scalability benchmarks provided in the summary.
The paper provides an initial mapping from diagnosis to intervention strategies (therapeutics) — i.e., treatment planning for model dysfunctions.
Conceptual mapping and proposed intervention strategies documented in the therapeutics section (initial mappings; not claimed as exhaustive).
Policy recommendation: governments should shift from direct administrative provision toward a strategic purchaser role using digital platforms to foster inclusive labor market access.
Policy implication derived from empirical pattern of platform-mediated employment growth and the identified Fiscal-Digital Synergy; recommendation based on observed heterogeneity by digital infrastructure and procurement channels (280-city analysis).
Public cultural services can function as productive social infrastructure that advances SDG 8 (decent work) provided adequate digital capacity exists.
Interpretation of empirical results showing employment gains contingent on digital infrastructure; normative linkage to SDG 8 drawn by authors based on observed Fiscal-Digital Synergy effects (empirical sample: 280 cities, 2008–2021).
AI should serve precision and purpose in public policy — improving foresight, enabling better trade-offs, and preserving democratic accountability.
Normative policy prescription and conceptual argumentation in the book; no empirical testing or quantified outcomes reported.
AI-driven systems should empower people with knowledge and pathways to participate in global markets rather than concentrate gains.
Normative recommendation derived from policy analysis and value judgments in the book; not supported by empirical evidence in the blurb.
Algorithmic transparency and auditability can reduce systemic risk from opaque automated lending decisions and improve regulator oversight and macroprudential policy.
Conceptual/systemic-risk argument in the "Systemic risk & governance externalities" section; no empirical systemic-risk analysis provided.
Improved algorithmic transparency could reduce information asymmetries, lowering adverse selection and moral hazard over time and potentially expanding credit to underserved populations.
Conceptual economic argument in the "Credit allocation & pricing" section; based on theory rather than empirical testing.
If properly designed and enforced, the protocol measures can improve credit access for underserved populations and reduce biased exclusion, supporting inclusive growth.
Normative claim supported by doctrinal arguments, comparative regulatory literature and technical fairness literature synthesized in the audit (no controlled empirical evaluation reported).
Firms that effectively implement governed hyperautomation may realize sustainable efficiency and reliability advantages, potentially increasing market concentration in some sectors unless governance costs level the playing field.
Strategic and competitive-dynamics argument derived from case examples and best-practice synthesis; no sector-level empirical concentration measures presented.
Standardized governance patterns reduce information asymmetries, enabling insurers and regulators to better price and manage enterprise AI risks.
Policy implication argued from the existence of standardized governance artifacts (audit trails, certifications) and industry practice; conceptual, no empirical insurer/regulator data presented.
Embedding governance reduces downside risks (compliance fines, data breaches), improving expected net returns of automation investments and lowering the adoption threshold for risk-averse firms.
Conceptual cost-benefit argument and industry best-practice examples; lacking quantitative measurement of returns or threshold shifts.
High non-wage costs (NWC ≈ 51%) and a large formalization premium (CFIL ≈ +88%) increase the private incentive to substitute labor with capital, including AI/automation, especially for routine tasks.
Policy implication derived from the measured 2023 NWC and CFIL values for the 19-country sample combined with economic substitution logic (cost of labor relative to capital/technology); no direct empirical firm-level evidence of automation responses presented in the note.
VIS can be integrated into macro/meso AI-economics models (input–output general equilibrium, growth models) to capture embodied labor and capital effects and to enable counterfactual analysis of AI diffusion scenarios.
Authors propose methodological extensions and modeling directions that embed VIS-style accounting into larger economic models for scenario analysis (conceptual suggestion).
VIS metrics can inform policy decisions (workforce retraining, sectoral subsidies, taxation) by revealing where AI-induced productivity changes will propagate through supply chains.
Authors argue policy relevance based on VIS’s ability to map upstream/downstream labor effects; presented as an implication rather than empirically validated policy outcomes.
VIS-based measures can improve measurement of AI’s productivity impacts by better capturing indirect labor displacement or augmentation from AI-driven automation across supply chains.
Conceptual extension: VIS framework captures indirect labor effects that would matter when assessing AI-driven automation impacts; not empirically tested for AI within the paper.
Research should prioritize more granular skill-to-AI-capability mappings, longitudinal tracking of adoption vs. exposure, and integration of firm behavior and regulatory dynamics into agent-based models to move from exposure assessment toward outcome prediction.
Paper's recommendations for future work built on acknowledged limitations and the gap between capability exposure and realized outcomes.
Incentives for human‑augmenting AI (e.g., subsidies or tax incentives tied to task redesign and training) can promote inclusive adoption patterns.
Policy analysis and comparative case studies; theoretical models that predict firm adoption responses to incentives, but limited causal empirical evidence specific to AI-targeted incentives.
By synthesizing computer science, engineering, and financial policy insights, DRL should be viewed not merely as a mathematical tool but as a transformative agent within the global socio-technical infrastructure of capital markets.
High-level synthesis and interdisciplinary argumentation in the paper; no empirical evidence or longitudinal studies are cited in the excerpt to demonstrate systemic transformation.
Research agenda items include quantifying social returns to different alignment interventions, studying market equilibria under participatory vs. opaque strategies, and modeling optimal regulatory mixes under uncertainty about harms and capability growth.
Prescriptive research agenda derived from the paper's economic analysis and identified knowledge gaps; presented as proposed studies rather than completed research.
If conformal filtering produces vacuous outputs at factuality levels customers demand, adoption in knowledge-intensive domains may be limited until methods simultaneously provide robustness and informativeness; vendors using efficient verifiers and robust calibration may gain competitive advantage.
Paper's market/economic discussion drawing on empirical trade-offs (informativeness vs. factuality) and cost comparisons; this is an applied implication rather than a direct experimental result.
Authors propose the 'AI orchestra' concept: future development will involve coordinated ensembles of specialized AI agents (code generation, test generation, dependency analysis, security scanning) orchestrated by humans and higher-level controllers.
Theoretical/conceptual argument by the authors grounded in qualitative findings from Netlight (practitioner reports of multiple tools and coordination frictions); this is a forward-looking synthesis rather than an empirically established fact.
Modular and cell‑free platforms could enable decentralized, localized manufacturing of specialty compounds, potentially altering trade flows away from centralized petrochemical hubs.
Conceptual synthesis plus small-scale demonstrations of modular/cell-free units in the reviewed literature; limited pilot projects and discussion of potential scalability and portability.
Canvas Design Principles aimed at reducing algorithmic myopia matter for welfare and regulatory concerns: better adaptive behavior reduces mispricing/misattribution risks but raises questions about transparency, accountability, and systemic amplification of shocks.
Policy and governance implication inferred from the claimed reductions in algorithmic myopia and increased adaptivity; study does not report direct welfare/regulatory impact measurements.
Faster, more accurate identification of demand shifts can compress the window for first‑mover advantages, intensify competitive dynamics, and raise the premium on organizational agility and human–AI integration capabilities.
Theoretical implication derived from observed improvements in signal detection (~5.8×) and resilience; not directly measured as market‑level competitive outcomes in the study.
Product teams evaluating LLM-powered features rely on a spectrum of practices—from informal “vibe checks” to organizational meta-work—to cope with LLMs’ unpredictability.
Qualitative interview study with 19 practitioners; thematic coding of transcripts produced descriptions of a range of evaluation practices used by teams.
Platform design choices (property rights, portability, reputation, tokenization, escrowed memories) will shape incentives for contributions to shared knowledge and agent improvement.
Policy and mechanism-design implications drawn from observed phenomena (shared memories, contributions, and trust) in the qualitative dataset; recommendation rather than empirically tested claim.
Shared memory architectures create public-good–like externalities (knowledge diffusion and spillovers) that may be underprovided absent coordination or platform governance.
Qualitative observations of shared memories and diffusion patterns plus theoretical economic interpretation; no empirical quantification of spillover magnitudes provided.
Easier specification of constraints can reduce some harms (clear safety violations) but centralizes normative power (who defines constraints) and creates international/cultural externalities and risks of regulatory capture.
Normative and economic argument in the paper combining technical tractability of constraints with governance concerns; this is an inference about likely distributional effects rather than empirically established fact.
Adoption of C.A.P. may reduce demand for routine oversight/clarification roles and increase demand for higher-skill roles such as prompt/system designers and dialogue curators.
Labor demand and task composition analysis presented as a conceptual projection in the paper; no labor-market empirical study reported.