Evidence (4049 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Governance
Remove filter
We currently lack an understanding of how political parties perceive the potential impact AI has on employment, the role of regulations in protecting workers from AI-related job losses, and the importance of AI educational and training programs.
Statement of a literature/knowledge gap motivating the study (assertion by the authors; no empirical basis provided in the excerpt).
Signal legitimacy was validated through negative control experiments.
Experimentation claim: the paper asserts that negative control experiments were run to validate that signals are not due to memorized ticker associations. The excerpt does not specify the design, number, or results of these negative controls.
Pidgin should not be treated as 'broken English' but as necessary linguistic infrastructure for repaired, sustainable development; failures often reflect language-sovereignty crises requiring political solutions.
Normative claim supported by mixed-methods findings on comprehension, adoption, and legitimacy, and Critical Discourse Analysis of institutional language hierarchies.
The paper advances a new conceptual framework called 'Developmental Sociolinguistics' and formalizes Three Laws of Linguistic Justice (Epistemic Access, Discursive Parity, Sovereignty), operationalized via a proposed 'Pidgin Protocol' for decolonized development practice.
Conceptual/theoretical contribution based on synthesis of field results and literature; proposal of framework and laws as normative prescriptions rather than empirically tested policy interventions.
Standards for provenance, labeling of AI-generated content, and interoperable evidence formats would lower verification costs and create beneficial network effects.
Policy recommendation derived from identified verification frictions and the study's analysis of data/model governance needs.
There is growing market demand for AI-assisted fact-checking tools, creating opportunities for software, monitoring services, and labeled datasets.
Analytic implication drawn from findings about increasing AI use and needs for automation/labeling; based on interviews and market inference in the study.
Regulators should consider guidelines on AI monitoring, algorithmic fairness in performance evaluations, and protections to prevent hybrid‑induced career penalties.
Policy recommendation based on conceptual assessment of risks identified in literature synthesis; not an empirical claim—no policy evaluation data provided.
Hybrid agency implies complementarity between GenAI and managerial/knowledge‑worker skills (curation, evaluation, coordination), potentially increasing returns to those skills while automating routine cognitive tasks—consistent with skill‑biased technological change.
Synthesis of recurring themes linking GenAI capabilities with managerial skill topics in the thematic clusters; positioned as an implication for labour demand and skill composition rather than an empirically tested effect.
Policy prescriptions for developing countries to mitigate these vulnerabilities include: diversify supply sources, invest in local human capital and mid-stream capabilities, create legal/regulatory flexibility to navigate competing standards, and pursue regional cooperation to build bargaining leverage.
Policy analysis and recommendations grounded in the mechanisms identified via process tracing and comparative cases; intended as prescriptive synthesis rather than empirically demonstrated interventions in the paper. (Based on inferred best-practice interventions; no empirical evaluation/sample size provided.)
There is demand for tooling that bridges evaluation outputs to actionable fixes (e.g., failure-mode libraries, standardized remediation templates, evaluation-to-priority mapping), signaling economic opportunities for third-party tools and consulting services.
Authors' inference based on the documented results-actionability gap and participants' descriptions of pain points; presented as a market implication rather than direct market measurement.
Firms that invest in instrumentation, cross-functional processes, and remediation levers capture more value from LLMs; organizations with better evaluation-to-action pipelines will obtain higher productivity gains and market edge.
Authors' inference from observed heterogeneity among teams in the interviews and comparison of practices in teams that reported more success converting evaluations into changes.
Public investments in standards, verification infrastructure, and public-interest datasets can correct market failures and support trustworthy AI.
Policy recommendation informed by governance and public-good theory and examples from the literature; the claim is prescriptive and not validated by new empirical evidence within the paper.
Policy instruments (law and markets) should be designed to remain institutionally and procedurally responsive to ethical claims that resist full codification (e.g., through participatory governance, oversight mechanisms, equitable redress, care-centered procurement standards).
Normative policy prescriptions derived from the Levinasian diagnosis and case illustrations; proposed measures are normative and not empirically evaluated within the paper.
Integrating Object-Oriented Ontology (OOO) and the material turn enables attention to nonhuman actors and assemblages without collapsing them into human-centered instrumentalism.
Theoretical synthesis of OOO/material-turn literature and argument that this synthesis offers analytic resources for socio-technical assemblages; illustrated conceptually in domains.
Humans who configure and teach agents gain understanding and skills themselves — learning-by-teaching generates human capital accumulation endogenous to agent deployment (bidirectional scaffolding).
Qualitative, naturalistic observations and comparative documentation of users configuring/teaching agents during the one-month study; no randomized assignment or pre/post quantitative skill testing reported.
Models trained primarily on negative constraints will generalize constraint adherence more robustly under distribution shift than models trained primarily on preference rankings.
Presented as a central, experimentally falsifiable prediction derived from the paper's theoretical account; the paper does not present large-scale empirical confirmation and recommends controlled experiments to test this.
Negative examples function as counterfactual eliminators that rule out regions of behavior space, allowing a model to settle on robust acceptable behavior, whereas positive preference signals require continual calibration in a high-dimensional, context-sensitive space.
Informal/structural theoretical argument and analogy to falsification presented in the paper; no direct empirical test reported there demonstrating this exact mechanism.
Regulators may prefer systems that support contestability and audit trails and could mandate argumentation-style explainability in certain sectors.
Speculative policy prediction; no regulatory statements or empirical policy adoption evidence cited.
Better contestability may reduce litigation and regulatory frictions if decisions are transparently defensible.
Speculative legal-economic claim; no case studies or empirical legal analysis provided.
New service layers may emerge (argumentation-as-a-service, audit firms, explanation certification, human-in-the-loop orchestration platforms).
Speculative market/industry evolution claim based on analogous tech-service cretions; no empirical evidence.
New metrics are needed to value resilience (robustness to out-of-distribution events, graceful degradation) in procurement and contracting; performance-based contracts and regulated minimums for oversight mode selection can help align incentives.
Prescriptive recommendation based on gaps identified in procurement and contracting practice; conceptual proposal without empirical testing.
Demand will grow for tools and services that enable oversight (auditability, explainability, safe fallbacks), creating markets for verification, certification, safety middleware, and human-in-the-loop platforms.
Market-structure and demand-side reasoning based on the proposed governance needs; forecast-style projection without empirical market-data analysis.
Allocation decisions should be explicit, auditable, and adaptive — with provisions for overriding, fallbacks, and graceful degradation during unanticipated conditions.
Normative recommendation based on safety and accountability principles combined with crisis-management practices; argued via conceptual analysis and illustrative design features.
Investment in multimodal continual learning, scalable and reliable knowledge-editing methods, and retrieval architectures that guarantee cross-modal consistency is economically justified.
Research/prioritization recommendations based on empirical benchmark findings showing current gaps; argumentation for R&D focus areas.
The findings argue for policies requiring disclosure of training-data timeframes and robust monitoring for time-sensitive factual accuracy in deployed systems.
Policy recommendations in the paper drawing on benchmark results and identified failure modes; prescriptive argumentation rather than empirical policy evaluation.
Models and platforms that offer transparent update mechanisms (frequent data updates, reliable RAG pipelines, clear training snapshot metadata) will have competitive advantages in the market.
Economic and market analysis in implications section recommending transparency and update mechanisms as differentiators; speculative/business-analytical evidence rather than experimental.
Embedding culturally aligned moderation and multi-layer safety orchestration can reduce regulatory frictions and increase adoption in conservative or tightly regulated markets.
Paper claims regulatory and safety economics implications from their safety/moderation architecture; this is an asserted implication rather than an empirically validated outcome in the summary.
The methods used (data quality focus, continual pre-training, model merging, modular product stacks) are potentially transferable to other underrepresented/low-resource languages, lowering barriers to regional AI competitiveness.
Paper posits this policy/transferability implication as an argument in the 'Implications for AI Economics' section; no cross-language experimental evidence provided in the summary.
Fanar 2.0 demonstrates that targeted data curation, continual pre-training, and model-merging can be a viable alternative to the raw-scale pre-training arms race for language-specific competitiveness.
Paper argues this implication based on achieving benchmark gains on Arabic and English using curated data (120B tokens), continual pre-training, model-merging, and a 256 H100 GPU training budget rather than massively larger-scale pre-training.
Oryx provides Arabic-aware image/video understanding and culturally grounded image generation.
Paper identifies Oryx as the vision component with Arabic-aware understanding and culturally grounded generation; no benchmark metrics are provided in the summary.
Exchanging generative modules (rather than raw data) and enabling modular unlearning improves auditability and aligns better with privacy/regulatory compliance than raw-data sharing.
Argument in the paper that module exchange and deterministic module deletion are more compatible with data sovereignty and regulatory requirements; no formal legal validation or compliance testing reported in the summary.
FederatedFactory enables new economic opportunities (module marketplaces, synthetic-data services) and affects incentives by shifting value toward modular generative assets and orchestration rather than raw centralized datasets.
Conceptual and economic discussion in the paper about potential implications; not based on empirical market data—presented as analysis and hypotheses about economic impact.
The single-round exchange decreases communication rounds and associated coordination/network costs compared to typical iterative federated learning.
Protocol design: single exchange of generative modules vs. typical multi-round weight-aggregation loops in standard FL; paper argues reduced networking/coordination cost. (No quantitative network-cost measurements provided in the summary.)
Tools that improve detection or quantification may reduce downstream costs from missed diagnoses or unnecessary follow-ups, improving cost-effectiveness in some scenarios.
Economic modeling and limited observational analyses that extrapolate diagnostic improvements to downstream resource use; direct empirical cost-effectiveness studies are scarce.
Policymakers should support standards for auditability, human‑in‑the‑loop thresholds and training subsidies to reduce coordination failures and make the social benefits of AI adoption more widely shared.
Normative policy recommendation derived from the paper’s analysis of risks, governance needs and distributional concerns; not empirically validated within the paper.
Organisations will invest more in training for AI‑related sensemaking, trust calibration and governance competencies; returns to such training should be evaluated relative to investments in model quality.
Prescriptive inference from the framework and human‑capital theory; supported by referenced literature but not empirically tested in this paper.
Explicit comparative‑advantage allocation will shift the composition of tasks across humans and AI, altering demand for routine versus non‑routine skills and potentially increasing demand for high‑level judgement, oversight and sensemaking skills.
Projected labour‑market implication based on theoretical reasoning and prior literature on task‑based skill demand; not empirically estimated in the paper.
Operationalising the four symbiarchic practices through updated HR systems lets firms capture AI‑enabled productivity gains without eroding trust, ethics or employee well‑being.
Normative claim based on theoretical synthesis and managerial prescription; no empirical testing or field data presented in the paper.
The paper provides a Differentiated Path reference for Emerging Economies to cope with Technological Nationalism.
Claim about the paper's contribution; based on authors' proposed policy framework and recommendations derived from literature review and theoretical analysis; not empirically validated for emerging economies in the excerpt.
The reduction of the AI Model Performance Gap between China and the United States to single digits highlights the new trend of Technology Competition.
Empirical/observational claim stated in the paper; no information in the excerpt about the benchmark metric used for model performance, measurement methodology, time frame, or data sources; 'single digits' not numerically specified.
Supportive regulatory frameworks and digital infrastructure development are important for leveraging AI technologies to improve global trade efficiency.
Study recommendation derived from empirical findings and discussion; this is a policy implication rather than a directly tested empirical claim (no policy evaluation data provided in the summary).
The study provides empirical support for digital transformation theories within financial intermediation.
Authors interpret quantitative results as empirical evidence consistent with digital transformation theories; specific theoretical tests, model fit statistics, and sample information are not included in the summary.
AI-enhanced compliance systems increased regulatory transparency.
Study reports improvements in regulatory transparency as part of operational efficiency gains attributed to AI-driven compliance systems in the quantitative analysis; precise transparency metrics and sample details not provided.
The system demonstrates 100% alignment with GAAP/IFRS regulatory compliance.
Reported regulatory compliance assessment or stakeholder validation claiming full alignment with GAAP/IFRS. (Summary lacks details on the compliance assessment method, criteria, or independent verification; sample/coverage not specified.)
AI has increased the accuracy of patient selection to 80–90%.
Stated performance range for AI-enabled patient selection in the review. The excerpt does not specify the datasets, evaluation metrics (e.g., accuracy vs. AUC), clinical contexts, or sample sizes used to obtain these numbers.
The practical value of the study lies in outlining an analytical framework that can support the design of adaptive workforce strategies, reduce vulnerability to technological disruption, and strengthen the capacity of economies to respond to ongoing digital change.
Claim about the paper's contribution based on the produced analytical framework; the paper presents the framework but does not report empirical validation or outcome measures from real-world implementations.
Integration of data-driven and AI-supported training tools is a critical component for effective reskilling and upskilling.
Argument based on theoretical analysis and review of practices; the paper recommends integration but does not present empirical performance metrics or randomized evaluations of such tools.
The study's implications include policy recommendations to foster responsible AI adoption and data utilization to mitigate economic risks.
Authors extend findings to policy recommendations in the discussion/conclusion of the paper (no specific policy proposals or evaluative evidence provided in the summary).
The research produced a practical framework to guide businesses in effectively leveraging AI and Big Data to navigate market volatility.
The paper's culmination is described as a practical framework derived from its mixed-methods findings (the summary does not provide the framework's components or empirical validation).
A broad-based consumption tax would rebalance a tax system that can no longer depend on taxing individual labor income.
Normative claim in the paper proposing consumption taxation as a corrective mechanism; no empirical evaluation of consumption tax effectiveness included in the excerpt.