Evidence (4333 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Governance
Remove filter
Drawing on analysis of agentic investment firm operational models demonstrating 50-70% cost reductions while maintaining fiduciary standards.
Internal analysis/modeling of agentic investment firm operational models reported by the authors; paper states the 50–70% cost reduction result but provides no sample size or detailed empirical validation in the provided text.
Using machine learning applied to news streams constitutes a practical method to augment existing fiscal surveillance tools.
Paper asserts practical applicability of ML + news for surveillance; presented as recommendation/claim rather than documented large-sample trial in the provided excerpt.
Incorporating news-based signals into machine-learning models can enhance regulatory practice by improving detection of potential fiscal instabilities.
Paper claims an empirical analysis and synthesizes findings linking news-derived signals and ML methods to improved regulatory monitoring; specific datasets, evaluation metrics, and sample sizes are not provided in the excerpt.
The framework offers a replicable model for governments and institutions seeking to proactively support high-potential innovations across sectors.
Paper asserts replicability and applicability to governments/institutions based on the described methods and outputs; no deployment case studies or empirical replication evidence reported in text provided.
A data-driven, foresight-based approach to policy design significantly enhances responsiveness, precision, and resource efficiency in science and technology governance.
Paper concludes this benefit based on its integrated framework, triangulation, Delphi/AHP validation and illustrative mapping; no quantified comparative metrics or experimental evaluation reported in text provided.
These findings provide quantitative foundations for AI capability-threshold governance.
Synthesis/interpretation of model results and empirical validation described in the paper (recommendation/implication).
The paper introduces the Distributed Human Data Engine (DHDE), a socio-technical framework previously validated in biological crisis management, and adapts it for regional economic flow optimization.
Author statement describing the DHDE and asserting prior validation in biological crisis management; adaptation described in paper (methodological description).
The ACT represents the first open-source effort to consolidate data on Africa's evolving HPC landscape, aiming to encourage more transparency from local AI stakeholders and facilitate broader access for AI developers.
Authors' characterization of ACT as a novel, open-source consolidation; assertion based on literature/tools review performed by the authors and on the tool's stated goals.
The results contribute to literature arguing that cloud-based GenAI is a source of enterprise value creation rather than merely an experimental technology.
Paper's stated addition to the existing literature based on the combined empirical and theoretical findings.
Orchestrated systems of smaller, domain-adapted models can mathematically outperform frontier generalist models in most institutional deployment environments.
Formal conditions and comparative analysis derived in the paper plus referenced/claimed empirical support across several domains (frontier lab dynamics, alignment evolution, sovereign AI pressures).
Debiasing via metadata redaction and explicit instructions restores detection in all interactive cases and 94% of autonomous cases.
Intervention experiments in Study 2 where metadata redaction and explicit instructions were applied to interactive assistants (e.g., GitHub Copilot) and autonomous agents (e.g., Claude Code); reported full restoration for interactive and 94% for autonomous.
An increasing number of enterprises are using the label of artificial intelligence merely as a cosmetic embellishment in their annual reports (the phenomenon of 'AI washing' is spreading).
Framing/background claim in the paper's introduction/abstract; implied support from the semantic analysis of annual report texts across Chinese A-share firms over 2006–2024.
There are ethical imperatives of fairness and transparency in automated wealth management, and the paper proposes a roadmap toward sustainable and interpretable financial AI.
Normative analysis and proposed roadmap described in the paper; the excerpt does not provide operationalized fairness metrics, interpretability methods, or evaluation results.
In environments characterized by high-frequency data, non-linear dependencies, and stochastic market regimes, autonomous DRL agents can learn optimal sequential decision-making policies that offer a compelling alternative to static or rule-based allocation strategies.
Argument based on theoretical suitability of DRL for sequential decision problems and the paper's system-level investigation; excerpt does not report specific experimental datasets, sample sizes, benchmarks, or performance metrics.
The integration of Deep Reinforcement Learning (DRL) into portfolio management represents a significant evolution from classical Mean-Variance Optimization and modern econometric frameworks.
Conceptual comparison and synthesis presented in the paper; no empirical sample size or experimental results are provided in the excerpt to quantify the degree of improvement.
Blindfolding (anonymizing identifiers) allows verification of whether meaningful predictive signals persist (i.e., predictions reflect legitimate patterns rather than pre-trained recall of tickers).
Combined methodological-and-result claim: approach described (anonymization) plus stated objective and reported validation (negative controls and reported Sharpe under anonymization). Specific experimental protocol and quantitative results isolating the effect of anonymization are not provided in the excerpt.
On 2025 year-to-date (through 2025-08-01), the system achieved Sharpe 1.40 +/- 0.22 across 20 random seeds.
Backtest/performance claim: reported Sharpe ratio with reported uncertainty and a sample size of 20 seeds; time window specified as 2025 YTD through 2025-08-01. No further details on portfolio construction, leverage, transaction costs, or benchmark adjustment provided in the excerpt.
Regulatory sandboxes offer a flexible and innovation-friendly governance model compared to traditional command-and-control mechanisms.
Normative and comparative analysis within a law & economics framework; no empirical performance data reported in the abstract.
Comparative insights from FinTech identify the institutional design features necessary to ensure the effectiveness and resilience of regulatory sandboxes.
Comparative case-based reasoning drawing on FinTech regulatory sandbox experience (abstract does not report number or selection of cases).
AI regulatory sandboxes may correct specific government failures, including regulatory capture, rent-seeking, and knowledge gaps.
Analytical claims supported by comparative reasoning (FinTech examples) and economic analysis of government failure; no empirical testing or sample size reported in the abstract.
AI regulatory sandboxes facilitate iterative regulatory learning while promoting responsible AI innovation.
Theoretical argument using experimentalist governance concepts and law & economics reasoning; comparative insights referenced but no empirical sample detailed in the abstract.
AI regulatory sandboxes can reduce negative externalities associated with AI deployment.
Conceptual and economic analysis in the paper (no empirical quantification or sample size reported in the abstract).
AI regulatory sandboxes can mitigate information asymmetries between regulators and firms.
Analytical application of an economic analysis of law framework; theoretical argumentation rather than reported empirical measurement in the abstract.
A well-established legal framework for data privacy (e.g., PIPL) enhances the benefits of big data for corporate performance.
Inference drawn from the observed stronger positive big-data effect on firm value after PIPL implementation, as reported by the paper's moderation analysis.
Robust sensitivity tests confirm the main findings, indicating that the results are not driven by model specification or sample selection.
Paper reports multiple robustness/sensitivity checks (unspecified in summary) that the authors state produce consistent results supporting the primary conclusions.
The positive impact of big data on firm performance is strengthened following the implementation of China's Personal Information Protection Law (PIPL).
Moderation/interacted-specification analysis in the paper comparing pre- and post-PIPL periods (or interacting big-data measure with a PIPL indicator), showing a larger positive effect on firm value after PIPL implementation.
The positive effect of big data on firm value operates through improving operational efficiency and reducing costs.
Mechanism analysis reported in the paper indicating mediation/channel tests where big data adoption is associated with measures of operational efficiency and cost reductions, which in turn relate to higher firm value.
Big data application significantly improves firm value.
Results from fixed-effects regressions on the 2007–2021 panel showing a statistically significant positive coefficient for the big-data keyword-frequency measure on firm value (paper reports significance and effect direction).
It is optimal to start taxing AI when cognitive workers start to consider switching to manual jobs.
Analytical result derived from the extended dynamic taxation model and its comparative-static/optimal-policy analysis; the timing rule for introducing an AI tax follows from the model's equilibrium conditions and welfare optimization.
The model implies testable governance diagnostics linking latent fragility to observable patterns: recorded dissent (anonymous vs. formal voting gaps), scenario-set diversity, pipeline and method concentration, and anchor lag.
Theoretical mapping from model primitives and observable quantities to proposed diagnostics; the paper enumerates observable patterns that should correlate with model-implied fragility. This is a theoretical implication rather than an empirically validated claim.
Pidgin significantly outperformed standard English on measures of knowledge transfer across agriculture, education, and health domains.
Aggregate analysis of questionnaire comprehension items (44-item instrument) across domain-specific modules administered to 45 participants; comparative language-performance results reported in study.
Volunteers who used proverbs and vernacular registers were incorporated into local kinship structures, granted traditional titles, and perceived as legitimate development actors rather than outsiders.
Qualitative evidence from participant observation and discourse samples collected during fieldwork; interview and questionnaire items on perceptions of volunteer legitimacy and social integration.
Agricultural techniques taught in Pidgin were nearly universally adopted by recipients.
Self-reported adoption/behavior-change items in the 44-item questionnaire and corroborating qualitative observation of agricultural practice among participants in the sample (N = 45).
Pidgin-mediated interventions achieved large comprehension gains on health messaging, exceeding 30 percentage points compared with standard English.
Quantitative comparison derived from the 44-item field questionnaire (comprehension items) administered to the 45-participant sample; reported percentage-point difference (>30 pp) in health-message comprehension by language of instruction.
Using Cameroon Pidgin English as the primary medium for Peace Corps development work produced substantially better knowledge transfer, uptake, and social legitimacy than standard English.
Mixed-methods field study of Peace Corps interventions in Cameroon's Northwest: 44-item questionnaire administered to 45 participants across agriculture, education, and health; quantitative measures of comprehension and self-reported adoption; supplemented by qualitative observation and discourse samples.
AI adoption will shift fact-checking tasks (more monitoring, less rote verification), creating a need for reskilling and new roles (AI tool operators, analysts); donor and public investments should fund capacity building for local organizations.
Workforce implications inferred from interview reports about changing task mixes and the study's interpretive recommendations.
Investments should prioritize hybrid models where automation provides scale and humans handle contextual, adversarial, and legally sensitive judgments.
Recommendation based on interview findings about AI benefits and limitations and the study's interpretive synthesis.
The study distills context-sensitive best practices for fact-checking in restrictive environments, including safety protocols, local partnerships, and hybrid verification workflows.
Synthesis of findings from document analysis and interviews producing a set of recommended practices documented in the study's outputs.
AI can lower verification costs and scale reach by automating tasks such as classification, clustering, alerting, and translation.
Interview reports from platform staff and interpretive analysis identifying AI-assisted use cases for prioritization, monitoring, and translation.
Community reporting and audience-focused formats are used to improve engagement.
Platform outputs and staff interviews describing deployment of community-reporting mechanisms and tailored audience formats.
Platforms form partnerships with media outlets, academic institutions, and civil-society actors to amplify reach and secure data.
Interview accounts and organizational documents describing cross-sector partnerships and collaboration arrangements.
Transparent workflows and clear labeling are used to build credibility with audiences.
Document analysis of platform outputs and guidelines showing explicit workflow transparency and labeling practices, supported by interview statements.
Platforms emphasize local-language expertise and culturally grounded sourcing as a strategy to improve verification and credibility.
Observed practices and platform guidelines derived from document analysis and staff interviews describing the use of local-language expertise and sourcing.
Practical policy recommendation: require transparent documentation and third‑party auditing for high‑impact LLM deployments and subsidize public‑interest evaluation infrastructure.
Policy prescription supported by the paper's normative and economic analysis; no pilot implementation or empirical evaluation of the recommendation is provided.
Policy levers that can address alignment externalities include disclosure requirements (data provenance, evaluation practices), mandatory participatory evaluation for high‑impact systems, standards for auditing, procurement rules favoring participatory transparency, and liability/certification regimes.
Policy recommendation based on economic and governance reasoning and synthesis of prior regulatory proposals; no policy pilot data or impact evaluation is reported.
Economics research should develop multi‑dimensional metrics capturing welfare, distributional impacts, and autonomy rather than relying on single aggregate accuracy or safety scores.
Prescriptive recommendation grounded in critique of current benchmarking practices and theoretical desiderata; no new metric is empirically validated in the paper.
Dynamic constraints (continuous monitoring, feedback loops, and configurable safety settings that adapt post‑deployment) are preferable to static pre‑deployment-only safety fixes.
Conceptual argument and synthesis of deployment experience and monitoring literature; suggestions for operational tooling and monitoring rather than empirical evaluation.
Participatory governance—includes varied stakeholders such as users, affected communities, domain experts, and regulators in design, evaluation, and deployment decisions—will improve alignment outcomes and legitimacy.
Theoretical and normative argument citing participatory design literature and ethical governance scholarship; paper offers procedural recommendations but no empirical trial of governance models.
Alignment should shift from static, post‑training constraints (one‑off fixes like safety filters or RLHF alone) to dynamic, participatory systems that explicitly protect pluralism, autonomy, and justice.
Normative argument and conceptual synthesis drawing on literature in AI safety, value alignment, and participatory design; prescriptive reasoning rather than original empirical results.
Investment choices in collaboration AI and digital infrastructure become central strategic decisions affecting firms' comparative advantage.
Management literature synthesis and illustrative multinational cases; argument is conceptual without firm‑level comparative empirical data presented in the paper.