Evidence (8066 claims)
Adoption
5586 claims
Productivity
4857 claims
Governance
4381 claims
Human-AI Collaboration
3417 claims
Labor Markets
2685 claims
Innovation
2581 claims
Org Design
2499 claims
Skills & Training
2031 claims
Inequality
1382 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 417 | 113 | 67 | 480 | 1091 |
| Governance & Regulation | 419 | 202 | 124 | 64 | 823 |
| Research Productivity | 261 | 100 | 34 | 303 | 703 |
| Organizational Efficiency | 406 | 96 | 71 | 40 | 616 |
| Technology Adoption Rate | 323 | 128 | 74 | 38 | 568 |
| Firm Productivity | 307 | 38 | 70 | 12 | 432 |
| Output Quality | 260 | 71 | 27 | 29 | 387 |
| AI Safety & Ethics | 118 | 179 | 45 | 24 | 368 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 75 | 37 | 19 | 312 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 74 | 34 | 78 | 9 | 197 |
| Skill Acquisition | 98 | 36 | 40 | 9 | 183 |
| Innovation Output | 121 | 12 | 24 | 13 | 171 |
| Firm Revenue | 98 | 35 | 24 | — | 157 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 87 | 16 | 34 | 7 | 144 |
| Inequality Measures | 25 | 76 | 32 | 5 | 138 |
| Regulatory Compliance | 54 | 61 | 13 | 3 | 131 |
| Task Completion Time | 89 | 7 | 4 | 3 | 103 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 33 | 11 | 7 | 98 |
| Wages & Compensation | 54 | 15 | 20 | 5 | 94 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 27 | 26 | 10 | 6 | 72 |
| Job Displacement | 6 | 39 | 13 | — | 58 |
| Hiring & Recruitment | 40 | 4 | 6 | 3 | 53 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 11 | 6 | 2 | 41 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 6 | 9 | — | 27 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
The benefits of AI for energy justice are concentrated in China’s advanced eastern region.
Spatial heterogeneity analysis reported in the paper showing stronger positive effects in the eastern region compared to other regions.
The positive effect of AI on energy justice is amplified by better digital infrastructure.
Heterogeneity/interaction analysis reported in the paper showing larger AI effects where digital infrastructure is stronger.
The positive effect of AI on energy justice is amplified by stricter environmental regulations.
Heterogeneity/interaction analysis reported in the paper showing stronger AI effects in contexts with stricter environmental regulation.
AI’s positive effect on energy justice is mediated by reduced industrial density.
Mediation/pathway analysis reported in the paper identifying reductions in industrial density as a mechanism.
AI’s positive effect on energy justice is mediated by higher energy prices.
Reported mediation/pathway results indicating higher energy prices are a channel for AI’s impact on the energy justice index.
AI’s positive effect on energy justice is mediated by green innovation.
Mediation/pathway analysis in the paper identifies green innovation as a mechanism through which AI affects energy justice.
AI’s positive effect on energy justice is mediated by improved energy efficiency.
Mediation/pathway analysis reported in paper identifying energy efficiency as one mechanism linking AI adoption to energy justice improvements.
AI adoption significantly enhances overall energy justice.
Panel regression analysis using the constructed energy justice index as outcome; significance reported in findings (based on the stated empirical results across 30 provinces, 2008–2022).
GenAI implementations that are strategically deployed in managed Azure cloud infrastructure provide a positive ROI over time when aligned with business processes, enterprise architecture, and performance metrics.
Conclusion drawn from the paper's mixed-method analysis (quantitative ROI modelling, cost–benefit analysis, and case study synthesis).
Close coupling among Azure OpenAI Service, Azure Machine Learning, and cost governance tooling (FinOps) significantly decreases overall cost of ownership and enhances scalability and compliance.
Architectural analysis of Azure-native GenAI services and cost/governance tooling reported in the paper.
Measurable ROI from GenAI on Azure is mainly driven by improvements in productivity, optimization of operational costs, faster decision making, and increased speed of innovation across business functions.
Reported results from the paper's mixed-method study combining quantitative ROI modelling and cost–benefit analysis plus qualitative synthesis of secondary enterprise case studies.
Microsoft Azure has become one of the first enterprise-scale platforms facilitating GenAI-driven change.
Statement in the paper's abstract asserting Azure's market position as an early enterprise-scale platform for GenAI.
This synthesis bridges the gap between values and practice, offering a policy-ready model for secure and sustainable AI governance.
Authors' concluding claim that their integrated governance risk framework and risk-tiering matrix operationalize ethical principles into auditable technical controls and are policy-ready.
The study aligns its integrated risk-tiering model with Sustainable Development Goal 9 on industry, innovation and infrastructure.
Authors state that the developed integrated risk-tiering model is aligned with SDG 9 as part of the study framing and intended policy relevance.
The analysis produced a heat map of governance frameworks, a co-occurrence network of themes, a cluster analysis of framework coverage and an integrated governance risk framework supported by a risk-tiering matrix.
Authors report specific analytical outputs (heat map, co-occurrence network, cluster analysis) and that they developed an integrated governance risk framework with a risk-tiering matrix based on their analysis.
Our empirics demonstrate that self-evolving AI offers a scalable and interpretable paradigm.
Empirical results on the U.S. equity market are cited as evidence; the paper claims scalability and interpretability based on those empirical demonstrations and the architecture of the system.
Applying this methodology to the U.S. equity market, long-short portfolios formed on the simple linear combination of signals deliver a return of 59.53% (annualized).
Empirical backtest/application to the U.S. equity market reported in the paper; specific annualized return percentage is provided. Sample period, universe, and number of observations not stated in the excerpt.
Applying this methodology to the U.S. equity market, long-short portfolios formed on the simple linear combination of signals deliver an annualized Sharpe ratio of 3.11.
Empirical backtest/application to the U.S. equity market reported in the paper; specific performance metric (annualized Sharpe) is provided. Sample period, universe, and number of observations not stated in the excerpt.
To mitigate data snooping biases, the closed-loop system imposes strict empirical discipline through out-of-sample validation and economic rationale requirements.
Description of model validation protocol in the paper (use of out-of-sample validation and economic rationale filters); supports claim that these steps are used to reduce data-snooping risk.
The approach operationalizes the model as a self-directed engine that endogenously formulates interpretable trading signals (rather than relying on sequential manual prompts).
Methodological description and implementation details in the paper describing how the model generates signals autonomously and interpretable outputs; empirical example applied to U.S. equity market is referenced to illustrate operation.
We develop an autonomous framework for systematic factor investing via agentic AI.
Statement of methodological contribution in the paper (framework description); no sample size or empirical test required for the descriptive claim.
The technology particularly benefits less experienced practitioners by providing comprehensive starting points for legal research, while experienced attorneys can use it for quality control and initial drafts.
Authors' interpretation of AI outputs from the experiment and reasoning about how those outputs map onto different practitioner needs (qualitative judgment).
The analysis reveals AI’s potential to transform law firm economics by dramatically reducing research time while maintaining analytical quality, though careful attorney oversight remains essential.
Inference from the experimental finding that four AI systems produced substantive analysis comparable to junior-associate work on one transcript and the stated observation about traditional research time (8–40 hours); authors' qualitative judgment about economic implications and need for oversight.
Statutory and regulatory citations proved generally accurate and useful.
Authors' examination of statutory and regulatory references produced by the four AI engines in the experiment, judged to be generally correct and helpful.
All four engines successfully spotted legal issues, assessed claim strengths and weaknesses, and suggested follow-up investigation—tasks that traditionally required eight to forty hours of junior attorney research time.
Observed outputs from the four AI engines on the single transcript showing issue-spotting, strengths/weaknesses assessment, and suggested follow-ups; comparison to typical junior attorney research time (stated as 8–40 hours).
Contemporary generative AI performs sophisticated legal analysis comparable to experienced associates, correctly identifying major employment law claims including ADA violations, Title VII discrimination, OSHA retaliation, FMLA interference, and workers’ compensation retaliation.
Qualitative assessment of outputs from the four AI engines applied to the single hypothetical transcript; comparison against expected legal claims (authors' judgment that outputs matched those an experienced associate would produce).
Four major generative AI engines—DeepSeek, Claude, ChatGPT, and Grok—are useful legal analysis tools for employment law practitioners.
Experimental evaluation in which a single hypothetical client interview transcript was submitted to each of the four AI systems and their outputs were assessed by the authors.
Policy recommendations: increase investment in AI research and expansion; promote AI-driven robotics in key sectors; provide targeted skilling programs for elderly workers; invest in digital infrastructure and the ageing industry; and leverage and develop elderly human capital to support inclusive and sustainable economic development.
Paper discussion/conclusion draws policy implications based on empirical finding that AI adoption mitigates negative ageing effects on GDP growth.
Robustness checks using the old-age dependency ratio as the proxy for ageing deliver consistent results.
Paper reports robustness verification: replacing the primary ageing measure with the old-age dependency ratio yields similar threshold/mitigation findings.
When AI adoption (industrial robot penetration) surpasses a critical threshold, the negative effect of ageing on GDP growth is significantly mitigated.
Threshold interaction result from panel threshold regression: AI adoption (robot penetration) as threshold variable; paper reports that beyond a critical robot-adoption threshold the negative ageing–GDP relationship is significantly weakened.
Through a comparative analysis of Pax Romana, Pax Britannica, Pax Americana, and the emerging U.S. techno-security architecture, the article demonstrates continuity in the logic of hegemonic control centered on infrastructures.
Comparative historical analysis of four hegemonic/regime examples as described in the paper; methodological approach is comparative and qualitative (no quantitative sample size given).
Hegemonic orders can be conceptualized as historically specific logistical regimes — the material basis of hegemony evolves but the underlying logic remains constant: control over the infrastructures that organize global circulation.
Conceptual claim grounded in synthesis of structural power theory, global value chain analysis, and infrastructure studies and illustrated through comparative historical examples (Pax Romana, Pax Britannica, Pax Americana, emerging U.S. techno-security architecture).
The article develops a theoretical framework of logistical hegemony to explain how infrastructures, chokepoints, and global production networks structure the exercise of power in the world economy.
Primary claim of the paper: theoretical development drawing on structural power theory, global value chain analysis, and infrastructure studies; conceptual/theoretical argumentation rather than empirical sample-based evidence.
The specification provides mechanisms for interoperability between institutions.
Design claim in the specification describing mechanisms enabling institutional interoperability.
ACP operates as an additional layer on top of RBAC and Zero Trust, without replacing them.
Design statement in the specification describing ACP's relationship to existing RBAC and Zero Trust architectures.
ACP defines the mechanisms of cryptographic identity, capability-based authorization, deterministic risk evaluation, verifiable chained delegation, transitive revocation, and immutable auditing that a system must implement for autonomous agents to operate under explicit institutional control.
List of mechanisms and required features presented in the specification text.
ACP is the admission control layer between agent intent and system state mutation: before any agent action reaches execution, it must pass a cryptographic admission check that validates identity, capability scope, delegation chain, and policy compliance simultaneously.
Explicit behavioural/design claim in the specification text describing the admission-control role and the checks performed prior to action execution.
ACP is a formal technical specification for governance of autonomous agents in B2B institutional environments.
Stated in the v1.13 specification header/abstract and repository description (specification text and repository link provided).
Organizational support and continuous learning are important to maximize the benefits of AI integration in startup environments.
Conclusions drawn from thematic analysis of interviews with 12 startup employees emphasizing need for organizational support and ongoing learning.
AI functions as a workforce augmentation tool that enhances human capabilities rather than replacing employees.
Reported perceptions from 12 startup employees in semi-structured interviews; thematic coding indicated view of AI as augmentation rather than replacement.
Most employees demonstrated progressive adjustment and competency improvement over time after initial adaptation.
Interview data from 12 startup employees with thematic analysis indicating progressive adjustment and competency gains over time.
AI improves employee performance by supporting more accurate decision-making and increasing work effectiveness and output quality.
Findings from semi-structured interviews of 12 startup employees, analyzed via thematic coding and frequency scoring, reporting improved decision accuracy and output quality with AI support.
AI integration contributes to competency development, particularly in digital literacy, analytical thinking, and adaptive learning.
Qualitative semi-structured interviews with 12 startup employees; thematic coding highlighted competencies (digital literacy, analytical thinking, adaptive learning).
AI significantly enhances employee productivity by accelerating task completion, reducing manual workload, and improving workflow efficiency.
Qualitative study using semi-structured interviews with 12 startup employees; data analyzed with thematic coding, frequency scoring, and visualized analysis.
Experiments highlight a reward anatomical structure that balances income, profit, efficiency, fairness, and customer retention, moving beyond income-only goals.
Experimental design / reward engineering reported in paper; claim supported by experiments (no quantitative metrics or sample size given in excerpt).
Training strength is validated by benchmarking against fixed, rule-based models and cost-plus in controlled experimentation.
Paper reports controlled experiments benchmarking ARL models against fixed/rule-based and cost-plus baselines; specific experimental design and sample sizes not provided in excerpt.
Inventory challenges are addressed by utilizing a curated dataset that has been enhanced through feature engineering, transformation, and systematic cleaning, providing reliable inputs for training.
Methodological claim about dataset curation and preprocessing used to train ARL agents; no dataset size or quantitative validation reported in excerpt.
Profitability in a dynamic marketplace is enhanced through an Adaptive Reinforcement Learning (ARL)-based pricing framework that utilizes Q-Learning and Deep Q-Networks (DQN) for real-time optimization in response to changing market conditions, competition, and inventory levels.
Paper proposes and experiments with an ARL-based pricing framework (methods include Q-Learning and DQN); validation claimed via benchmarking/controlled experimentation against baselines (details not provided in excerpt).
Dynamic pricing is crucial for maximizing revenue and maintaining competitiveness in markets with fluctuating demand, perishable goods, and diverse customer preferences.
Conceptual claim stated in paper's introduction/motivation; no empirical sample or experiment specified in the statement.
In the long term, big data promotes sustained improvements in individuals’ welfare.
Theoretical long-run growth analysis in the model showing that sustained data sharing leads to long-run welfare improvements (analytic/model-based, no empirical/sample data).