Evidence (4793 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Productivity
Remove filter
The analysis implies specific implications for healthcare leadership and procurement (e.g., procurement and leadership should consider incentive and risk-allocation effects, not just task optimisation).
Authors' conclusions/recommendations drawn from the theoretical analysis and typology (prescriptive claim in the paper; no empirical evaluation reported in the abstract).
Autonomous coding agents, able to create branches, open pull requests, and perform code reviews, now actively contribute to real-world projects.
Empirical observations reported in the dataset and study showing agent-originated branches, PRs, and review actions in open-source projects (paper asserts these actions occurred in real projects).
Workplace organization (W) materially modifies the augmentation function so that two firms with identical technology investments can realize 'radically different' augmentation outcomes.
Conceptual claim supported by the paper's theoretical model (phi(D,W)) and cited empirical illustration (Colombia EDIT survey interaction result).
Because other AI systems exhibit similar scaling-law economics, the mechanisms identified extend beyond computer vision, reinforcing that partial automation is often the economically rational long-run outcome, not merely a transitional phase.
Theoretical argument generalized from scaling-law evidence in the paper; no additional cross-domain empirical evidence reported in the summary.
These findings support the practical value of structured intent representation as a robust, protocol-like communication layer for human-AI interaction.
Aggregate interpretation of the experimental results (cross-language variance reduction, model compensation pattern, equivalence of structured frameworks, and user-study improvements).
We further provide initial evidence that this AI-for-AI paradigm can transfer beyond the AI stack through experiments in mathematics and biomedicine.
Reported preliminary experiments in mathematics and biomedicine intended to test transfer beyond the AI development stack.
To our knowledge, ASI-Evolve is the first unified framework to demonstrate AI-driven discovery across three central components of AI development: data, architectures, and learning algorithms.
Authors' claim of primacy based on reported experiments demonstrating AI-driven discovery in pretraining data curation, neural architecture design, and reinforcement learning algorithm design.
The growth of digital platforms contributes to the decentralization of job creation.
Paper cites contemporary data on the growth of digital platforms as part of its analysis (no specific platform-level datasets or sample sizes cited in the abstract).
The paper's predictions are consistent with practitioner reports.
Authors claim qualitative consistency with practitioner reports (no systematic survey/sample size provided in the provided text).
The paper's predictions are consistent with empirical observations from scientific productivity data.
Authors state they compare model predictions to scientific productivity data (no sample sizes or dataset details provided in the provided text).
The paper's predictions are consistent with empirical observations from AI coding benchmarks.
Authors state they compare model predictions to AI coding benchmark results (no sample sizes or specific benchmarks reported in the provided text).
An AI planner that uses a mix of static analysis with AI instructions can create migration plans for very complex code components that are reliably followed by the combination of an orchestrator and coders, using AI-generated example-based playbooks.
Methodological description and reported demonstrations in the paper (planner + orchestrator + coders following playbooks); no numeric sample size reported in abstract.
With experience, users issue more targeted queries and engage more deeply with supporting citations.
Longitudinal analysis of user behavior in the Asta dataset showing changes over time/with experience: increased use of targeted queries and higher engagement (clicks/inspect actions) with citations.
Users treat generated responses as persistent artifacts, revisiting and navigating among outputs and cited evidence in non-linear ways.
Interaction-log analysis showing patterns of revisits, non-linear navigation between generated outputs and cited evidence within sessions in the Asta dataset.
Users treat the system as a collaborative research partner, delegating tasks such as drafting content and identifying research gaps.
Qualitative and quantitative analysis of interaction logs in the Asta dataset showing user behaviors where the system is used to draft content and identify gaps (examples and aggregated counts described in paper).
Users submit longer and more complex queries than in traditional search.
Comparative analysis of query length/complexity in the Asta Interaction Dataset (>200,000 queries) versus traditional search baselines (as reported in the paper); measurement of query length and complexity metrics across logs.
ASR-assisted transcription offers a practical pathway toward scalable, technology-supported documentation of endangered languages.
Authors' interpretive conclusion based on the corpus creation, ASR model performance (CER ~15%), and reported reductions in transcription time/cognitive load; presented as a recommendation/implication rather than a directly measured outcome.
ASR integration can substantially reduce cognitive load for transcribers.
Paper reports evaluation of ASR assistance including cognitive-load outcomes (authors claim cognitive load is reduced); details of measurement instrument, sample size, and statistical results are not given in the abstract.
ASR integration can substantially reduce transcription time.
Paper reports an evaluation of the impact of ASR assistance on the efficiency of speech transcription (comparison of ASR-assisted vs manual transcription). The abstract asserts a substantial reduction in transcription time but does not provide numeric details in the provided text.
Drawing on analysis of agentic investment firm operational models demonstrating 50-70% cost reductions while maintaining fiduciary standards.
Internal analysis/modeling of agentic investment firm operational models reported by the authors; paper states the 50–70% cost reduction result but provides no sample size or detailed empirical validation in the provided text.
The proposed system architectures and findings provide practical implications for future development of agentic AI systems for engineering design.
Concluding/implicational claim based on the methods and experimental findings reported in the paper (battery pack design experiments); no empirical test of 'practical implications' is provided in the excerpt.
Fostering digital transformation alongside workforce reskilling and innovation-ecosystem development is essential for sustainable industrial growth and strengthening Kazakhstan’s global economic position.
Policy and strategic recommendations based on the study's empirical results, case studies, and macro-level index comparisons.
Digital transformation combined with workforce retraining optimizes labor costs and enhances productivity.
Synthesis of enterprise-level case examples and aggregated regression/correlation findings at industry and national levels that link digitalization and retraining programs to labor-cost and productivity indicators.
Overall, the DRL framework enhances traffic capacity and fuel efficiency without compromising safety.
Aggregate interpretation of simulation results comparing DRL-based AV control to IDM across capacity, fuel efficiency, and safety metrics within the simulated scenarios. Specific safety metrics and sample sizes are not described in the claim text.
These results establish agent scaling as a practical and effective axis for HLS optimization.
Synthesis/interpretation of empirical results (including mean 8.27× speedup and per-benchmark gains) reported in the paper.
Across benchmarks, agents consistently rediscover known hardware optimization patterns without domain-specific training.
Qualitative and empirical observations across the evaluated benchmarks (12) reporting that agents found recognized hardware optimization patterns despite no hardware-specific training.
This work demonstrates the technical feasibility of scalable, AI-augmented quality assessment for early childhood education and lays a foundation for continuous, inclusive AI-assisted evaluation enabling systemic improvement and equitable growth.
Overall results of dataset release, Interaction2Eval performance (agreement), and deployment efficiency reported in the paper; used by the authors to argue broader feasibility and potential systemic impact.
AI-assisted monitoring could shift assessment practice from annual expert audits to monthly AI-assisted monitoring with targeted human oversight.
Authors' synthesis combining dataset-scale results, Interaction2Eval performance (agreement), and deployment efficiency gains to argue feasibility of more frequent monitoring.
The work advances theory on human performance in complex negotiations and offers validated design guidance for interactive systems.
Authors' stated contributions: theoretical advancement and validated design guidance, grounded in the presented empirical results and the validated visualization tested in the N=32 experiment.
The paper introduces the Distributed Human Data Engine (DHDE), a socio-technical framework previously validated in biological crisis management, and adapts it for regional economic flow optimization.
Author statement describing the DHDE and asserting prior validation in biological crisis management; adaptation described in paper (methodological description).
This systematic framework can help predict at a detailed level where today's AI systems can and cannot be used and how future AI capabilities may change this.
Interpretive/utility claim: authors argue that the ontology plus classification results serve as rough predictive tools for AI applicability across work activities.
EnterpriseLab provides enterprises a practical path to deploying capable, privacy-preserving agents without compromising operational capability.
Conclusion drawn by the authors based on the platform design and the reported empirical results (performance parity with GPT-4o, cost reductions, benchmark robustness). The abstract offers this as a high-level takeaway rather than a quantified empirical claim.
The results contribute to literature arguing that cloud-based GenAI is a source of enterprise value creation rather than merely an experimental technology.
Paper's stated addition to the existing literature based on the combined empirical and theoretical findings.
When compared to baseline approaches, the ARL-based model's accuracy in revenue and price optimization decreased by less than 20%, indicating that it can adapt and optimize pricing techniques in intricate, cutthroat markets.
Reported experimental comparison versus baselines (fixed/rule-based and cost-plus); specific metrics, dataset size, and whether 'decrease' refers to error or accuracy are not clarified in the excerpt.
Our results substantiate the potential of large language models as a foundational pillar for high-fidelity, scalable decision simulation and latter analysis in the real economy based on foundational database.
High-level conclusion drawn from the paper's experiments and methodological contributions; generalization claim asserting LLMs' potential as foundational tools for scalable, high-fidelity decision simulation.
Experiments demonstrate that our framework achieves improved simulation stability compared to existing economic and financial LLM simulation baselines.
Empirical claim: experiments vs. baselines showing improved simulation stability (paper statement that framework improved simulation stability, without quantitative details in the excerpt).
Experiments demonstrate that our framework achieves significant improvements in purchase quantity prediction compared to existing economic and financial LLM simulation baselines.
Empirical claim: experiments comparing MALLES against existing baselines; paper reports 'significant improvements' in purchase quantity prediction (no numerical values provided in the excerpt).
Experiments demonstrate that our framework achieves significant improvements in product selection accuracy compared to existing economic and financial LLM simulation baselines.
Empirical claim: experiments comparing MALLES against existing economic and financial LLM simulation baselines; paper reports 'significant improvements' in product selection accuracy (no numerical values provided in the excerpt).
This preference-learning approach enables the models to internalize and transfer latent consumer preference patterns, thereby mitigating the data sparsity issues prevalent in individual categories.
Claim based on the paper's reported approach: cross-category post-training and transfer of latent preferences; supported by experiments (paper states mitigation of data sparsity).
A well-established legal framework for data privacy (e.g., PIPL) enhances the benefits of big data for corporate performance.
Inference drawn from the observed stronger positive big-data effect on firm value after PIPL implementation, as reported by the paper's moderation analysis.
Robust sensitivity tests confirm the main findings, indicating that the results are not driven by model specification or sample selection.
Paper reports multiple robustness/sensitivity checks (unspecified in summary) that the authors state produce consistent results supporting the primary conclusions.
The positive impact of big data on firm performance is strengthened following the implementation of China's Personal Information Protection Law (PIPL).
Moderation/interacted-specification analysis in the paper comparing pre- and post-PIPL periods (or interacting big-data measure with a PIPL indicator), showing a larger positive effect on firm value after PIPL implementation.
The positive effect of big data on firm value operates through improving operational efficiency and reducing costs.
Mechanism analysis reported in the paper indicating mediation/channel tests where big data adoption is associated with measures of operational efficiency and cost reductions, which in turn relate to higher firm value.
Big data application significantly improves firm value.
Results from fixed-effects regressions on the 2007–2021 panel showing a statistically significant positive coefficient for the big-data keyword-frequency measure on firm value (paper reports significance and effect direction).
It is optimal to start taxing AI when cognitive workers start to consider switching to manual jobs.
Analytical result derived from the extended dynamic taxation model and its comparative-static/optimal-policy analysis; the timing rule for introducing an AI tax follows from the model's equilibrium conditions and welfare optimization.
For AI agent tool design, surfacing contextual information outperforms prescribing procedural workflows.
Authors' conclusion drawn from the suite of experiments (GraphRAG vs TDD prompting vs auto-improvement) showing better regression reduction and/or resolution when contextual information is surfaced.
An autonomous auto-improvement loop raised resolution from 12% to 60% on a 10-instance subset with 0% regression.
Reported experiment on a 10-instance subset where an auto-improvement loop was applied (numbers provided in the excerpt).
Smaller models benefit more from contextual information (which tests to verify) than from procedural instructions (how to do TDD).
Inferred from comparative results across models (Qwen3-Coder 30B vs Qwen3.5-35B-A3B) and interventions (contextual test-surfacing vs TDD prompting) reported in the paper.
When deployed as an agent skill, GraphRAG improved resolution from 24% to 32%.
Empirical comparison reported in the evaluation on SWE-bench Verified (same experimental context as above).
TDAD's GraphRAG workflow reduced test-level regressions by 70% (from 6.08% to 1.82%).
Empirical result reported from the SWE-bench Verified evaluation using the GraphRAG workflow (sample details: Qwen3-Coder 30B on 100 instances and Qwen3.5-35B-A3B on 25 instances as reported).