Evidence (3492 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 736 | 1615 |
| Governance & Regulation | 664 | 329 | 160 | 99 | 1273 |
| Organizational Efficiency | 624 | 143 | 105 | 70 | 949 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 348 | 109 | 48 | 322 | 836 |
| Output Quality | 391 | 120 | 44 | 40 | 595 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 275 | 143 | 62 | 34 | 521 |
| AI Safety & Ethics | 183 | 241 | 59 | 30 | 517 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 105 | 40 | 6 | 187 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 78 | 8 | 1 | 151 |
| Regulatory Compliance | 69 | 64 | 14 | 3 | 150 |
| Training Effectiveness | 81 | 15 | 13 | 18 | 129 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
Innovation
Remove filter
DPPs generate high-quality, structured product and lifecycle data that are non-rivalrous and highly reusable, raising firm-level incentives to invest in AI models (forecasting, optimization, provenance verification) that exploit this data to capture value across production, secondary markets, and services.
Economic/technical implication drawn from the empirical characterization of DPP data and stakeholder interviews; this is an inferential claim linking DPP data properties to incentives for AI investment rather than a directly measured outcome in the surveys.
Practical DPP deployment must combine standards, governance, and user-centric design to unlock circular-economy benefits.
Inference from empirical mapping of barriers/drivers (survey and qualitative stakeholder input) and multivariate analyses showing interplay of technical and organizational factors; sample sizes not reported.
DPPs should be seen as both technical data platforms and participatory tools that enable collaborative value creation and responsible consumption (thus supporting SDG 12: responsible consumption and production).
Conceptual interpretation synthesized from empirical findings (surveys + multivariate analyses) and theoretical framing in the paper; empirical grounding via stakeholder responses but largely a conceptual contribution.
Successful DPP adoption requires matching technical functionalities (data granularity, interoperability, user interfaces) with firm-level readiness and strategies to engage different consumer segments.
Logistic regression and PCA mapping relationships among DPP features, organizational practices and consumer profiles arising from the two online surveys and mixed-method analysis; sample sizes not reported.
DPPs facilitate knowledge sharing and open innovation across firms, embedding sustainability and knowledge management into operational practice.
Qualitative and survey responses from industry stakeholders in the two sectors; analyses reported include mapping of cross‑firm knowledge exchange and organizational practices (methods: mixed methods, logistic regression/PCA); sample sizes not reported.
DPPs enhance transparency and traceability across complex supply chains, enabling material circularity and more resilient sourcing decisions.
Survey-based evidence and multivariate analyses (PCA, logistic regression) from stakeholders in Italian fashion and cosmetics indicating perceived/observed links between DPP functionalities (data granularity, interoperability) and traceability/circularity outcomes; sample sizes not reported.
Digital Product Passports (DPPs) function as a socio-technical, cognitive infrastructure that, when DPP technical capabilities are aligned with organizational readiness and consumer engagement, materially support circularity (raw-material reuse), supply-chain resilience, and cross-firm knowledge exchange, thereby turning sustainability from a compliance burden into a source of innovation and value in fashion and cosmetics.
Mixed-methods empirical study in Italian fashion and cosmetics using two online surveys, PCA and logistic regression to map relationships among technical features, organizational practices and consumer profiles; sample sizes not reported in the summary.
Economists and AI practitioners will need capacity-building in Earth-system knowledge to ensure models capture cumulative and systemic environmental risks rather than only firm-level signals.
Recommendation based on gap analysis between current disciplinary skills and systemic-environmental modeling needs; no survey or training-efficacy data offered.
There is a need for standards for data provenance, auditability, and adversarial robustness to prevent greenwashing and model manipulation.
Policy recommendation grounded in conceptual risk analysis; no technical standard proposals or threat-model evaluations provided.
Open environmental disclosure data supports reproducible empirical research in AI economics (causal inference, counterfactuals, macro-financial modeling) on effects of regulation and capital flows on environmental outcomes.
Logical argument about data availability enabling reproducible research; no empirical examples or reproducibility metrics provided.
More reliable environmental disclosures enable algorithmic investors and market models to price externalities more accurately and to implement sustainability-aligned strategies at scale.
Conceptual argument about improved information enabling market mechanisms; no empirical market-impact study included.
Open data facilitates automated, lower-cost reporting tools (NLP extraction, sensor/IoT integration, ETL pipelines) that reduce administrative burden and increase reporting frequency and timeliness.
Conceptual claim linking open standardized data to automation potential; no implemented cases or cost estimates provided.
Improved, standardized environmental disclosures improve training data quality for predictive models, reducing measurement error and bias.
Theoretical claim about data quality effects on model performance; no empirical evaluation provided.
Better, standardized, open environmental data unlocks AI/ML opportunities, enabling scalable models for firm- and system-level environmental risk assessment, scenario analysis, stress testing, and portfolio optimization.
Conceptual implications and use-case enumeration; no empirical model-building or benchmarking presented.
Applying assurance standards and regulatory oversight analogous to financial reporting will improve environmental data quality.
Normative recommendation; argument by analogy to financial assurance practices; no empirical assessment included.
Standardization (common taxonomies, units, definitions) and machine-readability are necessary to ensure comparability of environmental disclosures.
Methodological recommendation based on conceptual analysis of data interoperability issues; no empirical demonstration provided.
Treating environmental data with the same rigor as financial data (governance, standardization, auditing) would markedly improve investor, regulator, and public agency ability to assess environmental pressures, hold firms accountable, and align capital with sustainability objectives.
Conceptual causal claim argued from analogy to financial reporting; no empirical testing provided in the paper.
Corporate sustainability reporting is a powerful lever for changing corporate behavior; improving it can influence investment flows and corporate practice.
Normative/conceptual claim supported by literature synthesis and policy reasoning rather than new empirical testing.
The 2018 Supply Chain Innovation and Application Pilot Program can be used as a quasi‑natural experiment (treatment) to identify causal effects of SCD on firm outcomes.
Difference-in-differences design comparing treated (pilot-designated) versus control firms pre/post-2018; treatment defined by designation as pilot enterprise under the 2018 program.
The SCD → green innovation effects are larger for large firms (by firm size).
Heterogeneity analysis splitting sample by firm size (large vs small) with results indicating stronger SCD effects on green innovation for larger firms.
The SCD → green innovation effects are larger for state‑owned enterprises (SOEs).
Heterogeneity analysis by ownership type (SOE vs non‑SOE) showing larger and more significant coefficients for SOEs in the SCD effect on green innovation.
Carbon information disclosure (CID) is a key mediating channel: SCD increases the likelihood and quality of CID, which in turn promotes substantive green innovation.
Mediation analysis using observed CID indicators (likelihood/quality of carbon disclosure) in a causal pathway framework; SCD raised CID metrics in first-stage regressions and CID was positively associated with subsequent substantive green innovation in mediation tests.
Policy responses (active labor-market interventions, reskilling, lifelong learning, social insurance, redistribution) are needed to manage transitional inequality caused by AI-driven structural shifts in labor demand.
Policy implication drawn from reviewed empirical and theoretical literature on labor-market transitions and distributional impacts; presented as a recommendation without new empirical evaluation in this paper.
Economists should refine methods to measure AI adoption and incorporate AI-driven productivity gains into growth accounting while accounting for measurement challenges (quality change, task reallocation).
Methodological recommendation based on the review's identification of measurement difficulties in the existing empirical literature; the paper itself provides conceptual guidance rather than new measurement results.
AI has materially increased operational efficiency and productivity in industry, changing production processes and firm organization.
Qualitative integration of prior empirical studies and firm-level case studies cited in the literature review (industry analyses, adoption case examples); the paper itself does not provide new quantitative estimates or causal identification.
Policy priorities to improve China's digital services exports include: strengthening participation in global rule‑making, building internationally competitive platforms and cloud infrastructure, expanding targeted support for firms (especially SMEs) to internationalize, and refining data governance to balance security/privacy with cross‑border interoperability.
Derived recommendations from the integrative literature and policy review and comparative diagnosis (interpretive, not empirically validated within the paper).
Participation in international rule formation (standards and data rules) influences which AI/data standards prevail and therefore which firms gain comparative advantage in global markets.
Conceptual argument and policy literature reviewed on standards, governance, and competitive advantage (qualitative synthesis).
China's export competitiveness in digital services depends critically on participation in international rule‑making, stronger platform infrastructure, targeted support for firms going global, and improved data governance.
Synthesis of reviewed studies, institutional diagnosis, and comparative analysis (interpretive policy conclusion rather than empirically quantified effect sizes).
Digital services have become a key indicator of a country's export competitiveness because they reshape global trade structure and labor specialization within global value chains.
Review of theoretical mechanisms and empirical literature in the integrative review; comparative policy analysis (qualitative synthesis rather than original quantification).
SlideFormer generalizes beyond a single GPU vendor (the design achieves high utilization on both NVIDIA and AMD GPUs).
Reported experiments and utilization measurements on both NVIDIA (RTX 4090) and AMD GPUs showing sustained >95% peak performance, implying cross-vendor applicability. The summary does not specify which AMD models or the breadth of tested kernels.
Custom Triton kernels and advanced I/O integration remove key bottlenecks in single-GPU fine-tuning pipelines and contribute to the observed throughput gains.
Paper reports the use of custom Triton kernels for performance-critical primitives and improved I/O integration; throughput gains (1.40×–6.27×) are attributed in part to these optimizations. The summary does not isolate ablation results quantifying each optimization's contribution.
Heterogeneous memory management (multi-tier placement across GPU, CPU, and storage) materially reduces peak on-device memory requirements.
Authors describe an efficient memory layout and placement strategy across GPU, host RAM, and storage tiers and report lowered peak device memory use (≈2× reduction). The summary does not include low-level placement parameters or traces.
SlideFormer sustains >95% peak performance (high utilization) on both NVIDIA and AMD GPUs.
Reported sustained peak utilization measurements on experiments run on NVIDIA (e.g., RTX 4090) and AMD GPUs; the summary states >95% peak performance but does not give per-workload/utilization measurement methodology.
SlideFormer supports up to 8× larger batch sizes and up to 6× larger models on the same GPU relative to prior single-GPU baselines.
Reported comparisons to prior single-GPU baselines measuring achievable batch size and model-size capacity on the same GPU; exact baselines, workloads, and experimental configurations are not detailed in the summary.
SlideFormer reduces peak CPU and GPU memory usage by approximately 2× (roughly halving memory requirements).
Authors report peak memory measurements showing about a 2× reduction in both GPU and CPU memory compared to baselines; memory accounting method and baselines are not fully specified in the summary.
SlideFormer achieves 1.40×–6.27× higher throughput versus baseline systems.
Quantitative evaluation comparing throughput (reported as tokens/sec or updates/sec) against state-of-the-art single-GPU and multi-GPU fine-tuning pipelines (baselines are unnamed in the summary). Measurements reported across single-GPU experiments (hardware includes RTX 4090 and AMD GPUs).
SlideFormer enables fine-tuning very large LLMs (reported up to 123B+ parameters) on a single GPU (e.g., RTX 4090).
Authors report experiments and capability claims for single-GPU setups including an NVIDIA RTX 4090; model size stated as 123B+ in the paper summary. Details on exact model family, sequence length, or batch size used for the 123B+ claim are not enumerated in the summary.
Clear agent identity and provenance simplify liability attribution and enable markets for certified components, attestation services, and compliance tooling.
Legal/economic reasoning about traceability and liability plus systems design suggestions; no legal case analysis or market data presented.
Lifecycle service models (leasing, 'agent as a service', update/maintenance contracts) will become economically important to manage long‑lived physical assets with fast‑moving AI stacks.
Business model reasoning and analogy to service models in other capital‑intensive sectors; no market empirical study or business case analysis provided.
Observability and attestation reduce uncertainty for insurers and regulators, lowering risk premia and insurance costs for agent deployments.
Argument from information economics/insurance theory and analogy to fields where observability reduces asymmetric information; no empirical insurance cost data or pilot programs reported.
Open interoperability standards and agent identities can lower entry barriers, increase competition, and accelerate complementary innovation.
Economic and policy reasoning referencing benefits of standards/open ecosystems; no empirical intervention or controlled comparison provided.
Design choices will shape capital intensity and replacement cycles; architectures that support upgradeability and modularity lower expected upgrade costs and stranded‑asset risk.
Economic reasoning and analogy to modular design benefits in other industries; conceptual argument without empirical capital‑allocation data or simulations.
Architectural components such as agentic identity and attestation, secure communication protocols, semantic layers and interchange formats, policy engines, and observability pipelines are necessary to enable safe, economic multi‑agent ecosystems.
Architectural blueprint proposed via conceptual systems design; justification by analogy to existing security/identity/semantic frameworks; no empirical testing reported.
Design principles — modularity, clear agentic identity, secure agent‑to‑agent communication, policy‑governed runtimes, semantic interoperability, and observability/governance frameworks — will mitigate the architectural risks identified.
Normative systems design proposition grounded in systems engineering reasoning and historical lessons; no experimental validation or deployment studies provided.
New capabilities (edge hardware, sensing, connectivity, and AI) now enable agents that not only sense/report but also perceive, reason, and act autonomously and cooperatively in real time.
Technological trend synthesis and systems reasoning; examples of mature edge hardware and advances in real‑time ML are used illustratively; no experimental validation provided.
Treating evolution, trust, and interoperability as first‑class requirements (rather than afterthoughts) is essential to avoid costly lock‑in, premature ossification, fragmentation, and negative externalities observed with IoT.
Normative prescription motivated by historical/comparative analysis of Internet and IoT (qualitative examples of fragmentation and lock‑in); no controlled study or quantitative validation presented.
The next phase of the Internet will be the "Internet of Physical AI Agents" — distributed, long-lived, embodied systems that perceive, reason, and act autonomously in real time.
Predictive/conceptual argument based on observed technological trends (advances in edge hardware, sensing, connectivity, and AI). Position paper with historical/comparative reasoning and illustrative examples; no primary empirical dataset or quantified projection.
Open-source orchestration and evaluation harnesses plus a self-contained evaluation pipeline improve reproducibility for the Speedrunning Track.
Paper claims and documents the release of orchestration and evaluation code and describes the self-contained pipeline designed for deterministic reproducible evaluation.
Version 1.0 marks integration into operational workflows and establishes a base for future capabilities.
Authors report that v1.0 has been used in verification and mask-refinement loops for real datasets (MeerKAT, ASKAP, APERTIF); no detailed deployment metrics provided.
Immersive inspection tools like iDaVIE are complements to automated ML pipelines by helping generate higher-quality labels and curated training examples.
Paper argues conceptual complementarity and cites iDaVIE's use for mask refinement and curated subcube export; no experimental comparison of label quality or downstream ML performance provided.