Evidence (4793 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Productivity
Remove filter
In production, the system received high satisfaction from both domain experts and developers, with all participants reporting full satisfaction with communication efficiency.
Post-deployment user feedback / satisfaction reports mentioned in paper (no numeric participant count provided).
The automated workflow saved an estimated 979 engineering hours.
Aggregate time-savings estimate reported in paper (derived from per-API time reduction × number of APIs).
The automated workflow reduces per-API development time from approximately 5 hours to under 7 minutes.
Time-per-API comparison reported in paper based on evaluation on spapi (comparison of manual vs automated per-API time).
The automated workflow achieves 93.7% F1 score.
Empirical evaluation on spapi (F1 reported); presumably computed over the evaluated API items/endpoints.
We address this gap through a graph-based workflow optimization approach that progressively replaces manual coordination with LLM-powered services, enabling incremental adoption without disrupting established practices.
Description of proposed method (graph-based workflow + LLM-powered services) and claim of design enabling incremental adoption; supported by subsequent case evaluation.
Organizational size moderates the adoption–efficiency relationship such that larger firms realize proportionally greater efficiency gains from AI adoption.
Reported moderation effect in the PLS-PM analysis testing organizational size as a moderator of the relationship between AI adoption and recruitment efficiency metrics across sampled organizations.
Procedural fairness perceptions positively predict employee experience outcomes, including organizational commitment, job satisfaction, and employer trust.
PLS-PM paths from procedural fairness perceptions to employee experience measures (organizational commitment, job satisfaction, employer trust) using survey data from HR professionals' reports.
Algorithmic transparency is a strong predictor of procedural fairness perceptions.
PLS-PM results linking measured algorithmic transparency to procedural fairness perceptions in the survey data (n=523 respondents).
AI adoption is positively associated with improvements in quality-of-hire.
PLS-PM association between AI adoption and reported quality-of-hire improvement from HR respondents across sampled organizations.
AI adoption is positively associated with reductions in cost-per-hire.
PLS-PM association between AI adoption and cost-per-hire reduction reported in the survey (firm-level outcomes across sampled organizations).
AI adoption is positively associated with reductions in time-to-hire (recruitment time).
PLS-PM association between AI adoption and recruitment efficiency metrics reported in the survey (firm-level outcomes across sampled organizations).
Top management support and HR digital readiness are both positively associated with organizational AI adoption, with top management support demonstrating greater explanatory power.
PLS-PM tests of organizational antecedents predicting organizational AI adoption using survey responses aggregated to organization level (184 organizations referenced).
Perceived usefulness and perceived ease of use significantly predict AI adoption intention, with perceived usefulness exhibiting a stronger effect.
PLS-PM results on relationships between TAM constructs (perceived usefulness, perceived ease of use) and adoption intention using survey data (n=523).
A large portion of the interactive activities' AI market value (26%) involves transferring information.
Descriptive subcategory statistic: within interactive activities, authors report 26% of market value pertains to information transfer tasks.
Interactive activities (which include both information-based and physical activities) account for 48% of AI market value.
Descriptive aggregate: authors define an 'interactive' category spanning info and physical activities and report it holds 48% of AI market value.
A substantial portion of AI market value (36%) is used in activities that involve creating information.
Descriptive aggregate: subcategory within information-based activities—authors report 36% of market value allocated to 'creating information'.
Most of the AI market value is used in information-based activities (72%).
Descriptive aggregate: authors categorize activities into information-based vs physical and report that 72% of estimated AI market value maps to information-based activities.
There is a highly uneven distribution of AI market value across activities: the top 1.6% of activities account for over 60% of AI market value.
Descriptive statistical result from mapping estimated AI market values to the ~20K activities; authors report concentration metrics (top 1.6% share >60%).
We use the data about AI software and robotic systems to generate graphical displays of how the estimated units and market values of all worldwide AI systems used today are distributed across the work activities that these systems help perform.
Analytic/mapping procedure: authors combine classifications of software (13,275) and robots (20.8M) with market-value estimates to create visual distributions across activities.
We classify a worldwide tally of 20.8 million robotic systems using the developed work-activity ontology.
Empirical classification/counting: authors report mapping 20.8 million robotic systems worldwide to the activity ontology.
We classify descriptions of 13,275 AI software applications using the developed work-activity ontology.
Empirical classification: authors state they mapped 13,275 AI software application descriptions to the ontology.
We disaggregate and then substantially reorganize the approximately 20K activities in the US Department of Labor's O*NET occupational database to produce a comprehensive ontology of work activities.
Methodological: authors report transforming the O*NET activity taxonomy (~20,000 activity-level records) by disaggregation and reorganization into a new ontology.
Models trained in EnterpriseLab remain robust across diverse enterprise benchmarks, including EnterpriseBench (+10%) and CRMArena (+10%).
Benchmark evaluations reported in the paper showing reported +10% improvements on EnterpriseBench and CRMArena relative to baseline; exact baselines, statistical tests, and sample sizes are not specified in the abstract.
8B-parameter models trained in EnterpriseLab reduce inference costs by 8-10x compared to frontier models (implied GPT-4o).
Empirical cost comparison reported in the paper; the abstract states an 8-10x reduction in inference costs for the 8B models trained in EnterpriseLab versus the referenced frontier model(s). Detailed cost accounting and sample sizes not provided in the abstract.
8B-parameter models trained within EnterpriseLab match GPT-4o's performance on complex enterprise workflows.
Empirical evaluation reported in the paper comparing 8B-parameter models trained in EnterpriseLab to GPT-4o on complex enterprise workflows; specific benchmark tests and metrics are referenced but details (sample sizes, exact metrics) are not provided in the abstract.
We validate the platform through EnterpriseArena, an instantiation with 15 applications and 140+ tools across IT, HR, sales, and engineering domains.
Reported instantiation/experimental setup in the paper: EnterpriseArena contains 15 applications and 140+ tools spanning specified domains.
EnterpriseLab provides integrated training pipelines with continuous evaluation.
System/design claim in paper describing integrated training and evaluation tooling as part of the platform.
EnterpriseLab includes automated trajectory synthesis that programmatically generates training data from environment schemas.
System/design claim described in paper; supported by the authors' description of an automated data-generation component.
EnterpriseLab provides a modular environment exposing enterprise applications via a Model Context Protocol, enabling seamless integration of proprietary and open-source tools.
Feature/design claim in paper; supported by implementation details of the 'Model Context Protocol' and reported integration capabilities in the platform description.
We introduce EnterpriseLab, a full-stack platform that unifies tool integration, data generation, and training into a closed-loop framework.
System/design claim describing the contribution of the paper (platform implementation and architecture); supported by the paper's implementation description rather than independent validation.
The paper reports details from a 100% deployment of DRL with policy regularizations on Alibaba's e-commerce platform, Tmall.
Direct statement in the abstract claiming full deployment across Tmall; implies a real-world, company-scale deployment but the abstract provides no operational metrics or counts.
Imposing policy regularizations improves the final performance of several DRL methods for inventory management.
Empirical claim supported by the paper's synthetic experiments and reported production deployment on Alibaba/Tmall (as stated in the abstract); no quantitative effect sizes provided in the abstract.
Imposing policy regularizations, grounded in classical inventory concepts such as 'Base Stock', can significantly accelerate hyperparameter tuning for DRL methods.
Paper reports synthetic experiments and a production deployment (Alibaba/Tmall) where policy regularizations were applied; abstract claims acceleration in hyperparameter tuning but does not report numeric tuning-time metrics in the abstract.
Deep Reinforcement Learning (DRL) provides a general-purpose methodology for training inventory policies that can leverage big data and compute.
Argument/assertion made in the paper's introduction/abstract (conceptual claim about DRL capabilities); no empirical sample or quantitative test reported in the abstract.
Human-replacing technologies have a strategic role in enhancing industrial productivity and ensuring the long-term resilience of Ukraine’s mining and metallurgical sector amid workforce shortages and structural labour-market changes due to war and demographic decline.
Integrated sectoral assessment in the paper combining current context (workforce shortages, structural changes), literature on technology-driven productivity/resilience, and industry-specific considerations; presented as a high-level conclusion.
Integrating ergonomic assessments and human–systems–interaction approaches into automation projects is important to prevent cognitive overload, occupational stress and operational risks for control‑room operators.
Recommendation and emphasis in the paper, supported by references to ergonomics and human-factors literature; presented as a preventive/mitigative approach rather than a quantified empirical result for the sector.
Successful technological modernization requires continuous investment in human capital, reskilling and the development of digital and engineering competencies.
Policy/recommendation based on the paper's synthesis of the sector analysis and literature on skill requirements and technology adoption; not presented as an original empirical estimate in the summary.
Higher robot density is associated with productivity gains, particularly in low-robotized sectors such as Ukraine’s mining and metallurgical industry.
Empirical evidence cited from international and industry-specific studies reviewed in the paper (literature review/meta-analytic style evidence); no Ukraine-specific causal estimate with sample size reported in the summary.
Human-replacing technologies also have an indirect impact on productivity by increasing total factor productivity (TFP).
Analytical argumentation in the paper supported by references to empirical studies showing TFP effects of automation/digitalization; literature synthesis rather than a new econometric estimate presented for Ukraine.
Human-replacing technologies (mechanization, automation, robotization, digitalization and AI-augmentation) make a direct contribution to labour productivity growth in Ukraine's mining and metallurgical sector.
Sectoral analysis and synthesis in the paper drawing on empirical international and industry-specific studies; literature review of productivity impacts of mechanization/automation/robotization/digitalization/AI in industrial contexts.
Industrial intelligence and the digital economy can be leveraged as a 'dual engine' to boost regional TFCP and advance high-quality green and low-carbon economic development, supporting differentiated regional coordination policies.
Synthesis/implication drawn from the paper's empirical findings (SDM results on 30 provinces, 2010–2023) showing positive total/spillover effects and regional heterogeneity.
Green finance has an insignificant positive effect on regional TFCP.
Coefficient on green finance control variable in the Spatial Durbin Model (30 provinces, 2010–2023) is positive but not statistically significant.
The digital economy presents different regional driving patterns: a 'local-spillover dual drive' in the east, a 'local-dominated drive' in the central region, and a 'spillover-dominated drive' in the west.
Regional/subsample Spatial Durbin Model estimates for digital economy variables across east, central, and west subsamples (30 provinces, 2010–2023) with reported direct and indirect effects.
The digital economy exerts a significantly positive direct effect on local TFCP and a strong positive spatial spillover effect, forming a 'local driving + spatial radiation' promotion pattern.
Spatial Durbin Model estimates on panel data (30 provinces, 2010–2023) showing statistically significant positive direct and indirect (spillover) coefficients for digital economy variables.
Regional TFCP shows significant positive spatial autocorrelation.
Spatial analysis (Spatial Durbin Model and spatial statistics) applied to panel of 30 provincial-level regions; reported significant spatial autocorrelation (e.g., positive Moran's I implied).
Across 378 hardware validated experiments, concise human-expert skills with structured expert knowledge enable near-perfect success rates across platforms.
Reported experimental results: 378 hardware-validated experiments across platforms comparing agent configurations; finding reported that human-expert skills produce near-perfect success rates (no numeric success rate provided in excerpt).
Large language models (LLMs) and agentic systems have shown promise for automated software development.
Statement in paper referencing prior successes of LLMs and agentic systems for automated software development (no empirical data reported in this excerpt).
The largest gains appear when AI is embedded in an orchestrated workflow rather than deployed as an isolated coding assistant.
Central thesis supported by comparisons across five delivery configurations (traditional baseline and V1–V4) in a retrospective longitudinal field study of the Chiron platform applied to three real software modernization programs; authors observe greater portfolio-level improvements when AI is integrated into coordinated workflows.
V3 and V4 add acceptance-criteria validation, repository-native review, and hybrid human-agent execution, simultaneously improving speed, coverage, and issue load.
Observed differences across the five delivery configurations (baseline, V1–V4) in the field study of three modernization programs; authors link feature additions in V3/V4 to measured improvements in stage durations, coverage, and validation issues.
First-release coverage rises from 77.0% to 90.5% across the portfolio as platform versions progress.
Observed first-release coverage measured in the retrospective longitudinal field study of three real modernization programs, reported as percentages across delivery configurations.