Evidence (5539 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Adoption
Remove filter
Harnessing AI's potential requires moving beyond measuring technical model performance (e.g., predictive accuracy) to measuring strategic impact.
Authors argue this as a conceptual requirement for realizing AI's benefits in R&D; presented as a recommendation rather than supported by quantified empirical evidence in the excerpt.
Preliminary analyses suggest that 'AI-native' companies may be outpacing traditional peers.
Explicitly stated in the paper as based on preliminary analyses; the excerpt provides no details on the analyses, metrics, or sample sizes.
The broad introduction of AI into the R&D landscape over the last years holds the promise to lift pharmaceutical R&D out of its productivity problem.
Framed as an expectation/promise in the paper; based on recent broad adoption trends of AI in R&D (no specific empirical evaluation or sample size reported in the excerpt).
In this verifiable domain, simple arbitrage strategies generate net profit margins of up to 40%.
Empirical result from the SWE-bench case study comparing arbitrage strategy returns using GPT-5 mini and DeepSeek v3.2 (reported maximum net profit margin = 40%).
Generative AI can autonomously produce novel content, including text, images, models, and scenarios.
General technical/descriptive claim stated in the paper's background/introduction; not an empirically tested claim within the provided excerpt.
Generative AI facilitates the synthesis of structured and unstructured information from diverse sources, enabling managers to explore multiple decision pathways, identify potential risks, and optimize strategic choices.
Descriptive/functional claim made in the paper's introduction and conceptual framing; the empirical component (survey + SEM) is described generally but no specific measures or effect sizes for information synthesis or these capabilities are provided in the excerpt.
Generative AI augments human creativity by producing innovative solutions and scenario-planning alternatives that may not emerge through conventional analytical approaches.
Stated in the conceptual/argumentative portion of the paper; may be supported by survey items but no explicit empirical measure or effect size for creativity is provided in the provided text.
Decision quality and strategic agility positively influence organizational performance.
Reported SEM results from the paper linking the constructs (decision quality and strategic agility) to organizational performance using survey data from senior managers and AI adoption specialists; method = SmartPLS.
Generative AI adoption significantly enhances strategic agility.
Same empirical source as above: survey of senior managers/decision-makers/AI adoption specialists; tested via Structural Equation Modeling (SmartPLS) as reported in the paper.
Generative AI adoption significantly enhances decision quality.
Empirical analysis reported in the paper: survey data collected from senior managers, decision-makers, and AI adoption specialists across multiple industries; relationships assessed using Structural Equation Modeling (SmartPLS). No numeric sample size or effect estimate reported in the provided text.
By enabling developers without initial capital to participate in the digital economy, RSI could unlock the 'latent jobs dividend' in low-income countries and help address local challenges in health, agriculture, and services.
Societal-impact argument in the paper linking the RSI model to potential employment gains and localized solutions; speculative extrapolation, no empirical employment estimates or pilot studies reported.
The RSI model could stimulate innovation in the ecosystem.
Argument based on lowered financial barriers and incentive structures from the paper's theoretical comparative analysis; no empirical measures of innovation provided.
The RSI model aligns stakeholder interests (platforms and developers).
Theoretical argument and incentive-alignment reasoning in the paper's comparative framework; no empirical validation presented.
A comparative analysis in the paper shows that the RSI model lowers entry barriers for developers.
Detailed comparative (theoretical) analysis within the paper contrasting existing models and RSI; no empirical trial, sample, or randomized test reported.
Generative AI platforms (Google AI Studio, OpenAI, Anthropic) provide infrastructures (APIs, models) that are transforming the application development ecosystem.
Statement in paper based on literature review and descriptive framing of current platforms; no empirical sample or quantitative test reported.
Financial digital intelligence enhances innovation by strengthening regional industry–university–research collaboration.
Authors report this channel from mechanism/mediation tests using the same empirical sample (5,731 observations, 2015–2022); specific measures of collaboration or identification strategy not provided in excerpt.
Financial digital intelligence enhances innovation by reducing transaction costs.
Mechanism analysis reported by authors on the same panel dataset (5,731 observations, 2015–2022); reduction in transaction costs is presented as a mediating channel (details of measurement/identification not included in excerpt).
Financial digital intelligence enhances innovation by improving corporate information disclosure.
Mechanism analysis reported in paper using same empirical sample (5,731 observations, 2015–2022); authors identify corporate information disclosure as a mediating channel (specific identification strategy not provided in excerpt).
Financial digital intelligence remarkably boosts the innovative development of strategic emerging industries.
Empirical analysis using panel data from 2015–2022 comprising 5,731 observations covering 789 listed companies and 114 prefecture-level cities in China (methods not specified in excerpt; presumably regression analysis on firm/city-level panel).
In production, the system received high satisfaction from both domain experts and developers, with all participants reporting full satisfaction with communication efficiency.
Post-deployment user feedback / satisfaction reports mentioned in paper (no numeric participant count provided).
The automated workflow saved an estimated 979 engineering hours.
Aggregate time-savings estimate reported in paper (derived from per-API time reduction × number of APIs).
The automated workflow reduces per-API development time from approximately 5 hours to under 7 minutes.
Time-per-API comparison reported in paper based on evaluation on spapi (comparison of manual vs automated per-API time).
The automated workflow achieves 93.7% F1 score.
Empirical evaluation on spapi (F1 reported); presumably computed over the evaluated API items/endpoints.
We address this gap through a graph-based workflow optimization approach that progressively replaces manual coordination with LLM-powered services, enabling incremental adoption without disrupting established practices.
Description of proposed method (graph-based workflow + LLM-powered services) and claim of design enabling incremental adoption; supported by subsequent case evaluation.
The work underscores the urgency of tangible actions aimed at closing the AI divide and allowing Africa to actively shape its AI future.
Concluding normative claim in the paper, supported by the paper's synthesis of identified infrastructural and policy barriers and the illustrative ACT tool.
We introduce the Africa AI Compute Tracker (ACT), an interactive map to monitor the availability of AI-ready HPC systems throughout the continent.
Paper reports development and introduction of the ACT tool; the claim is about the authors' own deliverable (an interactive map consolidating HPC availability data).
Sustainable AI adoption requires robust digital foundations through balanced access to compute, data, and the energy that makes it possible (the 'right enablers').
Normative claim grounded in the paper's stated quantitative and qualitative analysis and synthesis of official declarations; presented as a central conceptual conclusion.
Organizational size moderates the adoption–efficiency relationship such that larger firms realize proportionally greater efficiency gains from AI adoption.
Reported moderation effect in the PLS-PM analysis testing organizational size as a moderator of the relationship between AI adoption and recruitment efficiency metrics across sampled organizations.
Procedural fairness perceptions positively predict employee experience outcomes, including organizational commitment, job satisfaction, and employer trust.
PLS-PM paths from procedural fairness perceptions to employee experience measures (organizational commitment, job satisfaction, employer trust) using survey data from HR professionals' reports.
Algorithmic transparency is a strong predictor of procedural fairness perceptions.
PLS-PM results linking measured algorithmic transparency to procedural fairness perceptions in the survey data (n=523 respondents).
AI adoption is positively associated with improvements in quality-of-hire.
PLS-PM association between AI adoption and reported quality-of-hire improvement from HR respondents across sampled organizations.
AI adoption is positively associated with reductions in cost-per-hire.
PLS-PM association between AI adoption and cost-per-hire reduction reported in the survey (firm-level outcomes across sampled organizations).
AI adoption is positively associated with reductions in time-to-hire (recruitment time).
PLS-PM association between AI adoption and recruitment efficiency metrics reported in the survey (firm-level outcomes across sampled organizations).
Top management support and HR digital readiness are both positively associated with organizational AI adoption, with top management support demonstrating greater explanatory power.
PLS-PM tests of organizational antecedents predicting organizational AI adoption using survey responses aggregated to organization level (184 organizations referenced).
Perceived usefulness and perceived ease of use significantly predict AI adoption intention, with perceived usefulness exhibiting a stronger effect.
PLS-PM results on relationships between TAM constructs (perceived usefulness, perceived ease of use) and adoption intention using survey data (n=523).
A large portion of the interactive activities' AI market value (26%) involves transferring information.
Descriptive subcategory statistic: within interactive activities, authors report 26% of market value pertains to information transfer tasks.
Interactive activities (which include both information-based and physical activities) account for 48% of AI market value.
Descriptive aggregate: authors define an 'interactive' category spanning info and physical activities and report it holds 48% of AI market value.
A substantial portion of AI market value (36%) is used in activities that involve creating information.
Descriptive aggregate: subcategory within information-based activities—authors report 36% of market value allocated to 'creating information'.
Most of the AI market value is used in information-based activities (72%).
Descriptive aggregate: authors categorize activities into information-based vs physical and report that 72% of estimated AI market value maps to information-based activities.
There is a highly uneven distribution of AI market value across activities: the top 1.6% of activities account for over 60% of AI market value.
Descriptive statistical result from mapping estimated AI market values to the ~20K activities; authors report concentration metrics (top 1.6% share >60%).
We use the data about AI software and robotic systems to generate graphical displays of how the estimated units and market values of all worldwide AI systems used today are distributed across the work activities that these systems help perform.
Analytic/mapping procedure: authors combine classifications of software (13,275) and robots (20.8M) with market-value estimates to create visual distributions across activities.
We classify a worldwide tally of 20.8 million robotic systems using the developed work-activity ontology.
Empirical classification/counting: authors report mapping 20.8 million robotic systems worldwide to the activity ontology.
We classify descriptions of 13,275 AI software applications using the developed work-activity ontology.
Empirical classification: authors state they mapped 13,275 AI software application descriptions to the ontology.
We disaggregate and then substantially reorganize the approximately 20K activities in the US Department of Labor's O*NET occupational database to produce a comprehensive ontology of work activities.
Methodological: authors report transforming the O*NET activity taxonomy (~20,000 activity-level records) by disaggregation and reorganization into a new ontology.
Models trained in EnterpriseLab remain robust across diverse enterprise benchmarks, including EnterpriseBench (+10%) and CRMArena (+10%).
Benchmark evaluations reported in the paper showing reported +10% improvements on EnterpriseBench and CRMArena relative to baseline; exact baselines, statistical tests, and sample sizes are not specified in the abstract.
8B-parameter models trained in EnterpriseLab reduce inference costs by 8-10x compared to frontier models (implied GPT-4o).
Empirical cost comparison reported in the paper; the abstract states an 8-10x reduction in inference costs for the 8B models trained in EnterpriseLab versus the referenced frontier model(s). Detailed cost accounting and sample sizes not provided in the abstract.
8B-parameter models trained within EnterpriseLab match GPT-4o's performance on complex enterprise workflows.
Empirical evaluation reported in the paper comparing 8B-parameter models trained in EnterpriseLab to GPT-4o on complex enterprise workflows; specific benchmark tests and metrics are referenced but details (sample sizes, exact metrics) are not provided in the abstract.
We validate the platform through EnterpriseArena, an instantiation with 15 applications and 140+ tools across IT, HR, sales, and engineering domains.
Reported instantiation/experimental setup in the paper: EnterpriseArena contains 15 applications and 140+ tools spanning specified domains.
EnterpriseLab provides integrated training pipelines with continuous evaluation.
System/design claim in paper describing integrated training and evaluation tooling as part of the platform.
EnterpriseLab includes automated trajectory synthesis that programmatically generates training data from environment schemas.
System/design claim described in paper; supported by the authors' description of an automated data-generation component.