Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
Generative AI can autonomously produce novel content, including text, images, models, and scenarios.
General technical/descriptive claim stated in the paper's background/introduction; not an empirically tested claim within the provided excerpt.
Generative AI facilitates the synthesis of structured and unstructured information from diverse sources, enabling managers to explore multiple decision pathways, identify potential risks, and optimize strategic choices.
Descriptive/functional claim made in the paper's introduction and conceptual framing; the empirical component (survey + SEM) is described generally but no specific measures or effect sizes for information synthesis or these capabilities are provided in the excerpt.
Generative AI augments human creativity by producing innovative solutions and scenario-planning alternatives that may not emerge through conventional analytical approaches.
Stated in the conceptual/argumentative portion of the paper; may be supported by survey items but no explicit empirical measure or effect size for creativity is provided in the provided text.
Decision quality and strategic agility positively influence organizational performance.
Reported SEM results from the paper linking the constructs (decision quality and strategic agility) to organizational performance using survey data from senior managers and AI adoption specialists; method = SmartPLS.
Generative AI adoption significantly enhances strategic agility.
Same empirical source as above: survey of senior managers/decision-makers/AI adoption specialists; tested via Structural Equation Modeling (SmartPLS) as reported in the paper.
Generative AI adoption significantly enhances decision quality.
Empirical analysis reported in the paper: survey data collected from senior managers, decision-makers, and AI adoption specialists across multiple industries; relationships assessed using Structural Equation Modeling (SmartPLS). No numeric sample size or effect estimate reported in the provided text.
Human-like presentations increased perceived usefulness and agency in certain tasks.
Experimental manipulation of the human-likeness of AI presentation in the study's three tasks; the abstract reports increased perceived usefulness and agency for human-like presentations in some tasks. No sample sizes, task specifics, or effect magnitudes reported in abstract.
A single dissent within a panel reduced pressure to conform.
Experimental manipulation of within-panel consensus (introducing a single dissent) in the study's three tasks; abstract reports that a single dissent lowered conformity pressure. No numerical data provided in abstract.
Accuracy improved for small panels relative to a single AI.
Reported experimental result from the paper's study: participants completed three tasks and received advice from AI panels; panel size was manipulated (small panels vs single AI). The abstract states this accuracy improvement for small panels. (Sample size and exact tasks not reported in abstract.)
By enabling developers without initial capital to participate in the digital economy, RSI could unlock the 'latent jobs dividend' in low-income countries and help address local challenges in health, agriculture, and services.
Societal-impact argument in the paper linking the RSI model to potential employment gains and localized solutions; speculative extrapolation, no empirical employment estimates or pilot studies reported.
The RSI model could stimulate innovation in the ecosystem.
Argument based on lowered financial barriers and incentive structures from the paper's theoretical comparative analysis; no empirical measures of innovation provided.
The RSI model aligns stakeholder interests (platforms and developers).
Theoretical argument and incentive-alignment reasoning in the paper's comparative framework; no empirical validation presented.
A comparative analysis in the paper shows that the RSI model lowers entry barriers for developers.
Detailed comparative (theoretical) analysis within the paper contrasting existing models and RSI; no empirical trial, sample, or randomized test reported.
Generative AI platforms (Google AI Studio, OpenAI, Anthropic) provide infrastructures (APIs, models) that are transforming the application development ecosystem.
Statement in paper based on literature review and descriptive framing of current platforms; no empirical sample or quantitative test reported.
In production, the system received high satisfaction from both domain experts and developers, with all participants reporting full satisfaction with communication efficiency.
Post-deployment user feedback / satisfaction reports mentioned in paper (no numeric participant count provided).
The automated workflow saved an estimated 979 engineering hours.
Aggregate time-savings estimate reported in paper (derived from per-API time reduction × number of APIs).
The automated workflow reduces per-API development time from approximately 5 hours to under 7 minutes.
Time-per-API comparison reported in paper based on evaluation on spapi (comparison of manual vs automated per-API time).
The automated workflow achieves 93.7% F1 score.
Empirical evaluation on spapi (F1 reported); presumably computed over the evaluated API items/endpoints.
We address this gap through a graph-based workflow optimization approach that progressively replaces manual coordination with LLM-powered services, enabling incremental adoption without disrupting established practices.
Description of proposed method (graph-based workflow + LLM-powered services) and claim of design enabling incremental adoption; supported by subsequent case evaluation.
A large portion of the interactive activities' AI market value (26%) involves transferring information.
Descriptive subcategory statistic: within interactive activities, authors report 26% of market value pertains to information transfer tasks.
Interactive activities (which include both information-based and physical activities) account for 48% of AI market value.
Descriptive aggregate: authors define an 'interactive' category spanning info and physical activities and report it holds 48% of AI market value.
A substantial portion of AI market value (36%) is used in activities that involve creating information.
Descriptive aggregate: subcategory within information-based activities—authors report 36% of market value allocated to 'creating information'.
Most of the AI market value is used in information-based activities (72%).
Descriptive aggregate: authors categorize activities into information-based vs physical and report that 72% of estimated AI market value maps to information-based activities.
There is a highly uneven distribution of AI market value across activities: the top 1.6% of activities account for over 60% of AI market value.
Descriptive statistical result from mapping estimated AI market values to the ~20K activities; authors report concentration metrics (top 1.6% share >60%).
We use the data about AI software and robotic systems to generate graphical displays of how the estimated units and market values of all worldwide AI systems used today are distributed across the work activities that these systems help perform.
Analytic/mapping procedure: authors combine classifications of software (13,275) and robots (20.8M) with market-value estimates to create visual distributions across activities.
We classify a worldwide tally of 20.8 million robotic systems using the developed work-activity ontology.
Empirical classification/counting: authors report mapping 20.8 million robotic systems worldwide to the activity ontology.
We classify descriptions of 13,275 AI software applications using the developed work-activity ontology.
Empirical classification: authors state they mapped 13,275 AI software application descriptions to the ontology.
We disaggregate and then substantially reorganize the approximately 20K activities in the US Department of Labor's O*NET occupational database to produce a comprehensive ontology of work activities.
Methodological: authors report transforming the O*NET activity taxonomy (~20,000 activity-level records) by disaggregation and reorganization into a new ontology.
Models trained in EnterpriseLab remain robust across diverse enterprise benchmarks, including EnterpriseBench (+10%) and CRMArena (+10%).
Benchmark evaluations reported in the paper showing reported +10% improvements on EnterpriseBench and CRMArena relative to baseline; exact baselines, statistical tests, and sample sizes are not specified in the abstract.
8B-parameter models trained in EnterpriseLab reduce inference costs by 8-10x compared to frontier models (implied GPT-4o).
Empirical cost comparison reported in the paper; the abstract states an 8-10x reduction in inference costs for the 8B models trained in EnterpriseLab versus the referenced frontier model(s). Detailed cost accounting and sample sizes not provided in the abstract.
8B-parameter models trained within EnterpriseLab match GPT-4o's performance on complex enterprise workflows.
Empirical evaluation reported in the paper comparing 8B-parameter models trained in EnterpriseLab to GPT-4o on complex enterprise workflows; specific benchmark tests and metrics are referenced but details (sample sizes, exact metrics) are not provided in the abstract.
We validate the platform through EnterpriseArena, an instantiation with 15 applications and 140+ tools across IT, HR, sales, and engineering domains.
Reported instantiation/experimental setup in the paper: EnterpriseArena contains 15 applications and 140+ tools spanning specified domains.
EnterpriseLab provides integrated training pipelines with continuous evaluation.
System/design claim in paper describing integrated training and evaluation tooling as part of the platform.
EnterpriseLab includes automated trajectory synthesis that programmatically generates training data from environment schemas.
System/design claim described in paper; supported by the authors' description of an automated data-generation component.
EnterpriseLab provides a modular environment exposing enterprise applications via a Model Context Protocol, enabling seamless integration of proprietary and open-source tools.
Feature/design claim in paper; supported by implementation details of the 'Model Context Protocol' and reported integration capabilities in the platform description.
We introduce EnterpriseLab, a full-stack platform that unifies tool integration, data generation, and training into a closed-loop framework.
System/design claim describing the contribution of the paper (platform implementation and architecture); supported by the paper's implementation description rather than independent validation.
Organizations can design more effective recruitment strategies by signaling AI adoption to increase attractiveness to prospective applicants.
Practical implication drawn from the combined experimental findings (Study 1 N = 145; Study 2 N = 240; total N = 385) showing AI-adoption signals increase organizational attractiveness via perceived innovation ability, particularly for applicants with high AI self-efficacy.
Conceptualizing AI adoption as an organizational signal extends signaling theory to the context of technology-infused recruitment.
Theoretical argumentation in the paper, supported by the two experimental studies (Study 1 and Study 2) that test signaling mechanisms in recruitment contexts.
The positive indirect effect of AI-adoption signals on organizational attractiveness via perceived innovation ability is stronger for job seekers with high AI self-efficacy (Study 2 moderated mediation).
Study 2: moderated mediation model showing AI self-efficacy moderates the mediated relationship; sample size N = 240; participants were active job seekers.
Perceived innovation ability mediates the positive association between AI-adoption signals and organizational attractiveness (Study 2).
Study 2: moderated mediation analysis in an experiment recruiting active job seekers; sample size N = 240; mediation of AI-signal -> perceived innovation ability -> organizational attractiveness was validated.
AI-adoption signals are significantly positively associated with organizational attractiveness (Study 1).
Study 1: scenario-based experiment comparing AI-adoption signal vs no-signal conditions; sample size N = 145.
The effect is amplified in Japanese, where experiential queries draw 62.1% non-OTA citations compared to 50.0% in English.
Subset analysis by language within the audited sample comparing non-OTA citation shares for experiential queries in Japanese vs English; percentages reported in paper.
Experiential queries draw 55.9% of their citations from non-OTA sources, compared to 30.8% for transactional queries — a 25.1 percentage-point gap (p < 5 × 10^{-20}).
Quantitative comparison of citation-source types in the audited sample (1,357 citations across 156 queries), classifying queries as 'experiential' vs 'transactional' and computing share of citations from non-OTA sources; reported p-value indicates statistical test of difference.
An approach is needed focused on emerging and future interdependencies between professionals and generative machine learning, implying extending but also reimagining theoretical perspectives on expertise, work and organizations.
Paper's central argument based on theoretical reasoning and literature synthesis about generative ML characteristics and their implications for professionals; method: conceptual/theoretical development; no empirical sample.
Existing theories need to be extended whilst also responding to the distinctive characteristics of generative machine learning and the implications for how we theorize change.
Argumentative/theoretical claim in the paper based on comparison of features of generative ML with prior digital/algorithmic technologies; method: conceptual analysis and literature engagement; no empirical sample.
We develop an approach using insights from existing literature on digital, algorithmic and artificial intelligence technologies.
Paper's stated contribution: theoretical development based on synthesis of existing literature (digital, algorithmic, AI). Method: conceptual synthesis; no empirical testing or sample reported.
There is a need for an approach to theorizing professional work and professional service firms in the generative machine learning age.
Conceptual argument presented in the paper (literature-based rationale); method is theoretical/literature review and argumentation; no empirical sample reported.
GenAI implementations that are strategically deployed in managed Azure cloud infrastructure provide a positive ROI over time when aligned with business processes, enterprise architecture, and performance metrics.
Conclusion drawn from the paper's mixed-method analysis (quantitative ROI modelling, cost–benefit analysis, and case study synthesis).
Close coupling among Azure OpenAI Service, Azure Machine Learning, and cost governance tooling (FinOps) significantly decreases overall cost of ownership and enhances scalability and compliance.
Architectural analysis of Azure-native GenAI services and cost/governance tooling reported in the paper.
Measurable ROI from GenAI on Azure is mainly driven by improvements in productivity, optimization of operational costs, faster decision making, and increased speed of innovation across business functions.
Reported results from the paper's mixed-method study combining quantitative ROI modelling and cost–benefit analysis plus qualitative synthesis of secondary enterprise case studies.