Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
Projection congruence — alignment of forecasts/plans across heterogeneous agents — becomes a central metric for assessing alignment in agentic human–AI teams.
Conceptual modeling and proposal in the paper; introduced as a new measurable construct (projection congruence indices) for future empirical work.
The DAR framework reframes human oversight as a dynamic, auditable process whose micro-level mechanics and macro-level legitimacy have direct economic consequences for productivity, contracting, regulation, and welfare.
Synthesis claim based on the conceptual framework, formal modeling, derived propositions, and policy/economics implications sections. The claim is theoretical and synthesizing rather than empirically validated.
The Reversal Register will create granular, time-stamped administrative data valuable for structural estimation of trust, error externalities, and productivity comparisons between automation and human judgment.
Design claim linking register contents to potential econometric uses; no empirical data shown—claim about potential data utility.
Reversal Register logs can enable descriptive and causal analyses of handovers and support experimental/quasi-experimental tests (e.g., randomized hysteresis thresholds, A/B override policies).
Implied empirical strategies and instrumentation described; paper outlines how register data would be used for experiments and causal inference. No empirical implementation or sample reported.
Operationalizing reversible AI leadership via DAR can preserve human accountability while enabling AI-led decisions where appropriate.
Conceptual argument supported by the combined use of authority states, Reversal Register logging, and override mechanisms; no field validation provided.
DAR incorporates stabilizing mechanisms—hysteresis bands and safe-exit timers—to reduce rapid oscillation of authority and improve stability of handovers.
Formal model components and design proposals (hysteresis and timers) with conceptual argument that these damp oscillation; no empirical validation reported.
Continuous human-in-the-loop oversight, monitoring, and retraining are required to maintain quality and prevent model drift.
Practitioner reports and conceptual literature synthesized in the review advocating monitoring and retraining; no longitudinal empirical study provided here.
Transparent disclosure to customers about AI involvement helps preserve trust.
Conceptual analyses and referenced empirical/regulatory discussions in the literature aggregated by the review; this paper presents no new experimental evidence on disclosure effects.
Hybrid designs that automate low-risk, high-volume tasks while routing complex, judgment-sensitive cases to humans produce the best operational outcomes.
Inferred best-practice from aggregated empirical studies, industry examples, and conceptual reasoning; no controlled comparative trials presented in this review.
Agent augmentation via suggested responses, summarization, and information retrieval improves agent productivity.
Aggregated evidence from prior empirical research and practitioner reports cited in the review; no new measurements or sample sizes presented here.
Generative AI enables personalization at scale through automated tailoring of messaging and recommendations.
Qualitative synthesis of empirical studies and industry reports showing automated personalization use-cases; no systematic effect-size estimates or new quantitative data in this review.
Generative AI provides 24/7 availability and cost-effective scaling of routine interactions.
Industry case examples and prior empirical studies aggregated in the review; no original data or quantified sample sizes provided in this paper.
Generative AI can materially transform customer service and strategic communication by enabling continuous automation, scalable hyper-personalization, and effective agent augmentation.
Nano review: qualitative aggregation and synthesis of existing empirical studies, industry case examples, and conceptual analyses. No novel primary data or sample size; conclusion drawn from heterogeneous secondary sources and practitioner reports (not a systematic meta-analysis).
There is a need for standards around evaluation, bias mitigation, provenance, and accountability in AI-assisted ideation and design.
Policy recommendation motivated by documented biases, errors, and provenance issues in the reviewed studies; grounded in the synthesis's critique of existing practice.
There will likely be complementarity-driven increases in demand for evaluative, integrative, and domain-expert roles (curators, synthesizers, implementation experts).
Inference from task-level studies and economic reasoning about complementarities between AI generative capability and human evaluative skills; empirical labor-market evidence is limited in the reviewed literature.
Lower search and idea-generation costs enabled by LLMs may speed early-stage R&D and increase the gross flow of candidate innovations.
Theoretical economic interpretation supported by empirical findings of increased idea volumes in experimental/field studies summarized in the review; no long-run causal firm-level evidence presented.
Generative AI accelerates early-stage hypothesis and prototype development by providing scaffolded prompts and procedural suggestions.
Applied case evidence and experimental studies summarized in the review showing reduced time or increased productivity in early-stage experimental/design tasks when using LLM assistance; no pooled effect size presented.
Empirical studies document that AI-assisted tools can help break cognitive fixation and generate cross-domain analogies.
Cited experimental tasks and lab studies in the literature showing higher incidence of analogical or cross-domain suggestions from LLMs and improvements on fixation-related task metrics; heterogeneity across tasks and measures.
Generative AI provides scaffolded, structured support that aids systematic hypothesis formation, prototyping steps, and decomposition of complex problems.
Review of design/ideation studies and applied case evidence where LLMs produced stepwise plans, decomposition prompts, or hypothesis scaffolds; evidence drawn from multiple short-term experimental and applied studies, sample sizes and exact designs vary by study.
Generative models rapidly produce many candidate ideas, analogies, and associative prompts that help overcome cognitive fixation.
Synthesis of experimental ideation and design studies reporting increases in number of ideas and examples of reduced fixation when participants used LLM outputs; heterogeneous sample sizes across cited studies (not reported in review).
Generative AI can raise per-worker productivity for tasks involving brainstorming, drafting, and prototyping, but realized gains depend on downstream filtering and implementation costs.
User studies showing higher output on specific tasks (brainstorming/drafting), combined with qualitative reports of filtering/implementation effort; many studies measure immediate task output but not net realized productivity after implementation.
Generative AI can increase creative output in both lab and field tasks as judged by external raters.
Controlled experiments and field studies reporting higher judged creativity/novelty scores for AI-assisted outputs versus controls; judged creativity/novelty is typically assessed by human raters using rubric-based scoring.
AI assistance helps people overcome fixation and produces cross-domain analogies that they might not generate alone.
Experimental studies and qualitative analyses documenting reductions in fixation effects and increases in cross-domain analogical suggestions when participants use generative models.
Generative AI supports systematic problem breakdown and early-stage prototyping, accelerating hypothesis generation and prototype development.
Field case studies of AI-supported prototyping and lab/user studies reporting reduced time-to-prototype and generated hypotheses; measures include time-to-prototype and user-reported usefulness.
Generative AI boosts ideational fluency—the quantity and diversity of ideas produced in brainstorming tasks.
Controlled experiments and user studies measuring number and diversity of ideas with and without AI assistance; typical study designs compare participant idea counts/uniqueness across conditions (note: many studies use small or convenience samples).
When used as a 'cognitive co-pilot' that expands the solution space and challenges assumptions while humans curate and evaluate, generative AI generates economic value.
Inferred from experimental and field findings showing increased idea quantity/diversity and faster prototyping combined with qualitative studies showing human curation is needed; economic interpretation drawn from the review rather than direct macroeconomic measurement.
Generative AI serves a dual cognitive role: (1) a high-volume catalyst for divergent idea generation and cross-domain analogy-making, and (2) a structured assistant for deconstructing complex problems and scaffolding hypotheses and prototypes.
Synthesis of controlled experiments, lab studies, field case studies, and qualitative analyses summarized in the review; evidence includes measures of idea fluency/diversity, examples of analogy production, and observations of AI-assisted problem decomposition in prototyping tasks. (Note: underlying studies are heterogeneous and often short-term or convenience samples.)
Perceptions—specifically trust and perceived accuracy—are central frictions in AI adoption within finance; interventions that raise perceived and demonstrable accuracy (e.g., explainability, transparent validation) will increase uptake and productivity gains.
Study finds correlations between perceptions and adoption/productivity proxies from questionnaire and performance data; authors combine these empirical associations with qualitative insights to recommend explainability/validation as interventions. Evidence is correlational and inferential (causal impact of interventions not estimated in summary).
Higher perceived accuracy of AI outputs is associated with increased perceived utility of AI for forecasting and risk-management tasks.
Survey items measuring perceived accuracy and perceived utility for specific tasks (forecasting, risk management) and quantitative association analysis; supported by interview excerpts illustrating task-specific utility; exact effect sizes and sample counts not provided in summary.
Greater trust in AI correlates with greater willingness to adopt AI tools and to incorporate AI recommendations into decisions.
Correlational findings from structured questionnaires linking measures of trust with adoption intentions and self-reported incorporation of AI recommendations; supported by qualitative interview evidence; sample across multinational financial institutions (size not specified).
When trust and accuracy are high, human–AI collaboration improves organizational agility, enabling faster, data-driven strategic pivots and better risk management.
Quantitative analysis estimating relationships between perceived trust/accuracy and organizational agility indicators (speed of strategic pivots, risk-management metrics) augmented by interview accounts describing faster responses; sample: finance professionals across multinational financial institutions (sample size and exact agility metrics not specified).
Perceived accuracy of AI-generated insights increases decision confidence and perceived utility for forecasting and risk management.
Quantitative questionnaire measures of perceived accuracy correlated with self-reported decision confidence and perceived utility for forecasting/risk management, with qualitative interviews used to explain mechanisms; sample: finance professionals across multinational financial institutions (sample size not specified).
Perceived trust in AI tools is a key driver of finance professionals' willingness to use AI and their confidence in AI-assisted decisions.
Mixed-methods: quantitative analysis of structured questionnaires measuring perceived trust together with measures of willingness to use AI and decision confidence, supplemented by semi-structured interview evidence; sample described as finance professionals across multinational financial institutions (sample size not specified in summary).
Policy measures are needed to support reskilling, algorithmic accountability, data governance standards, and protections against discriminatory automated decisions to ensure equitable benefits from data-driven HRM adoption.
Policy implications section of the review synthesizing concerns and recommendations from the included literature.
Richer firm-level HR data resulting from data-driven HRM enables economists to better identify causal effects of workforce policies and technology adoption.
Methodological implication stated in the review: improved measurement and data availability noted across included studies as aiding empirical identification.
Data-driven HRM can raise firm productivity by reducing turnover costs, improving matching quality, and enabling targeted training, potentially increasing firm-level returns to AI adoption.
Reported benefits and theoretical mechanisms summarized from the reviewed literature; however the review also notes gaps in causal long-run evidence.
Adoption of data-driven HRM is likely to increase demand for data-literate HR professionals, data scientists, and AI tool vendors while requiring complementary upskilling for managers and employees.
Implication drawn in the review based on patterns in the literature; synthesis infers labor demand shifts from technologies and required capabilities reported in included studies.
Documented benefits of data-driven HRM include better anticipation of disruptions, optimized hiring and internal mobility, targeted well-being interventions, and improved HR operational efficiency.
Synthesis across included studies reporting empirical or observational benefits; collated as 'benefits documented' in the review (47-study sample).
Machine learning and AI support recruitment, performance evaluation, and personalized employee development.
Theme from the review: multiple peer-reviewed studies (within the 47) describe ML/AI applications in recruitment, performance evaluation, and personalization (thematic synthesis).
Information systems such as dashboards and real-time monitoring improve the responsiveness of workforce decision-making.
Recurring theme in the review: included studies document use of dashboards/real-time systems and report improved responsiveness in HR operations (thematic synthesis of 47 studies).
Predictive analytics enhances workforce resilience by forecasting turnover, absenteeism, and skill gaps.
Theme extracted from multiple included studies that report or evaluate predictive models for turnover, absenteeism, and skills forecasting (synthesis across reviewed literature).
Analytics shifts HR from an administrative function to a strategic decision-making role.
Thematic analysis across the 47 included studies identified 'strategic imperative of data-driven HRM' as a central theme discussed across multiple papers.
Data-driven HRM (predictive analytics, AI-driven workforce analytics, and real-time monitoring) enables organizations to better anticipate workforce disruptions, improve talent acquisition, and support employee well-being, thereby strengthening workforce resilience.
Synthesis (thematic analysis) of a PRISMA-based systematic review of 47 peer-reviewed studies (2012–2024) identified from Scopus, Web of Science, and Google Scholar; claim derived as the main finding across included studies.
Investment in intangible assets — data governance, process documentation, and change management — is economically essential to appropriate AI value and is costly to build and hard to imitate.
Consistent treatment across conceptual and practitioner literature in the review; grounded in resource-based view framing and multiple case observations.
Returns are highest where AI augments skilled workers (decision support) rather than simply replacing routine tasks; investments in training and new roles are economic complements.
Synthesis of case studies and theoretical literature included in the review emphasizing human-AI complementarity; practitioner reports on training/upskilling outcomes.
AI-enabled ERP can raise measured productivity via faster decisions and automation, but benefits depend on complementary investments in organizational capital; standard productivity metrics may understate gains from improved decision quality.
Conceptual arguments and limited empirical evidence from the literature; review notes scarcity of large-scale causal estimates and measurement challenges.
In supply-chain functions AI is used for demand forecasting, inventory optimization, dynamic routing, and exception management.
Aggregated evidence from case studies, simulation studies, and practitioner reports in the systematic review demonstrating these use cases and reported benefits.
In manufacturing AI supports predictive maintenance, quality control, and production scheduling optimization.
Technical evaluations and empirical case studies included in the review document these applications and associated operational improvements.
In procurement AI is applied to spend analytics, supplier risk scoring, and automated ordering / contract compliance.
Synthesis of practitioner reports and case studies from the 2020–2025 literature showing applied deployments and reported functional impacts.
In finance functions AI is used for automated close, anomaly detection, improved forecast accuracy, and scenario planning.
Multiple case studies and practitioner reports in the reviewed literature describing deployments and measured improvements in financial processes and outputs.