Evidence (4137 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Governance
Remove filter
Task content and valence were randomized in the experiment.
Methodological statement in the abstract that task assignments, including their content and valence, were randomized across participants.
The paper presents relevant tradeoffs and design choices across human-LLM archetypes, including decision control, social hierarchies, cognitive forcing strategies, and information requirements.
Qualitative analysis and discussion in the paper synthesizing insights from the literature review and empirical evaluation. Method: thematic synthesis and design analysis. Sample size: based on the review of 113 papers and the clinical-case evaluation (details in full text).
We describe 17 human-LLM archetypes derived from a scoping literature review and thematic analysis of 113 LLM-supported decision-making papers.
Scoping literature review and thematic analysis method; corpus size = 113 LLM-supported decision-making papers (as reported in the paper).
The paper introduces the concept of human-LLM archetypes, defined as re-occurring socio-technical interaction patterns that structure the roles of humans and LLMs in collaborative decision-making.
Conceptual contribution presented in the paper (definition and framing). Method: theoretical/conceptual description in the manuscript. Sample size: not applicable.
By integrating dynamic capabilities theory with a micro foundations perspective, the study proposes a conditional model that reframes the essential challenge from technology adoption to organizational adaptation.
Model/theory construction presented in the paper (conceptual integration). This is a methodological/theoretical claim about the paper's contribution; no empirical validation provided.
This study identifies three types of AI triggers that target routines, cognitive frameworks, and resource allocation.
Proposed taxonomy / typology presented in the paper (theoretical classification). The claim is descriptive of the paper's contribution rather than empirically validated.
The analysis is framed through the integrated lens of the Technology-Organization-Environment (TOE) framework and Institutional Theory to provide a multi-faceted understanding of adoption dynamics.
Stated theoretical framing and analytical approach in the study (methodological claim).
The research synthesizes evidence from a wide array of sources, including recent academic literature by Nigerian scholars, NPA official performance reports, policy documents, and international trade facilitation reports (e.g., UNCTAD).
Explicit description of data sources in the study methodology; method: secondary data synthesis (no sample size applicable).
This study investigates the current state of adoption, the prevailing barriers, and the resultant performance outcomes of digital and AI-driven logistics within Nigeria’s maritime supply chain.
Stated study aim and scope; method: rigorous secondary data analysis drawing on multiple documentary sources (Nigerian academic literature, NPA reports, policy documents, UNCTAD).
This study uses a conceptual and analytical approach to examine the impact of AI and automation on work.
Stated methodology in the paper's abstract/introduction: methodological description that the study is conceptual and analytical; no empirical sample or quantitative data reported.
The study draws extensively on contemporary literature in sustainable supply chain management, healthcare procurement, and ESG governance.
Methodological claim about the paper's research approach: literature review/synthesis across the cited domains (bibliographic evidence within the paper).
A complete evaluation methodology is specified, including baselines and an ablation design.
Paper claims to specify evaluation methodology with baselines and ablation; details presumably in the methods section.
The paper formalizes two testable hypotheses on security coverage and latency overhead.
Explicit statement in the paper that two testable hypotheses are formalized (security coverage and latency overhead); no experimental results shown in the abstract.
The study analyzes the influence of artificial intelligence, financial technology, economic performance, monetary policy, financial development, and governance quality on the growth of G7 countries over 2000–2024 using the Method of Moments Quantile Regression (MMQR).
Statement in paper specifying use of Method of Moments Quantile Regression on G7 countries during 2000–2024. Implied panel sample: 7 countries × 25 years ≈ 175 country-year observations (if annual, balanced panel).
The paper empirically analyzes the algorithm-automated versus human decision-making debate using the AST and STS theoretical lenses.
Theoretical analysis and empirical synthesis across the reviewed studies (n=85), explicitly stated use of AST and STS frameworks to interpret findings.
To address the duality of benefits and harms, the paper proposes a dynamic Human-in-the-Loop (HITL) model that reconciles algorithmic determinism with normative HRM demands.
Conceptual/theoretical contribution presented in the paper (proposed HITL model based on synthesis of findings and theory).
There is substantial heterogeneity in effects (I^2 = 74%), indicating variability across studies.
Meta-analytic heterogeneity statistic reported in the paper (I^2 = 74%).
This study analyzes 28 papers (secondary studies and research agendas) published since 2023.
Systematic literature review conducted by the authors of secondary studies and research agendas; sample size explicitly reported as 28 papers; timeframe specified as 'since 2023'.
Three contributions are presented: the Agentic AI Framework (AAF 3.0); a cross-domain synthesis formalising the inverse evidence–complexity relationship; and a phased sociotechnical roadmap integrating governance sequencing, reimbursement reform, and equity safeguards.
Descriptive claim about the paper's outputs. These contributions are stated in the abstract as the study's deliverables based on the narrative review and synthesis of 81 sources.
Agentic AI is defined as autonomous, goal-directed systems capable of multi-step workflow coordination.
Definition provided by the authors within the paper (conceptual framing used for the review).
This structured narrative review of 81 sources (2020–2025) evaluates whether Agentic AI ... can support structural adaptation in ageing health systems.
Methodological statement in the paper: the study is a structured narrative review of 81 sources from 2020–2025.
The framework is depicted across organization areas with primary focus on strategic management and workforce decision-making and secondary focus on finance, operations, and marketing.
Descriptive claim based on the conceptual framework and its mapping to organizational domains within the paper. No empirical application or case studies reported.
This paper outlines a Human–AI Collaborative Decision Analytics Framework integrating five overlapping layers: data, AI analytics, business analytics interpretation, human judgment, and feedback learning.
Presentation of a conceptual framework developed by the authors (conceptual/modeling contribution). No empirical validation reported.
The results presented in the paper are based on a literature recherche, an analysis of individual tasks across different occupations (conducted within Erasmus+ projects), and discussions with trainers/educators.
Methodological statement from the paper; indicates the types of evidence used. The abstract does not provide numbers for analyzed tasks, the number of occupations, details of Erasmus+ projects, or counts of trainers/educators consulted.
The paper identifies key research gaps and proposes a future research agenda focused on human–AI interaction, organizational governance, and ethical accountability.
Conclusions/recommendations from the conceptual meta-analysis (paper-generated research agenda; no empirical testing reported in abstract).
This study presents a conceptual meta-analysis of interdisciplinary literature on AI-augmented decision-making in organizations.
Methodological statement of the paper (the paper itself is a conceptual meta-analysis); no primary empirical sample reported in the abstract.
Research has insufficiently modeled joint distributional outcomes and environmental performance, and lacks integrated evaluation of AI-enabled sustainable finance under heterogeneous disclosure regimes.
Review-level identification of methodological gaps across the surveyed literature (authors' synthesis of existing studies and their limitations).
There is a shortage of long-horizon causal evidence on non-linear coupling between digitalization and decarbonization, limiting robust policy inference.
Meta-level assessment in the review noting gaps in existing empirical literature (review authors' synthesis of the field; claim about research availability rather than primary data).
Research should prioritize dynamic, task-based models that include transitional frictions, heterogeneous agents, and sectoral structure to better measure AI exposure and impacts.
Methodological recommendation grounded in the paper's theoretical critique of static occupation-level automation metrics and noted empirical gaps.
Timing uncertainty and measurement challenges make forecasting the pace and scale of AI-induced employment change inherently uncertain.
Methodological limitations section noting uncertainty in AI adoption speed and difficulties mapping capabilities to tasks and predicting new occupation emergence.
Research agenda: there is a need for causal studies on AI’s impact on accounting labor demand and firm performance, analyses of distributional effects across firm sizes and industries, and evaluation of regulatory frameworks for reliable, interpretable AI in financial reporting.
Author-stated research priorities drawn from gaps identified in the literature review; not an empirical finding.
Policy implications include workforce retraining, standards for AI auditability and transparency, and regulation balancing innovation and controls (privacy, fraud prevention).
Policy recommendations based on identified risks and barriers discussed in the paper rather than empirical policy evaluation.
For stronger causal evidence, recommended empirical methods include difference-in-differences on adopting firms vs. controls, matched samples, and randomized pilots for particular tools, supplemented by qualitative interviews.
Methodological recommendations stated in the paper (not an empirical finding); no implementation/sample reported in the abstract.
Actionable research priorities include running larger-scale field trials linking game use to observed land-use and economic outcomes, developing validation protocols for game-backed models against empirical on-farm data, studying heterogeneity of impacts, and designing incentive mechanisms that leverage game-demonstrated profitability co-benefits.
Synthesis-driven recommendations based on identified evidence gaps—specifically the predominance of small-scale/qualitative studies and lack of long-term/causal evidence.
Rigorous economic evaluation (RCTs, quasi-experiments) is needed to quantify how game-enhanced DSTs affect investment, land-use choices, emissions outcomes, and farm incomes.
Chapter recommendation grounded in observed gaps: the literature lacks sufficiently rigorous causal impact evaluations; current evidence is largely qualitative or observational.
Personal data are nonrivalrous and highly replicable, so selling data does not follow ordinary scarcity logic.
Analytic/property claim about the economic characteristics of digital information; supported by conceptual definitions and common technical facts about data replication; no empirical sampling needed.
Framing claim: Ideological contests typically produce opposing normative visions (e.g., collectivized command economies vs. market democracies), which makes the development of Western economic theories that portray markets and democracy as dysfunctional puzzling.
Framing and motivation provided in the paper's introduction and background sections; synthesis of conventional expectations about ideological contest outcomes.
The paper uses a qualitative case‑study approach (archival and textual analysis, contextualization, interpretive synthesis) rather than attempting exhaustive quantitative causal identification.
Explicit methods description in the paper: in‑depth historical/institutional examination, archival/textual work, and interpretive synthesis.
The empirical strategy uses baseline panel regressions with standard controls (e.g., firm size, performance, leverage) and fixed effects to estimate the AI → pay relationship.
Methods section describing regression specifications including firm controls and fixed effects applied to the A-share firm panel.
Data consist of a panel of Chinese A-share listed companies covering 2007–2023.
Data description in the paper specifying the sample period and population (A-share listed firms, 2007–2023).
The firm-level AI application indicator is constructed via textual analysis of corporate disclosures (e.g., filings/annual reports) to capture AI application intensity.
Methodological description in the paper describing text-based construction of an AI application indicator from corporate disclosures for listed firms in the 2007–2023 sample.
The empirical approach tests for common long-run relationships across patenting series and identifies structural breaks concentrated after 2010.
Description of empirical strategy: time-series econometric analysis of patent filing series (1980–2019) including tests for common long-run relationships (cointegration) and structural break detection. The paper reports results of these tests (presence/absence of common trends and timing of breaks).
The paper highlights governance risks requiring transparency about LLM-derived mappings, mitigation of model biases, privacy-preserving data practices, and careful communication of uncertainty to avoid overconfident policy recommendations.
Explicit discussion of risks and governance considerations in the paper; this is an acknowledgment rather than an empirical claim. No implementation or audit evidence is provided.
Backtesting the architecture on historical automation waves and recent AI introductions will validate model design and calibration.
Paper explicitly proposes backtesting and holdout validation using historical automation episodes and recent AI adoption events; does not report completed backtests or empirical sample sizes.
Empirical validation of the integrated Kondratieff–Schumpeter–Mandel framework requires firm-level adoption and profitability data, sectoral investment series, and cross-country comparisons using panel methods and identification strategies (e.g., diff-in-diff, IV).
Methods/limitations section recommendation (explicitly states no single micro-econometric identification strategy was reported and outlines required data/methods).
The three frameworks (Kondratieff, Schumpeter, Mandel) are complementary: Kondratieff frames periodicity, Schumpeter provides micro-mechanisms of innovation-driven change, and Mandel foregrounds socio-political constraints and distributional outcomes.
Conceptual integration and comparative theoretical analysis (qualitative synthesis).
Kondratieff's framework is useful for identifying broad periodicities (recurring phases of expansion and stagnation) in capitalist development but is less specific about microeconomic mechanisms.
Theoretical review of Kondratieff literature and conceptual assessment (qualitative).
No new laboratory measurements or datasets are reported in the paper; the approach is methodological and conceptual rather than empirical.
Methods section and explicit statements within the paper noting absence of new data; verifiable by reading the paper.
These operators are presented as conceptual/theoretical bridges rather than immediately quantifiable laboratory units.
Explicit methodological statement in the paper emphasizing interpretive/theoretical intent; no empirical operationalization reported.
The literature is heterogeneous (different LLM families/sizes, prompting techniques, participant persona modeling, environments, and evaluation protocols), which impedes general conclusions about when LLMs reliably mimic humans.
Review notes wide variation across study designs and methods in the 182 studies; inability to produce a single performance estimate motivated unified conceptual framing.