Evidence (2215 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Innovation
Remove filter
Core AI techniques for these frameworks include supervised/unsupervised ML, NLP for unstructured text, anomaly detection for control/transaction monitoring, and reinforcement/prescriptive models for recommendations.
Methodological claim listing standard ML/NLP/anomaly-detection techniques and prescriptive approaches; statement of methods rather than an empirical comparison of alternatives.
Next‑gen frameworks use large-scale structured (transactions, ledgers, KPIs) and unstructured sources (reports, news, contracts, call transcripts) to power models.
Descriptive claim listing data types the paper recommends; presented as design input requirements rather than empirically validated data-integration projects.
The paper is entirely theoretical/analytical and does not report an empirical dataset.
Paper methodology section and abstract state primary tool is an analytical economic model; no empirical data or sample sizes are reported.
The same formal framework can be interpreted as a firm-level model where human skill investment maps onto AI/chatbot investment decisions.
Paper provides an alternative interpretation and formally maps agent skill-investment choices into an analogous firm R&D/AI-capital decision problem within the same mathematical framework.
Research priorities include causal studies on AI’s impacts on SME productivity, employment and inequality in LMICs; cost–benefit analyses of financing and policy interventions; evaluation of data governance models; and development of metrics/monitoring systems for inclusive adoption.
Authors' identification of evidence gaps from the structured literature review highlighting areas with insufficient causal or evaluative research.
Empirical causal evidence on long-run welfare, distributional outcomes, and labor effects of AI in LMIC SMEs remains thin.
Gap identified through the structured review: few causal studies (e.g., RCTs, natural experiments) addressing long-run effects in LMIC SME contexts.
Heterogeneity in SME types and sectors limits the generalizability of findings about AI adoption and impacts.
Authors' methodological limitation noted in the review: the evidence base spans diverse firm sizes, sectors, and contexts, constraining broad generalization.
Theoretical framing integrates Resource-Based View (RBV), Dynamic Capabilities (DC), Technology–Organization–Environment (TOE), and Diffusion of Innovation (DOI) to explain how firm resources, learning capacity, organizational and environmental factors shape AI adoption.
Conceptual synthesis performed as part of the literature review; integration based on existing theoretical literature rather than primary empirical testing.
The paper's empirical and policy conclusions are limited by its jurisdictional sample size (eleven) and reliance on available empirical/operational data, which the authors note is increasingly patchy due to declining transparency.
Methods and limitations sections explicitly noting sample size (eleven jurisdictions) and data availability constraints.
The paper's conclusions are limited by reliance on secondary sources, heterogeneous cross‑study comparisons, limited causal identification of long‑run macro effects, and measurement challenges for AI‑driven intangible capital.
Authors' stated limitations section summarizing the nature of evidence used (qualitative literature review, secondary macro indicators, sectoral examples); this is an explicit self‑reported methodological limitation rather than an external empirical finding.
Economists and policymakers should fund long‑run evaluations (RCTs, quasi‑experimental designs) to estimate causal effects of AI interventions on productivity, welfare, and environmental outcomes.
Evidence‑gap analysis and policy recommendations in the paper; explicit call for rigorous impact evaluation methods given current paucity of long‑run causal evidence.
There are limited long‑run randomized controlled trials (RCTs) on AI/IoT impacts for smallholders and scarce cross‑country data on distributional effects.
Literature review and evidence‑gap identification within the study; explicit statement that long‑run RCTs and cross‑country distributional data are scarce.
Heterogeneous contexts mean impacts vary; careful piloting, monitoring, and adaptive policy are necessary to manage uncertainty in outcomes.
Synthesis and explicit discussion of uncertainties; evidence gaps section noting variable results across regions and interventions.
The paper's conclusions are drawn from a mix of evidence types including literature review, surveys/interviews, case studies, usage-log or publication-metric analyses, and controlled experiments—although the abstract does not specify which of these were actually used or the sample sizes.
Explicitly noted in the Data & Methods summary as the likely underlying evidence types; the paper's abstract itself does not document original data or detailed methods.
This paper is a narrative review synthesizing heterogeneous studies and case reports rather than providing meta-analytic estimates of effect sizes.
Methods statement in the paper describing review type as narrative synthesis and noting limitations (no meta-analysis).
Another important gap is quantifying complementarities between AI and different skill types (evaluative vs. generative tasks).
Review observation that existing empirical work has not systematically quantified how AI productivity gains vary with worker skill composition and complementary roles.
Key research gaps include a lack of long-run causal evidence on the effects of LLMs on firm-level innovation rates, business formation, and industry structure.
Explicit identification of gaps in the literature within the nano-review; the review states that most studies are short-term, task-level, or descriptive.
There is a need for causal, longitudinal studies on how AI‑enabled fintech affects women's portfolio outcomes and on algorithmic interventions designed to reduce gender gaps.
Explicit statement in the paper noting limitations of existing literature (heterogeneity, limited longitudinal causal evidence, possible platform sample selection).
SECaaS offerings commonly include threat intelligence, managed detection & response (MDR), endpoint protection, IAM, CASB, security orchestration/automation, and compliance-as-a-service.
Survey of SECaaS product categories in industry reports and vendor catalogs; technical benchmarks describing typical feature sets.
Achieving CIA in the cloud requires technical controls (encryption, access controls, IAM, MFA, zero-trust), resilience measures (backups, redundancy, DR/BCP), and continuous monitoring (logging, SIEM, EDR/XDR).
Synthesis of technical best practices and vendor/industry guidance; supported by technical evaluations and case studies in the literature.
Core cloud security goals remain confidentiality, integrity, and availability (CIA).
Canonical security literature and standards cited in the chapter; general consensus across technical controls and industry best-practice frameworks (e.g., NIST, ISO).
Limitation: the study analyzes national‑level formal policy texts only and does not measure enforcement, implementation outcomes, or public reactions.
Author‑stated limitations in the paper specifying scope restricted to formal policy documents and absence of empirical enforcement/compliance data.
The paper uses qualitative content analysis, coding documents against the four analytical dimensions to generate a comparative typology of coordination approaches.
Method description: manual qualitative coding of the 36 documents into the specified dimensions, producing the typology distinguishing Chinese and U.S. approaches.
The study's empirical basis comprises 36 national‑level policy documents (18 from China; 18 from the United States) focused on scientific data governance.
Author‑reported dataset and sampling description in the Data & Methods section.
The comparative analysis is organized across four dimensions: coordination objectives, institutional actors, governance mechanisms, and stakeholder legitimacy.
Methodological design reported in the paper; documents were coded against these four analytic categories.
The legal arguments create some uncertainty about scope and enforcement timelines; economic actors will respond to expected enforcement probabilities and expected sanctions, so clarity from regulators or courts will shape the ultimate economic effects.
Doctrinal acknowledgement of legal uncertainty combined with standard economic modeling of regulatory expectations; no empirical modeling in the Article.
The paper is primarily legal/policy scholarship rather than an empirical assessment of the prevalence or magnitude of discrimination in EdTech; it does not provide econometric estimates of harm.
Explicit limitation noted in the Article (self‑reported).
The Article's evidence consists of illustrative case law and statutory text rather than empirical datasets; it builds doctrinal chains, hypotheticals, and applications of statutory language to modern procurement and EdTech deployment models.
Explicit description of evidence and limits in the Article (self‑reported).
Methodologically, the paper uses doctrinal legal analysis and policy argumentation — close reading of federal civil‑rights statutes, administrative guidance, and judicial decisions interpreting 'recipient' and 'federal financial assistance.'
Explicit methodological statement in the Article (self‑reported).
The legal argument is grounded in statutory interpretation and precedent about the scope of 'recipient' and how federal financial assistance flows and influence should be understood.
Doctrinal analysis of statutes, administrative guidance, and judicial decisions cited and discussed in the Article.
Techno‑economic assessments (TEA) and life‑cycle analyses (LCA) are necessary research tools to compare bio‑routes to incumbent chemical synthesis on cost and emissions, and current literature is incomplete in this regard.
Review notes the presence of some TEA/LCA studies but highlights gaps and heterogeneity in methods and results across case studies; many processes lack published TEA/LCA at commercial scales.
Dataset composition: 261 publicly traded U.S. financial firms matched to CFPB complaint records, monthly observations covering 2018–2023.
Data description in the paper: CFPB complaint records matched to 261 firms with monthly panel from 2018 through 2023 used in all reported analyses.
The paper does not make strong causal claims; causal interpretation is limited and future work should address endogeneity and reverse causality (e.g., with event studies or instrumental variables).
Authors explicitly note limitations on causal interpretation and recommend methods (event studies, IVs, natural experiments) for future causal identification.
Fixed-effects panel path models are used to control for firm-level heterogeneity and to estimate direct and mediated relationships between complaint features and abnormal returns.
Econometric approach described: panel path models with firm fixed effects (monthly firm–level data for 261 firms, 2018–2023) to parse direct/mediated associations between complaint measures and returns.
Econometric approach relies on cross-country panel regressions and interaction terms to assess direct effects and complementarities; identification is associative (panel variation + controls) rather than claiming causal identification using instruments or natural experiments.
Paper describes use of panel regressions with interaction terms and emphasizes that identification comes from panel variation and covariate controls, without detailing stronger causal identification strategies.
Models control for key macroeconomic covariates (e.g., GDP per capita, trade openness, human capital, institutional quality) to isolate technology effects.
Paper documents inclusion of macro controls in regression models to reduce omitted-variable bias.
Dependent variable is a composite national Sustainable Development Goal (SDG) performance index (aggregate/summary measure).
Paper specifies the dependent variable as an aggregate SDG performance measure used in the panel regressions.
Unit of analysis is country-year observations for G20 members covering 2015–2023.
Paper states sample and scope as a cross-country panel of G20 economies from 2015–2023 (panel dataset). (Up to 20 countries × 9 years = up to 180 country-year observations, depending on coverage).
The paper's empirical approach is primarily qualitative and interpretive: a systematic literature review plus comparative qualitative case studies, using policy documents, public diplomacy examples, development initiatives, technology export and standards behaviour, and secondary empirical studies as evidence.
Methods section of the paper explicitly states the approach and evidence types; sample of four comparative cases (US, China, EU, Russia) is specified.
The paper demonstrates different mixes and institutional practices of smart power in practice by applying the framework to the United States, China, the European Union, and Russia.
Explicit comparative qualitative case studies of four major international actors (sample size: four cases) using policy documents, public diplomacy examples, and development/technology initiatives as illustrative evidence.
Empirical validation of the book’s proposals would require complementary case studies, model documentation, and outcome measurements.
Author/reviewer recommendation in the blurb about methodological limitations and next steps; not an empirical finding.
The book is predominantly conceptual and policy-analytic and uses illustrative case vignettes rather than presenting a single empirical study.
Explicit methodological description in the Data & Methods blurb: synthesis of technical ideas, governance requirements, and illustrative vignettes; no empirical sample or experimental protocol described.
The research program is grounded in 12 years of forensic legal research spanning 2014–2026.
Author-stated research timeline and methodology (2014–2026 forensic legal research).
The protocol is underpinned by a forensic audit of approximately 4,200 specialized texts (legal doctrine, regulation, standards, technical literature).
Stated corpus and audit in the Methods section: ~4,200 texts reviewed as part of the forensic audit.
The protocol systematizes arguments for 16 projected rulings at Mexico’s Supreme Court (SCJN) to anchor the proposed rights and rules in constitutional practice.
Doctrinal projection and constitutional strategy section of the compendium describing 16 projected SCJN rulings (method: legal projection/modeling).
The compendium’s findings and recommendations are based on a forensic audit of approximately 4,200 specialized texts covering doctrine, jurisprudence, regulation and technical literature.
Stated methodological claim in the compendium: forensic corpus audit of ~4,200 texts (sample size reported).
Methodological claim: combining fixed-effects panel estimation, mediation analysis, and panel threshold models is an effective multi-method approach to (a) estimate average effects, (b) unpack causal channels, and (c) detect nonlinear stage-dependent impacts.
The paper's applied methodology: fixed-effects panel regressions, mediation framework, and panel threshold modeling on the 2012–2022 provincial panel.
The paper constructs a multidimensional digitalization index composed of digital infrastructure, digital service capacity, and the digital development environment.
Index construction described in data/methods: composite indicator combining measures of connectivity/broadband (infrastructure), e-commerce/digital finance (service capacity), and policy/institutional/human capital indicators (development environment).
Attributing productivity changes specifically to AI requires causal identification beyond VIS accounting (e.g., experiments, instrumental variables, difference-in-differences).
Paper notes that VIS is an accounting framework and that causal attribution to AI requires econometric/experimental methods beyond input–output accounting.
The method uses BEA for industry output and industry-by-industry transactions, BLS for employment and hours worked, and IMPLAN for detailed input–output structure and sector mapping; coverage period is 2014–2023.
Explicit data sources and time coverage stated: public BEA, BLS, and IMPLAN annual data 2014–2023 used to construct input–output matrices and labor measures.