The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (5267 claims)

Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 378 106 59 455 1007
Governance & Regulation 379 176 116 58 739
Research Productivity 240 96 34 294 668
Organizational Efficiency 370 82 63 35 553
Technology Adoption Rate 296 118 66 29 513
Firm Productivity 277 34 68 10 394
AI Safety & Ethics 117 177 44 24 364
Output Quality 244 61 23 26 354
Market Structure 107 123 85 14 334
Decision Quality 168 74 37 19 301
Fiscal & Macroeconomic 75 52 32 21 187
Employment Level 70 32 74 8 186
Skill Acquisition 89 32 39 9 169
Firm Revenue 96 34 22 152
Innovation Output 106 12 21 11 151
Consumer Welfare 70 30 37 7 144
Regulatory Compliance 52 61 13 3 129
Inequality Measures 24 68 31 4 127
Task Allocation 75 11 29 6 121
Training Effectiveness 55 12 12 16 96
Error Rate 42 48 6 96
Worker Satisfaction 45 32 11 6 94
Task Completion Time 78 5 4 2 89
Wages & Compensation 46 13 19 5 83
Team Performance 44 9 15 7 76
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 18 17 9 5 50
Job Displacement 5 31 12 48
Social Protection 21 10 6 2 39
Developer Productivity 29 3 3 1 36
Worker Turnover 10 12 3 25
Skill Obsolescence 3 19 2 24
Creative Output 15 5 3 1 24
Labor Share of Income 10 4 9 23
Clear
Adoption Remove filter
More theoretical work is needed to establish guarantees (consistency, asymptotic behavior, and frequentist coverage) for these networks when applied in economic settings.
Stated research need/caveat in the paper; no new theoretical proofs are provided in the summary to establish these properties.
high null result ForwardFlow: Simulation only statistical inference using dee... theoretical guarantees (absence of established consistency/asymptotic/coverage r...
The Boson Sampling Born Machine (BSBM) is a generative model whose model distribution is the output probability distribution of a linear-optical (bosonic modes) circuit.
Definition and constructive specification in the paper: model architecture described as linear-optical circuits with outputs given by bosonic-mode measurement probabilities (the paper's formal definition/construction). The claim is definitional/theoretical (no empirical sample size).
high null result Universality of Classically Trainable, Quantum-Deployed Boso... model distribution = linear-optical circuit output probabilities
Because this is a conceptual/systems-architecture paper, it does not present new empirical performance benchmarks.
Explicit statement in the paper's Data & Methods section that no new empirical benchmarks are presented.
high null result Reference Architecture of a Quantum-Centric Supercomputer presence or absence of new empirical performance benchmark data
The evaluated models consist of an MLP baseline and a GNN tailored to exploit relational/spatial structure among beams/antennas.
Model descriptions provided in the methods section: two supervised-learning architectures (MLP and GNN) used for beam prediction experiments.
high null result Federated Learning-driven Beam Management in LEO 6G Non-Terr... model architecture comparison (GNN vs MLP)
Using Federated Learning (FL) with orbital planes as distributed learners and HAPS for aggregation avoids centralization of raw channel data.
Method description: federated-learning architecture with clients mapped to orbital planes and HAPS performing coordination/aggregation; explicitly states no central pooling of raw channel samples.
high null result Federated Learning-driven Beam Management in LEO 6G Non-Terr... presence/absence of central pooling of raw channel data
DPS was empirically evaluated across diverse reasoning domains (mathematical reasoning, planning, and visual-geometry) to test generality.
Paper reports experiments on those three categories of tasks; they are listed as the evaluated tasks in the methods/experiments section.
high null result Dynamics-Predictive Sampling for Active RL Finetuning of Lar... task domains evaluated (mathematics, planning, visual-geometry)
DPS uses the inferred per-prompt state distributions as a predictive prior to select prompts estimated to be most informative, avoiding exhaustive candidate rollouts for filtering.
Method and selection mechanism described: predictive prior ranking/filtering replaces rollout-heavy candidate evaluation. (Procedure described in paper; empirical comparisons reported.)
high null result Dynamics-Predictive Sampling for Active RL Finetuning of Lar... selection of prompts (number of candidate rollouts avoided)
Dynamics-Predictive Sampling (DPS) models each prompt’s "extent of solving" under the current policy as a latent state in a dynamical system (a hidden Markov model) and performs online Bayesian inference on historical rollout reward signals to estimate that state.
Methodological description in the paper: DPS uses an HMM representation of per-prompt solving progress and applies online Bayesian updates using past rollout rewards. (No numerical sample size needed for this modeling claim.)
high null result Dynamics-Predictive Sampling for Active RL Finetuning of Lar... inferred latent state distribution / predicted expected learning progress per pr...
The paper does not present large-scale empirical validation; its evidence is primarily theoretical exposition, a constructed illustrative example, and a literature survey.
Explicit description of methods and data in the paper (analysis type: theoretical exposition + illustrative example; no experimental sample reported).
high null result Ergodicity in reinforcement learning presence/absence of empirical experiments or sample-based validation
Local stochastic fluctuations can undo early discovery leads, preventing transient superiority from becoming permanent unless additional asymmetries intervene.
Dynamical analysis of monopolization stage in the model and simulation trajectories showing reversal or loss of early leads in symmetric interaction regimes; theoretical demonstration that fluctuations can destabilize early footholds.
high null result Macroscopic Dominance from Microscopic Extremes: Symmetry Br... persistence of local leads over time (probability of lead reversal due to stocha...
Transient superiority (finding resources faster) by itself does not stabilize a system-wide monopoly; early leads are fragile and can be undone by local stochastic fluctuations.
Analysis of monopolization dynamics and absorbing-state stability within the stochastic spatial model, plus numerical simulations showing symmetric interaction scenarios do not produce robust absorbing monopolies. This is model-based (no empirical validation).
high null result Macroscopic Dominance from Microscopic Extremes: Symmetry Br... long-term persistence/probability of absorbing (system-wide monopoly) state give...
The authors recommend specific measurement metrics and empirical research priorities (e.g., MAPE, stockout frequency, inventory turns, lead times, fill rates, total supply chain cost, service-level volatility, resilience measures; causal studies like diff-in-diff or randomized interventions).
Explicit recommendations in the paper's measurement and research agenda sections.
high null result Optimizing integrated supply planning in logistics: Bridging... listed supply-chain performance and resilience metrics
The study's small sample size and qualitative design limit external generalizability and prevent causal effect size estimation; potential selection and reporting biases exist due to purposive sampling and interview-based data.
Authors explicitly state these limitations in the paper's limitations section.
high null result Optimizing integrated supply planning in logistics: Bridging... external generalizability and causal inference capability
The study is a qualitative multi-case study of five medium-to-large organizations, using semi-structured interviews across procurement, production planning, inventory management, and distribution, analyzed via cross-case comparison.
Methods section description provided by the authors (sample size n = 5, sectors, interview-based primary data, cross-case analysis).
high null result Optimizing integrated supply planning in logistics: Bridging... process-level, qualitative insights into ISP implementation
There is limited empirical causal evidence linking specific explanation types to long-term outcomes (safety, fairness, economic performance) in real-world deployments.
Meta-level finding of the review: authors report gaps in the literature—few causal or longitudinal studies of explanation interventions in deployed, high-stakes settings.
high null result Explainable AI in High-Stakes Domains: Improving Trust, Tran... evidence availability for causal effects on safety, fairness, economic performan...
The literature groups explainability impacts along three linked dimensions — user trust, ethical governance, and organizational accountability.
Analytical result of the review's thematic coding and synthesis across interdisciplinary literature (categorization derived from the reviewed corpus).
high null result Explainable AI in High-Stakes Domains: Improving Trust, Tran... categorization structure of explainability impacts (three-dimension taxonomy)
Research priorities include empirical measurement of task‑level automation rates, firm and industry productivity effects, wage impacts across occupations, and diffusion patterns.
Paper's stated research agenda and identification of measurement gaps; based on methodological critique of current evidence base.
high null result How AI Will Transform the Daily Life of a Techie within 5 Ye... future empirical research outputs on automation rates, productivity, wage impact...
Measuring these productivity gains will be challenging because quality improvements, faster iteration, and creative outputs are harder to price/observe than lines of code.
Methodological argument about measurement difficulty; based on conceptual considerations, not empirical validation.
high null result How AI Will Transform the Daily Life of a Techie within 5 Ye... observability and measurability of productivity gains (availability of suitable ...
Measuring AI's economic impact requires new metrics that account for decision-value uplift, reduced tail-risk exposures, and dynamic gains from continuous learning; causal identification will require experiments or staggered rollouts.
Methodological recommendation backed by conceptual discussion of measurement challenges; no implementation of such measurement approaches is reported in the paper.
high null result Next-Generation Financial Analytics Frameworks for AI-Enable... proposed measurement constructs (decision-value uplift, tail-risk reduction, lea...
Performance and evaluation should be measured using forecast accuracy, decision lift/value added, latency, and false positive/negative rates.
Paper-prescribed evaluation metrics; presented as recommended practice rather than derived from empirical testing within the paper.
high null result Next-Generation Financial Analytics Frameworks for AI-Enable... forecast accuracy, decision lift (value added), system latency, false positive/n...
Core AI techniques for these frameworks include supervised/unsupervised ML, NLP for unstructured text, anomaly detection for control/transaction monitoring, and reinforcement/prescriptive models for recommendations.
Methodological claim listing standard ML/NLP/anomaly-detection techniques and prescriptive approaches; statement of methods rather than an empirical comparison of alternatives.
high null result Next-Generation Financial Analytics Frameworks for AI-Enable... method adoption/type metrics (e.g., frequency of supervised vs. unsupervised met...
Next‑gen frameworks use large-scale structured (transactions, ledgers, KPIs) and unstructured sources (reports, news, contracts, call transcripts) to power models.
Descriptive claim listing data types the paper recommends; presented as design input requirements rather than empirically validated data-integration projects.
high null result Next-Generation Financial Analytics Frameworks for AI-Enable... data coverage and diversity (e.g., proportion of structured vs. unstructured inp...
There is a need for quantitative studies and microdata on firm-level RM practices, AI adoption, and performance outcomes to measure effect sizes and causal pathways.
Stated research gaps and limitations in the review (lack of primary empirical quantification; heterogeneity across contexts).
high null result The Role of Risk Management as an Organizational Management ... availability of quantitative evidence on RM effects (effect sizes, causal estima...
The review's conclusions are limited by reliance on published literature (potential bias toward successful implementations), lack of primary empirical quantification (no effect sizes), and heterogeneity across organizational contexts limiting direct generalizability.
Explicit limitations stated in the paper summarizing scope and method (qualitative literature review, secondary evidence only).
high null result The Role of Risk Management as an Organizational Management ... generalizability and empirical precision of review findings
Heterogeneity in system designs and deployment contexts complicates cross-site comparisons.
Limitations section and observed variation in platform architectures, degrees of automation, and governance across sites reported via descriptive data and interviews.
high null result The Role of Artificial Intelligence in Healthcare Complaint ... comparability across deployment sites (heterogeneity in systems and contexts)
Non-random selection of institutions limits causal inference and external generalizability of the study's findings.
Study limitations explicitly state non-random site selection and heterogeneous deployments; methodological note that causal claims are constrained.
high null result The Role of Artificial Intelligence in Healthcare Complaint ... generalizability and causal inference validity
The study uses a quantitative, cross-sectional survey-based research design of managers and educational administrators and employs descriptive statistics, correlation, and regression analyses.
Methods described in the summary explicitly state research design and analytical techniques; this is a methodological claim rather than an empirical substantive finding. (Sample size not provided in summary.)
high null result Algorithmic Trust and Managerial Effectiveness: The Role of ... research design / analytic approach (methodological description)
Key methodological details are missing or not reported: training/test split, cross-validation scheme, hyperparameter tuning, treatment of confounders/endogeneity, exact definition/measurement of the outcome, and whether results were validated out-of-sample or in field trials.
Summary lists these specific missing methodological elements as not provided in the paper.
high null result AI in food inequality: Leveraging artificial intelligence to... methodological reporting completeness
The paper does not report (or the summary omits) the sample size and full provenance of the Indian farm dataset.
Summary explicitly states that sample size and full provenance of the Indian dataset are not reported.
high null result AI in food inequality: Leveraging artificial intelligence to... reporting completeness for dataset (sample size/provenance)
Data sources used are FAO and Kaggle datasets for global context and a proprietary/field Indian farm dataset for modeling.
Paper cites FAO and Kaggle for global context and uses a proprietary Indian farm-level dataset for the core modeling work (summary notes that full provenance not reported).
high null result AI in food inequality: Leveraging artificial intelligence to... data provenance/source
The chosen ML technique is gradient boosting regression.
Explicit statement in the methods section that gradient-boosting regression was used for modeling.
high null result AI in food inequality: Leveraging artificial intelligence to... modeling technique used
Features used in modeling include pesticide/fertilizer use, farm size, crop type, harvest date, and climatic variables.
Listed predictor variables in the paper's modeling/methods section.
high null result AI in food inequality: Leveraging artificial intelligence to... predictor variables used in the ML model (feature list)
Instrumental-variable (IV) estimation is used to address endogeneity of AI adoption and to identify causal effects on employment and wages.
Paper states IV identification strategy applied to the 38-country panel; robustness checks and alternative specifications reported (paper refers to instrument details in full text).
high null result Artificial Intelligence and Labor Market Transformation: Emp... Causal estimate identification strategy for employment and wage outcomes
The AI Adoption Index is constructed as a composite measure combining enterprise investment in AI, AI-related patent filings, and workforce/firm surveys on AI use across 38 OECD countries (2019–2025).
Paper's methodological description of the index construction; data sources enumerated as investment, patenting, and survey measures over the panel period.
high null result Artificial Intelligence and Labor Market Transformation: Emp... AI adoption intensity (composite index)
There is a need for standardized metrics and measurement protocols for public-sector productivity and non-market outcomes (service quality, processing time, cost per transaction, transparency, trust).
Methodological critique within the review pointing to heterogeneity of outcome measures across studies and calling for standardized metrics; based on synthesis of reviewed literature.
high null result Digital Transformation and AI Adoption in Government: Evalua... existence/adoption of standardized measurement protocols and consistency of repo...
Much of the literature on public-sector digital/AI interventions is descriptive or case-based; causal, quantitative evidence on net productivity effects is limited and context-dependent.
Methodological assessment within the review noting heterogeneous study designs, reliance on secondary sources, and a lack of randomized or quasi-experimental studies; the review explicitly states this limitation.
high null result Digital Transformation and AI Adoption in Government: Evalua... availability of causal quantitative estimates of productivity impacts
Research and monitoring priorities for economists include task-level analyses of substitutability/complementarity, modeling adoption as a function of regulatory costs and reimbursement incentives, and evaluating long-run welfare and distributional effects.
Explicit research recommendations stated in the narrative review, based on gaps identified in the literature and evolving empirical questions.
high null result Will AI Replace Physicians in the Near Future? AI Adoption B... research activity in recommended areas; quality of evidence informing policy
Policymakers and payers should consider liability reform, reimbursement models that reward safe human–AI collaboration, funding for independent clinical validation, and measures to prevent market concentration.
Policy recommendations and implications derived from the narrative review's synthesis of regulatory, economic, and implementation challenges.
high null result Will AI Replace Physicians in the Near Future? AI Adoption B... policy actions implemented (liability reform, reimbursement changes, funding all...
Research priorities include causal studies on AI’s impacts on SME productivity, employment and inequality in LMICs; cost–benefit analyses of financing and policy interventions; evaluation of data governance models; and development of metrics/monitoring systems for inclusive adoption.
Authors' identification of evidence gaps from the structured literature review highlighting areas with insufficient causal or evaluative research.
high null result Artificial Intelligence Adoption for Sustainable Development... existence and quality of targeted causal and evaluative research on AI in LMIC S...
Empirical causal evidence on long-run welfare, distributional outcomes, and labor effects of AI in LMIC SMEs remains thin.
Gap identified through the structured review: few causal studies (e.g., RCTs, natural experiments) addressing long-run effects in LMIC SME contexts.
high null result Artificial Intelligence Adoption for Sustainable Development... availability of causal evidence on welfare, distributional effects, and labor ou...
Heterogeneity in SME types and sectors limits the generalizability of findings about AI adoption and impacts.
Authors' methodological limitation noted in the review: the evidence base spans diverse firm sizes, sectors, and contexts, constraining broad generalization.
high null result Artificial Intelligence Adoption for Sustainable Development... generalizability of reviewed findings across SMEs and sectors
Theoretical framing integrates Resource-Based View (RBV), Dynamic Capabilities (DC), Technology–Organization–Environment (TOE), and Diffusion of Innovation (DOI) to explain how firm resources, learning capacity, organizational and environmental factors shape AI adoption.
Conceptual synthesis performed as part of the literature review; integration based on existing theoretical literature rather than primary empirical testing.
high null result Artificial Intelligence Adoption for Sustainable Development... explanatory scope for AI adoption drivers (theoretical coherence rather than an ...
The systematic review followed PRISMA protocol and analyzed a corpus of 103 items (peer‑reviewed articles and institutional reports) published 2010–2024.
Explicit methodological statement in the paper describing PRISMA use and corpus size/timeframe.
high null result Models, applications, and limitations of the responsible ado... review methodology and corpus characteristics (sample size, timeframe)
Further longitudinal cost-benefit studies, scalability benchmarks, and cross-domain trials are needed to determine when on-prem RAG is the dominant economic choice.
Paper's research & evaluation recommendations calling for additional longitudinal and cross-domain empirical work; presented as a recommendation rather than an empirical finding.
high null result An Empirical Study on the Feasibility Analysis of On-Premise... need for further empirical evidence (longitudinal cost-benefit, scalability, cro...
Human-in-the-loop judgments were central to the paper's relevance/usefulness claims rather than relying solely on synthetic benchmarks.
Methods description explicitly states human evaluation by domain experts was used alongside quantitative benchmarks.
high null result An Empirical Study on the Feasibility Analysis of On-Premise... evaluation method (use of human expert judgments vs synthetic benchmarks)
Research gaps remain: quantifying welfare gains from specific AI applications in extraction (productivity, safety, emissions), evaluating cost-effectiveness of policy bundles, and estimating dynamic returns to data ecosystems and human capital.
Identification of gaps from literature and data coverage in the comparative analysis; calls for future empirical and modelling work.
high null result ADOPTION OF ARTIFICIAL INTELLIGENCE IN THE RUSSIAN EXTRACTIV... magnitude of welfare gains from AI applications; cost-effectiveness metrics for ...
The study is limited by being a single‑country case; contextual factors (regulatory regime, infrastructure capacity, procurement practices) may limit generalizability and the study emphasizes institutional and ethical analysis rather than quantitative measurement of economic impacts.
Explicit limitations reported in the paper summarizing scope and emphasis.
high null result Emerging ethical duties in AI-mediated research: A case of d... generalizability and scope limitations
Methods used include qualitative interviews with researchers and administrators, observation/documentation of tool use, mapping of data flows and third‑party dependencies, and normative/legal analysis contrasting local practices with GDPR principles.
Methods section of the paper as reported in the provided summary.
The study's empirical basis is a qualitative case study centered on environmental science research in Chile that adopts the GDPR as an organizing normative framework.
Paper description of study scope and normative framing (methods and focus described in Data & Methods).
high null result Emerging ethical duties in AI-mediated research: A case of d... study design / empirical basis
There is a need for validated administrative and firm-level data on AI adoption, workplace monitoring, and worker outcomes, and for evaluation of policy interventions (mandated impact assessments, transparency requirements, worker representation rules) using randomized or quasi-experimental designs where feasible.
Research and measurement priorities set out in the commentary based on identified gaps; prescriptive recommendation rather than evidence-based finding.
high null result AI governance under the second Trump administration: implica... availability of validated administrative and firm-level AI adoption data; existe...