The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (4333 claims)

Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 402 112 67 480 1076
Governance & Regulation 402 192 122 62 790
Research Productivity 249 98 34 311 697
Organizational Efficiency 395 95 70 40 603
Technology Adoption Rate 321 126 73 39 564
Firm Productivity 306 39 70 12 432
Output Quality 256 66 25 28 375
AI Safety & Ethics 116 177 44 24 363
Market Structure 107 128 85 14 339
Decision Quality 177 76 38 20 315
Fiscal & Macroeconomic 89 58 33 22 209
Employment Level 77 34 80 9 202
Skill Acquisition 92 33 40 9 174
Innovation Output 120 12 23 12 168
Firm Revenue 98 34 22 154
Consumer Welfare 73 31 37 7 148
Task Allocation 84 16 33 7 140
Inequality Measures 25 77 32 5 139
Regulatory Compliance 54 63 13 3 133
Error Rate 44 51 6 101
Task Completion Time 88 5 4 3 100
Training Effectiveness 58 12 12 16 99
Worker Satisfaction 47 32 11 7 97
Wages & Compensation 53 15 20 5 93
Team Performance 47 12 15 7 82
Automation Exposure 24 22 9 6 62
Job Displacement 6 38 13 57
Hiring & Recruitment 41 4 6 3 54
Developer Productivity 34 4 3 1 42
Social Protection 22 10 6 2 40
Creative Output 16 7 5 1 29
Labor Share of Income 12 5 9 26
Skill Obsolescence 3 20 2 25
Worker Turnover 10 12 3 25
Clear
Governance Remove filter
This is the first empirical, message-level study of verified chatbot-related psychological-harm cases (as opposed to speculative discussion).
Authors' positioning in paper; claim of novelty based on review of prior literature and their message-level, verified-case approach.
medium positive Characterizing Delusional Spirals through Human-LLM Chat Log... novelty / contribution described (message-level empirical analysis of verified h...
Adopting this approach shifts required skills and organizational roles away from lengthy parametric modeling toward data engineering, controller integration, and monitoring.
Authors' discussion of practical/organizational implications (qualitative); argument based on removal of model-building step and increased emphasis on data infrastructure and online operations.
medium positive Data-driven generalized perimeter control: Zürich case study changes in required skills/organizational roles (qualitative workforce compositi...
DeePC outperforms baseline controllers (e.g., fixed-time and standard adaptive schemes) in the simulated experiments.
Comparative simulation experiments reported in the paper where DeePC-controlled signals achieve superior system-level metrics relative to baseline controllers.
medium positive Data-driven generalized perimeter control: Zürich case study system-level outcomes (total travel time, CO2 emissions) compared across control...
The method was validated on a very large, high-fidelity microscopic closed-loop simulator of Zürich; the paper reports this as the largest such closed-loop urban-traffic simulation in the literature.
Authors' description of the experimental environment: city-scale microscopic simulator of Zürich with controller in the loop; explicit statement in the paper claiming it is the largest closed-loop urban-traffic simulation reported in the literature.
medium positive Data-driven generalized perimeter control: Zürich case study scale of validation (city-scale microscopic closed-loop simulation)
Regularization and the use of measured Hankel/data matrices make the method more robust to measurement noise and limited data.
Method description includes regularization terms in the DeePC optimization and use of Hankel matrices built from measured trajectories; simulation experiments show continued performance under noisy / limited-data conditions.
medium positive Data-driven generalized perimeter control: Zürich case study robustness to measurement noise and limited data (performance degradation metric...
DeePC handles sparse or limited traffic measurements better than many machine-learning methods.
Claims in the paper supported by experiments and methodological notes: use of Hankel structures and regularization in DeePC to operate with limited/sparse sensing; comparative statements versus generic ML methods (qualitative and simulation evidence).
medium positive Data-driven generalized perimeter control: Zürich case study controller performance (e.g., travel time, emissions) under sparse sensing / lim...
The DeePC-based approach avoids the expensive, time-consuming model-building step required by model-based control methods.
Methodological argument and demonstration that controller uses historical input–output trajectories directly rather than requiring separate parametric model identification; supported by simulation implementation that bypasses model identification.
medium positive Data-driven generalized perimeter control: Zürich case study need for explicit parametric model identification (development time/effort proxy...
Legible decision modes and recorded contest pathways improve verifiability and lower information asymmetries, aiding regulators and platforms in monitoring and reducing litigation/reputational risk.
Analytic claim in the implications section; argued conceptually and tied to proposed logging/audit tools; no empirical validation.
medium positive Designing for Disagreement: Front-End Guardrails for Assista... verifiability/auditability (availability of logs), regulator/platform monitoring...
The pattern can reduce costly misallocations caused by LLM unpredictability by constraining policy options, improving overall allocation efficiency in expectation.
Theoretical argument in the paper tying constrained policy space to reduced variability and misallocation risk; no empirical testing or quantitative model provided.
medium positive Designing for Disagreement: Front-End Guardrails for Assista... allocation efficiency (time-to-help, correct-priority assignments, resource util...
The pattern improves legibility, procedural legitimacy, and actionability compared to systems without these elements (proposed as evaluation goals).
Evaluation agenda and proposed user-study metrics in the paper (legibility tests, perceived fairness surveys, contest effectiveness measures); no empirical results yet.
medium positive Designing for Disagreement: Front-End Guardrails for Assista... legibility (user comprehension), procedural legitimacy (perceived fairness), act...
Bounded calibration with contestability avoids opaque silent defaults that mask value choices and avoids wide-open user-configurable value sliders that offload moral choice under stress.
Normative rationale and argumentation in the paper; compared qualitatively against two alternative design approaches; no empirical comparison.
medium positive Designing for Disagreement: Front-End Guardrails for Assista... reduction in hidden value-skews and offloaded moral choice (qualitative assessme...
Bounded calibration with contestability is a viable design pattern for LLM-enabled robots that must allocate scarce, real-time assistance among multiple people.
Conceptual/design proposal in the paper; illustrated with a concrete public-concourse robot vignette; no empirical deployment or sample data reported.
medium positive Designing for Disagreement: Front-End Guardrails for Assista... feasibility/viability of the design pattern (qualitative)
Modular strategy/execution architectures (like ESE) can materially improve the stability and efficiency of LLM-driven operational decision systems, increasing their attractiveness for deployment in retail, logistics, and supply-chain contexts.
Empirical improvements observed with ESE on RetailBench relative to monolithic baselines, coupled with analysis of deployment considerations and domain relevance discussed in the paper.
medium positive RetailBench: Evaluating Long-Horizon Autonomous Decision-Mak... operational stability and efficiency improvements as proxies for deployment attr...
ESE improves operational stability and efficiency relative to baselines that do not separate strategy from execution.
Empirical comparisons reported in the experiments: eight contemporary LLMs evaluated on multiple RetailBench environments, with ESE compared against monolithic LLM agents and other baselines using metrics of operational stability (e.g., variance or frequency of catastrophic failures) and efficiency (e.g., cost/profit/fulfillment).
medium positive RetailBench: Evaluating Long-Horizon Autonomous Decision-Mak... operational stability (variance/frequency of catastrophic failures) and efficien...
ESE enables interpretable and adaptive strategy updates intended to counteract error accumulation and environmental drift.
Design features of the strategy module (slower updates, interpretable strategy representation) and qualitative analysis in the paper linking these features to reduced error accumulation and strategy drift in experiments.
medium positive RetailBench: Evaluating Long-Horizon Autonomous Decision-Mak... interpretability of strategy updates and reduction in error accumulation/strateg...
Model-merging and targeted continual pre-training were used to amplify limited compute and improve performance without full from-scratch pre-training.
Paper describes using model-merging and targeted continual pre-training to leverage existing strong weights and inject language/domain data efficiently.
medium positive Fanar 2.0: Arabic Generative AI Stack performance improvement attributable to model-merging/continual pre-training met...
Prioritizing data quality over raw scale (curated 120B tokens instead of maximizing token counts) produced better Arabic and cross-lingual performance for the resource budget used.
Paper emphasizes a 'data quality over brute-force scale' strategy and reports benchmark improvements from the curated corpus and targeted training; the causal link is asserted via these results.
medium positive Fanar 2.0: Arabic Generative AI Stack model performance relative to data curation strategy
Those benchmark gains were achieved using roughly 1/8th the pre-training tokens of Fanar 1.0 (i.e., about 8× fewer pre-training tokens).
Paper states the approach used approximately 1/8th the pre-training tokens of Fanar 1.0 while improving benchmarks; exact token counts for Fanar 1.0 not provided in the summary.
medium positive Fanar 2.0: Arabic Generative AI Stack relative pre-training token count (Fanar 2.0 vs Fanar 1.0)
Fanar-27B reports benchmark gains relative to Fanar 1.0: Arabic knowledge +9.1 points, language ability +7.3 points, dialect handling +3.5 points, and English capability +7.6 points.
Paper reports these specific numeric benchmark improvements across Arabic knowledge, general language ability, dialects, and English capability; evaluation suite names, sample sizes, and statistical details are not specified in the summary.
medium positive Fanar 2.0: Arabic Generative AI Stack benchmark scores (Arabic knowledge, language ability, dialect handling, English ...
Using entailment-based verifiers can reduce inference compute cost by over two orders of magnitude, lowering marginal compute cost per query compared to LLM-based scorers.
Measured FLOP comparisons between lightweight entailment models and LLM-based scoring in the paper, with reported >100× FLOP reduction.
medium positive Is Conformal Factuality for RAG-based LLMs Robust? Novel Met... compute cost (FLOPs) per verification/query
Lightweight entailment-based verifiers match or exceed LLM-based confidence scorers for scoring atomic claims while consuming >100× fewer FLOPs.
Empirical comparisons in the paper between entailment (NLI) models and LLM-based scoring approaches across the evaluated datasets, with measured FLOPs showing more than two orders of magnitude lower compute for the entailment models alongside equal-or-better scoring performance.
medium positive Is Conformal Factuality for RAG-based LLMs Robust? Novel Met... claim-scoring accuracy/performance and compute cost (FLOPs)
FederatedFactory recovers centralized-model performance without pooling raw data or relying on a central dataset, thereby weakening dependence on foundation-model vendors and their pretrained priors.
Empirical claims that federated results match centralized upper bounds on tested datasets and methodological statement that no external pretrained priors are required; the economic interpretation is drawn from these empirical and methodological properties.
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... performance gap vs. centralized model; dependence on external pretrained priors
FederatedFactory enables exact modular unlearning: deterministic deletion of a client's generative module exactly removes that client's contribution to synthesized datasets.
Design claim in the paper: generative modules are modular assets, and deleting a module deterministically prevents its use when synthesizing the balanced dataset; paper asserts exact modular unlearning and reports it as a property of the method. (No formal auditing metrics or proofs provided in the summary.)
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... unlearning correctness (module-level removal effect on synthesized dataset compo...
Downstream discriminative models trained on the synthesized, balanced datasets avoid conflicting optimization trajectories that cause collapse in standard federated learning under mutually exclusive labels.
Methodological reasoning (balanced synthesized training data removes label heterogeneity across clients) plus empirical demonstrations where standard FL collapses under mutual exclusivity (e.g., CIFAR baseline) and FederatedFactory recovers performance.
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... optimization stability / avoidance of collapsed training (measured indirectly vi...
Across diverse medical imagery benchmarks (including MedMNIST and ISIC2019), FederatedFactory matches centralized upper-bound performance.
Empirical comparisons reported in the paper: FederatedFactory results are compared against a centralized upper bound on the same datasets and reported to be matched. (Details of which datasets and exact numeric comparisons beyond ISIC2019 are not enumerated in the summary.)
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... classification performance vs. centralized upper bound (accuracy/AUROC)
FederatedFactory restores ISIC2019 performance to AUROC = 90.57% under the tested regime.
Empirical experiment reported on ISIC2019 (dermatology images); paper reports AUROC value of 90.57% for FederatedFactory. (Exact train/test splits and client partitioning not specified in the summary.)
FederatedFactory operates without relying on external pretrained foundation models (zero-dependency).
Paper explicitly states the framework does not depend on pretrained foundation models; experiments are reported without using external pretraining (datasets: MedMNIST suite, ISIC2019, CIFAR-10).
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... dependency on pretrained models (binary: uses / does not use)
By synthesizing class-balanced datasets locally from exchanged generative modules, FederatedFactory eliminates gradient conflict among clients' discriminative updates.
Mechanistic argument in the paper (training discriminative models on locally synthesized, balanced data avoids heterogeneity-induced conflicting gradients) supported by empirical recovery of performance in experiments where baselines collapse under label heterogeneity.
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... reduction/elimination of gradient conflict (inferred via improved downstream per...
FederatedFactory reframes federated learning by exchanging generative modules (priors) instead of exchanging discriminative model weights.
Methodological description in the paper: design of FederatedFactory where each client trains/contributes generative modules (class-specific priors) and shares those modules rather than classifier weights. Evidence is the described protocol and experiments that implement that protocol on the reported datasets.
medium positive FederatedFactory: Generative One-Shot Learning for Extremely... unit of federation / protocol (generative modules vs. discriminative weights)
Practical recommendation: buyers and evaluators should demand contamination audits (triangulating lexical, paraphrase, and behavioral probes) and report both raw and contamination-adjusted scores, especially for high-stakes use.
Policy/recommendation section in paper motivated by experimental findings; recommended procedures follow the paper's triage methods (Experiments 1–3) applied to evaluations.
medium positive Are Large Language Models Truly Smarter Than Humans? improvement in evaluation reliability when contamination audits and adjusted rep...
Triangulation across methods reduces false positives and false negatives inherent to any single contamination-detection approach.
Methodological claim supported by design: use of lexical matching, paraphrase diagnostics, and behavioral probes to complement one another and offset single-method blind spots (as reported in robustness section).
medium positive Are Large Language Models Truly Smarter Than Humans? expected reduction in detection error (false positives/negatives) via multi-meth...
Estimated performance uplift from identified contamination ranges from +0.030 to +0.054 absolute accuracy points by category.
Experiment 1 translated contamination prevalence into estimated accuracy gains by simulating model behavior on known-exposed items (method described in paper; category-level simulations yield +0.030 to +0.054 point uplifts).
medium positive Are Large Language Models Truly Smarter Than Humans? estimated accuracy uplift (absolute accuracy points) attributable to contaminati...
Policymakers should prioritize retraining programs, strengthened social protection, and redistributive policies to mitigate automation-induced unemployment and inequality.
Policy recommendation based on the author's synthesis of risks and expert judgment; not based on an empirical intervention study in the paper.
medium positive DIGITAL TRANSFORMATION OF THE RUSSIAN FEDERATION’S SOCIOECON... mitigation of technological unemployment and inequality (employment rates, incom...
There has been progress in software import substitution, contributing to partial technological sovereignty in Russia.
Use of statistics on software import substitution (authors reference national statistics but do not report detailed numbers or methodology).
medium positive DIGITAL TRANSFORMATION OF THE RUSSIAN FEDERATION’S SOCIOECON... software import substitution rate / domestic share of software supply
Digitalization enables management optimization (improved management processes and decision-making) in Russian enterprises and public administration.
Qualitative analysis of policy documents and expert assessment by the author; no empirical evaluation or quantified effect sizes provided.
medium positive DIGITAL TRANSFORMATION OF THE RUSSIAN FEDERATION’S SOCIOECON... management efficiency/optimization (process improvements, decision-making qualit...
Digitalization has produced measurable labor productivity growth in segments of the Russian economy.
Author's interpretation drawing on national statistics and strategic documents; statistical details (period, sectors, sample sizes) not specified in the paper.
medium positive DIGITAL TRANSFORMATION OF THE RUSSIAN FEDERATION’S SOCIOECON... labor productivity (aggregate or sectoral productivity indicators)
Policy implication: prioritize large-scale, targeted reskilling and lifelong learning programs to enable workforce adaptability and capture AI complementarity gains.
Policy recommendations derived from the paper's findings (association between AI adoption and skill shifts, heterogeneous sectoral impacts) and the literature synthesis that links reskilling interventions to better labor outcomes; recommendation is prescriptive rather than empirically tested within the study.
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Policy effect is recommended but not empirically measured in the study (intended...
The paper provides empirical support for the complementarity hypothesis: AI tends to reconfigure jobs and create hybrid roles rather than eliminate employment wholesale.
Convergence of simulated sectoral employment patterns (some sectors showing net gains and hybrid-role growth), the strong correlation between AI adoption and skill shifts (r = 0.71), and corroborating studies from the literature synthesis emphasizing augmentation and hybridization mechanisms.
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Employment change and hybrid job share (evidence for complementarity vs. substit...
Institutional reskilling programs and governance frameworks markedly moderate labor-market outcomes: better frameworks correlate with more complementarities and lower net job loss.
Integration of literature-derived mechanisms with simulated empirical patterns; paper reports correlations/moderation-style comparisons across simulated sector-year cases incorporating policy/institutional variables (described in methods), supported by studies in the systematic review linking policy interventions to labor outcomes.
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Net employment change; measures of complementarity (e.g., hybrid share) conditio...
Healthcare and IT Services experienced net employment gains consistent with AI complementarity (augmented tasks and creation of new hybrid roles).
Simulated sectoral employment trends and net-change metrics for Healthcare and IT Services (2020–2024) presented in the paper, supported by literature synthesis examples showing human–AI complementarities in these sectors.
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Employment levels and net change by sector (Healthcare, IT Services)
The largest rises in hybrid jobs occurred in IT Services and Healthcare.
Sectoral decomposition of hybrid job share trends in the simulated dataset across the seven industries (2020–2024) and supporting qualitative/quantitative findings from the literature synthesis focused on IT Services and Healthcare.
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Hybrid job share by sector (IT Services, Healthcare)
Hybrid human–AI jobs increased substantially across all seven analyzed sectors between 2020 and 2024.
Descriptive trend analysis of the simulated dataset's hybrid job share metric (fraction of roles reclassified as human–AI hybrid) for the seven industries over 2020–2024, combined with corroborating examples from the literature synthesis (selected ACM/IEEE/Springer studies 2020–2024).
medium positive AI-Driven Transformation of Labor Markets: Skill Shifts, Hyb... Hybrid job share (sector-level, 2020–2024)
Responsible implementation requires legal/liability clarity, continuous monitoring for performance drift and distributional shifts, usable explanations, baseline AI literacy for clinicians, and co-design with frontline radiology teams.
Synthesis of governance literature, implementation best-practice reports, and recommendations from usability and deployment studies.
medium positive Human-AI interaction and collaboration in radiology: from co... successful deployment metrics, monitoring alerts for drift, clinician comprehens...
Triage and automation can shorten time-to-diagnosis, increase throughput, and reduce time spent on repetitive tasks.
Observational deployment reports and simulation studies that measured time-to-report or throughput improvements in pilot settings (evidence heterogeneous and context-dependent).
medium positive Human-AI interaction and collaboration in radiology: from co... time-to-diagnosis, studies-per-hour per radiologist, time spent on repetitive ta...
Integration points for AI across the imaging pathway include acquisition (image quality/protocol selection), triage (prioritization), interpretation/reporting (detection, quantification, report pre-population), and post-interpretation (teaching, QA, model improvement loops).
Descriptive synthesis of reported implementations and proposed use cases in the literature and deployment reports across multiple institutions.
medium positive Human-AI interaction and collaboration in radiology: from co... site-level implementation metrics by workflow stage (e.g., reduced repeat scans,...
Human-AI collaboration can produce synergistic gains (diagnostic complementarity) when errors are uncorrelated and tasks are allocated to leverage comparative strengths.
Theoretical/analytical models of error complementarity and empirical reader studies showing instances where combined readings outperform either agent alone (evidence drawn from multiple small-to-moderate reader studies and simulations).
medium positive Human-AI interaction and collaboration in radiology: from co... combined diagnostic accuracy (aggregate sensitivity/specificity), reduction in m...
AI in radiology has clear potential to improve diagnostic performance and workflow efficiency.
Narrative synthesis of laboratory evaluation studies, reader/comparison studies, and a limited number of observational deployment reports showing improved algorithm accuracy and some improvements in measured throughput or time-to-review in pilots (study sizes and settings heterogeneous; few large-scale RCTs).
medium positive Human-AI interaction and collaboration in radiology: from co... diagnostic accuracy (sensitivity/specificity), workflow efficiency (throughput, ...
Performance and reward structures must be redesigned to value oversight, hypothesis testing, escalation and governance behaviours that mitigate model risk but may not immediately increase output.
Managerial recommendation derived from the framework and organizational reward literature; no empirical evaluation provided.
medium positive Symbiarchic leadership: leading integrated human and AI cybe... alignment of incentives; frequency of oversight/governance behaviours; mitigatio...
Firms need new metrics to decompose value created by humans, AI, and their interaction (to distinguish complementarities versus substitution).
Analytic implication derived from the framework and literature on productivity measurement; presented as a recommendation for empirical work rather than tested evidence.
medium positive Symbiarchic leadership: leading integrated human and AI cybe... accuracy of productivity attribution; measurement of human–AI complementarities/...
Symbiarchic leadership is a practical, HR‑oriented framework for leading integrated human–AI “cyber teams,” specifying four linked leadership practices that make AI a co‑actor in knowledge work while preserving human judgement, accountability and organizational legitimacy.
Paper's central proposition based on theoretical synthesis of academic literature on human–AI collaboration, hybrid teams and digital‑era leadership plus illustrative practitioner examples; no original empirical data or experiments.
medium positive Symbiarchic leadership: leading integrated human and AI cybe... ability to lead integrated human–AI teams; preservation of human judgement, acco...