The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (4560 claims)

Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 378 106 59 455 1007
Governance & Regulation 379 176 116 58 739
Research Productivity 240 96 34 294 668
Organizational Efficiency 370 82 63 35 553
Technology Adoption Rate 296 118 66 29 513
Firm Productivity 277 34 68 10 394
AI Safety & Ethics 117 177 44 24 364
Output Quality 244 61 23 26 354
Market Structure 107 123 85 14 334
Decision Quality 168 74 37 19 301
Fiscal & Macroeconomic 75 52 32 21 187
Employment Level 70 32 74 8 186
Skill Acquisition 89 32 39 9 169
Firm Revenue 96 34 22 152
Innovation Output 106 12 21 11 151
Consumer Welfare 70 30 37 7 144
Regulatory Compliance 52 61 13 3 129
Inequality Measures 24 68 31 4 127
Task Allocation 75 11 29 6 121
Training Effectiveness 55 12 12 16 96
Error Rate 42 48 6 96
Worker Satisfaction 45 32 11 6 94
Task Completion Time 78 5 4 2 89
Wages & Compensation 46 13 19 5 83
Team Performance 44 9 15 7 76
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 18 17 9 5 50
Job Displacement 5 31 12 48
Social Protection 21 10 6 2 39
Developer Productivity 29 3 3 1 36
Worker Turnover 10 12 3 25
Skill Obsolescence 3 19 2 24
Creative Output 15 5 3 1 24
Labor Share of Income 10 4 9 23
Clear
Productivity Remove filter
Insurance markets may price AI-specific fraud risk, raising premiums or creating new products (AI-fraud insurance).
Speculative economic implication suggested by the authors; no market data or insurer statements cited.
low negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... changes in insurance pricing or product offerings attributable to AI-specific fr...
Vendors offering integrated governed hyperautomation stacks may capture premium pricing and increase switching costs, potentially widening adoption gaps between large incumbents and SMEs.
Market-structure and competitive dynamics discussed theoretically in the Implications section; no market-share or pricing data provided.
low negative Governed Hyperautomation for CRM and ERP: A Reference Patter... vendor pricing premiums; switching costs; differential adoption by firm size (ma...
Exposure to AI and platform work produces psychosocial effects for workers, including increased job insecurity, stress, and changing task content in surviving occupations.
Surveys, qualitative case studies, and workplace studies summarized in the review reporting worker‑reported insecurity and stress; the review also highlights inconsistent measurement and limited systematic evidence on psychosocial outcomes.
low negative The Impact of AI Machine Learning on Human Labor in the Work... job insecurity, stress, psychosocial wellbeing, and perceived changes in task co...
Regulators and standard-setters who value transparency and auditability will need to account for the gap between evaluation results and actionable fixes; firms may require incentives or rules to ensure evaluation leads to remediation, not just documentation.
Authors' policy implication derived from the study's finding of a results-actionability gap and discussion of auditability concerns; speculative recommendation rather than empirical finding.
low neutral Results-Actionability Gap: Understanding How Practitioners E... policy/regulatory effectiveness regarding evaluation leading to remediation (spe...
Delegation of oversight and reallocation of monitoring tasks due to AI integration changes transaction costs and affects organizational design and governance needs (e.g., more verification/audit effort or specialist oversight roles).
Based on participants' reported shifts in who performed monitoring/oversight tasks in the 40 interviews and the authors' interpretation of those shifts in organizational/economic terms.
low neutral AI in project teams: how trust calibration reconfigures team... transaction/monitoring costs and governance arrangements
The paper is the first systematic integration of XAI-based predictive modeling with counterfactual policy simulation specifically targeted at sustainability-oriented HR (Green HRM).
Authors' novelty claim stating this combination is novel in the Green HRM literature; no systematic literature review evidence provided in the summary to independently verify primacy.
low null result Explainable AI for Employee Retention in Green Human Resourc... novelty of methodological integration (claim about state-of-the-art)
The paper likely includes ablation studies and standard metrics (task success rate, step-wise error, plan coherence) to isolate contributions of the two training stages and to evaluate performance.
Summary states these analyses as 'likely additional methods' (i.e., typical but not fully detailed in the abstract); no direct confirmation or results provided in the provided text.
low null result Anticipatory Planning for Multimodal AI Agents task success rate, step-wise error, plan coherence (if present)
This study represents the first attempt to conduct a comprehensive evaluation of artificial intelligence (AI) and its influence on job displacement based on the existing body of literature.
Author assertion in the paper; the excerpt provides no external verification (no citation of prior reviews/meta-analyses to justify the 'first attempt' claim).
low null result A Study on Work-Life Balance of Women Employees in the IT Se... comprehensiveness of literature-based evaluation of AI's influence on job displa...
Results are robust across the authors' reported robustness checks.
Author statement that multiple robustness checks were performed and the main findings persist (the summary does not enumerate the checks or report their outcomes).
low null result Is digital trade affecting city house prices? An artificial ... city-level house prices
Observable firm-level and economy-wide moments—changes in spans of control, manager share of payroll, incidence of new tasks, employment growth, and shifts in the wage distribution—can be used to test the model's predictions.
Model-implied empirical identification strategy and suggested measurable moments in the paper's discussion/implications section (theoretical prediction, not an empirical test).
low null result AI as Coordination-Compressing Capital: Task Reallocation, O... empirical testable moments (spans of control, manager payroll share, new-task in...
The PIER architecture (physics-informed state construction, demonstration-augmented offline data, decoupled post‑hoc safety shield) transfers to wildfire evacuation, aircraft trajectory optimization, and autonomous navigation in unmapped terrain.
Claim of transferability stated in the paper; the excerpt does not include experimental details or quantitative results for these domains.
low positive Physics-informed offline reinforcement learning eliminates c... transferability of the PIER architecture to other domains (qualitative claim)
Expect rising demand and wage premia for managers with hybrid capabilities (systems thinking + computational literacy), with a risk of widening returns to managerial skill heterogeneity.
Theoretical implication from predicted complementarities and task reallocation; prescriptive economic inference without empirical labor-market evidence in the paper.
low positive Comparative analysis of strategic vs. computational thinking... labor demand, wage premia, and distributional widening across managerial skill t...
Managers’ time will be reallocated toward hybrid tasks (interpretation, oversight, ethical deliberation), increasing returns to combined strategic and computational skills.
Predictive inference from the role reconfiguration analysis and task-complementarity argument; forward-looking theoretical forecast (no empirical time-use data).
low positive Comparative analysis of strategic vs. computational thinking... managerial time allocation (share devoted to hybrid tasks) and returns/wage prem...
Standards for provenance, labeling of AI-generated content, and interoperable evidence formats would lower verification costs and create beneficial network effects.
Policy recommendation derived from identified verification frictions and the study's analysis of data/model governance needs.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... verification cost and interoperability/network effects
There is growing market demand for AI-assisted fact-checking tools, creating opportunities for software, monitoring services, and labeled datasets.
Analytic implication drawn from findings about increasing AI use and needs for automation/labeling; based on interviews and market inference in the study.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... market demand for AI tools and labeled datasets
Regulators should consider guidelines on AI monitoring, algorithmic fairness in performance evaluations, and protections to prevent hybrid‑induced career penalties.
Policy recommendation based on conceptual assessment of risks identified in literature synthesis; not an empirical claim—no policy evaluation data provided.
low positive The Sociology of Remote Work and Organisational Culture: How... existence/applicability of regulatory guidelines; protections against career pen...
Hybrid agency implies complementarity between GenAI and managerial/knowledge‑worker skills (curation, evaluation, coordination), potentially increasing returns to those skills while automating routine cognitive tasks—consistent with skill‑biased technological change.
Synthesis of recurring themes linking GenAI capabilities with managerial skill topics in the thematic clusters; positioned as an implication for labour demand and skill composition rather than an empirically tested effect.
low positive Generative AI and the algorithmic workplace: a bibliometric ... expected changes in returns to managerial/knowledge‑worker skills and automation...
There is demand for tooling that bridges evaluation outputs to actionable fixes (e.g., failure-mode libraries, standardized remediation templates, evaluation-to-priority mapping), signaling economic opportunities for third-party tools and consulting services.
Authors' inference based on the documented results-actionability gap and participants' descriptions of pain points; presented as a market implication rather than direct market measurement.
low positive Results-Actionability Gap: Understanding How Practitioners E... inferred market demand for evaluation-to-action tooling/services
Firms that invest in instrumentation, cross-functional processes, and remediation levers capture more value from LLMs; organizations with better evaluation-to-action pipelines will obtain higher productivity gains and market edge.
Authors' inference from observed heterogeneity among teams in the interviews and comparison of practices in teams that reported more success converting evaluations into changes.
low positive Results-Actionability Gap: Understanding How Practitioners E... relative productivity/value capture tied to evaluation-to-action capability (inf...
Public investments in standards, verification infrastructure, and public-interest datasets can correct market failures and support trustworthy AI.
Policy recommendation informed by governance and public-good theory and examples from the literature; the claim is prescriptive and not validated by new empirical evidence within the paper.
low positive The Evolution and Societal Impact of Artificial Intelligence... trustworthiness of AI systems and correction of market failures via public inves...
By lowering single-GPU resource requirements and improving throughput, SlideFormer can democratize domain adaptation and fine-tuning of large models on commodity single-GPU hardware (reducing the need for multi-GPU clusters).
Argumentative implication based on reported throughput, memory, and capacity improvements (e.g., enabling 123B+ models on a single RTX 4090 and reducing memory usage). This is an extrapolation from experimental results rather than a directly measured socio-economic outcome.
low positive An Efficient Heterogeneous Co-Design for Fine-Tuning on a Si... accessibility / feasibility of single-GPU fine-tuning (qualitative economic impl...
Regulators may prefer systems that support contestability and audit trails and could mandate argumentation-style explainability in certain sectors.
Speculative policy prediction; no regulatory statements or empirical policy adoption evidence cited.
low positive Argumentative Human-AI Decision-Making: Toward AI Agents Tha... regulatory adoption rate of contestability/audit-trail requirements
Better contestability may reduce litigation and regulatory frictions if decisions are transparently defensible.
Speculative legal-economic claim; no case studies or empirical legal analysis provided.
low positive Argumentative Human-AI Decision-Making: Toward AI Agents Tha... frequency/cost of litigation and regulatory disputes post-adoption of contestabl...
New service layers may emerge (argumentation-as-a-service, audit firms, explanation certification, human-in-the-loop orchestration platforms).
Speculative market/industry evolution claim based on analogous tech-service cretions; no empirical evidence.
low positive Argumentative Human-AI Decision-Making: Toward AI Agents Tha... emergence and market size of new service verticals around argumentative AI
Collaborative VR features can change team workflows (remote, synchronous inspection sessions), potentially lowering coordination costs across geographically distributed teams.
Paper lists collaborative multi-user sessions as a planned capability and posits organizational effects; no user studies or measurements of coordination cost savings presented.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... coordination costs / team workflow efficiency in distributed teams
Public funding for shared VR-capable data-exploration infrastructure could yield high leverage by improving returns on large observational investments.
Policy recommendation deriving from the platform and ROI arguments in the paper; no cost-benefit analysis or quantified ROI provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... policy leverage (ROI) from funding shared VR infrastructure
Using iDaVIE increases the usable fraction of large observational datasets by improving QC and annotation throughput, thereby raising returns to telescope investments and downstream AI efforts.
This is an inferred implication in the paper (returns-to-scale/platform effects) based on improved QC/annotation throughput; no empirical measurement of usable-fraction increases provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... usable fraction of observational datasets and downstream value for AI/modeling
Higher-quality labels produced via immersive inspection can reduce label noise and lower required training-data sizes for a target ML performance level.
Paper presents this as an implication/expected outcome based on improved annotation quality from immersive inspection; no empirical ML training experiments or quantitative reductions reported.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... label noise level and required training-data size for target model performance
iDaVIE demonstrably reduces cognitive load for multidimensional-data tasks compared with 2D-slice inspection.
Paper asserts reduced cognitive load and faster, more intuitive exploration as an aim and reported outcome; no formal user-study metrics, sample size, or statistical analysis provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... cognitive load (mental effort) for multidimensional-data inspection
The inverse-specification reward offers a domain-agnostic, holistic metric for fidelity to user intent and is recommended for measurement of model value/service quality.
Method introduces inverse-specification reward and asserts domain-agnostic applicability; recommendation based on its conceptual ability to recover briefs as fidelity measure (not necessarily validated across many domains).
low positive Learning to Present: Inverse Specification Rewards for Agent... Utility of inverse-specification recovery accuracy as a fidelity metric (concept...
High-quality automated slide generation has potential to reduce time spent on business presentation creation and produce productivity gains with partial substitution of routine creative/knowledge-worker tasks.
Empirical demonstration of near-SOTA automated slide generation capability on 48 briefs; domain-level economic implication extrapolated from performance improvements.
low positive Learning to Present: Inverse Specification Rewards for Agent... Potential time savings/productivity gains (not directly measured in the study)
Deploying BATQuant with reliable 4-bit weight/activation quantization for MXFP-capable accelerators reduces memory footprint and memory-bandwidth pressure, enabling higher throughput and lower per-token inference costs.
Argumentative / economic analysis in the paper linking reduced precision and parameter storage to lower memory/bandwidth requirements and inferred throughput/cost improvements; not presented as a direct empirical measurement of cost per token in production environments in the summary.
low positive BATQuant: Outlier-resilient MXFP4 Quantization via Learnable... Inferred system-level outcomes: memory footprint, memory-bandwidth usage, throug...
The methodological template (train an ML surrogate of a costly simulator and embed it in an optimizer) generalizes beyond Doherty power amplifiers to other analog/microwave components and broader engineering domains.
Paper proposes generality of approach in implications section; no experimental demonstrations beyond the Doherty PA case are provided in the summary.
low positive Deep Learning-Driven Black-Box Doherty Power Amplifier with ... applicability/generalizability of the surrogate+optimizer methodology to other d...
Investment in data quality and feature engineering yields tangible predictive gains for workforce performance models.
Paper emphasizes use of engineered features capturing engagement dynamics and learning trends and reports better model performance relative to baseline; however, no isolated ablation study quantifying the sole contribution of data-quality investments is reported in the summary.
low positive Adoption of AI-Based HR Analytics and Its Impact on Firm Pro... Predictive performance gains attributable to data quality/feature engineering (i...
Tools that improve detection or quantification may reduce downstream costs from missed diagnoses or unnecessary follow-ups, improving cost-effectiveness in some scenarios.
Economic modeling and limited observational analyses that extrapolate diagnostic improvements to downstream resource use; direct empirical cost-effectiveness studies are scarce.
low positive Human-AI interaction and collaboration in radiology: from co... downstream healthcare utilization (additional tests, treatments), cost per diagn...
The metacognitive reliability metric can reduce adoption risk for purchasers by providing transparent error-risk assessments and enabling performance-based autonomy thresholds.
Conceptual claim supported by the existence of an empirical confidence metric from the recursive meta-model and discussion of procurement/decision-making implications; not empirically tested with purchasers or procurement outcomes.
low positive Human Autonomy Teaming and AI Metacognition in Maritime Thre... adoption risk (qualitative or procurement decision proxies)
HACL/CS supports human trust and situational awareness.
Human factors measured with trust and situational awareness questionnaires in the simulation; summary reports supportive effects on trust and situational awareness but lacks sample-size/statistical detail.
low positive Human Autonomy Teaming and AI Metacognition in Maritime Thre... self-reported trust and situational awareness scores
Intelligent turn-level assignment can reduce costly human attention to only high-value moments, improving overall system productivity.
Conceptual implication from the assignment-layer design and empirical trade-offs reported; presented as an advantage in the paper rather than a directly measured economic productivity study.
low positive Hierarchical Reinforcement Learning Based Human-AI Online Di... distribution of human attention / system productivity (conceptual, not directly ...
HADT demonstrates a concrete way to substitute expensive human diagnostic labor with AI assistance while preserving high accuracy, implying reductions in marginal cost per consultation.
Inference drawn in the paper's implications section based on reported reductions in required human effort and maintained diagnostic accuracy (economic claim extrapolating from experimental results; not directly measured as cost in experiments).
low positive Hierarchical Reinforcement Learning Based Human-AI Online Di... implied marginal cost per consultation (not directly measured)
Organizational norms and UX influence adoption rates and diffusion of AI: social calibration processes at the team level matter for adoption beyond individual cost–benefit calculations.
Reported by interviewees (N=40) as factors shaping whether and how teams incorporated AI into routines; integrated into theoretical implications for diffusion modeling.
low positive AI in project teams: how trust calibration reconfigures team... AI adoption/diffusion rates at team/organization level
Well-calibrated trust tends to encourage AI being used as a complement to human labor (augmentation), increasing effective productivity; miscalibration (over- or under-trust) can lead to productivity losses.
Inferential claim drawn from interviewees' accounts of when teams appropriately relied on AI (augmentation) versus when inappropriate reliance or avoidance occurred; supported by thematic interpretation rather than quantitative measurement.
low positive AI in project teams: how trust calibration reconfigures team... productive use of AI (complementarity vs substitution) and effective productivit...
Policymakers should support standards for auditability, human‑in‑the‑loop thresholds and training subsidies to reduce coordination failures and make the social benefits of AI adoption more widely shared.
Normative policy recommendation derived from the paper’s analysis of risks, governance needs and distributional concerns; not empirically validated within the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... adoption of standards; breadth of social benefits; coordination failure reductio...
Organisations will invest more in training for AI‑related sensemaking, trust calibration and governance competencies; returns to such training should be evaluated relative to investments in model quality.
Prescriptive inference from the framework and human‑capital theory; supported by referenced literature but not empirically tested in this paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... training investment levels; returns on training; comparative returns vs model in...
Explicit comparative‑advantage allocation will shift the composition of tasks across humans and AI, altering demand for routine versus non‑routine skills and potentially increasing demand for high‑level judgement, oversight and sensemaking skills.
Projected labour‑market implication based on theoretical reasoning and prior literature on task‑based skill demand; not empirically estimated in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... task composition; demand for routine vs non‑routine skills; demand for oversight...
Operationalising the four symbiarchic practices through updated HR systems lets firms capture AI‑enabled productivity gains without eroding trust, ethics or employee well‑being.
Normative claim based on theoretical synthesis and managerial prescription; no empirical testing or field data presented in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... AI‑enabled productivity gains; employee trust; ethical outcomes; employee well‑b...
Public data sharing, reproducibility standards, and shared benchmarks could raise the floor of AI utility across the industry.
Policy implication grounded in arguments about data quality, coverage, and generalizability from the narrative review; speculative recommendation rather than evidence-backed empirical claim.
low positive Learning from the successes and failures of early artificial... baseline AI performance/utility across firms (industry-wide)
There is potential for consolidation as firms acquire data, talent, or validated AI-driven assets.
Industry-structure implication drawn from economics of complementary assets and observed M&A activity patterns; presented as a likely trend rather than demonstrated empirically in the paper.
low positive Learning from the successes and failures of early artificial... M&A activity targeting AI capabilities, data assets, or relevant talent
AI startups that demonstrate validated, reproducible wet-lab outcomes and access to high-quality data are more likely to command premium valuations.
Argument from observed market behavior and economics of complementary assets presented in the narrative; no systematic valuation analysis included.
low positive Learning from the successes and failures of early artificial... startup valuation premium tied to validated wet-lab results and data access
Investors should recalibrate expectations: greater value accrues to firms that integrate AI with experimental pipelines and proprietary data assets rather than firms that only possess AI capability.
Economics-focused implications drawn from thematic analysis of heterogeneity in firm outcomes and integration requirements; market-practice inference rather than empirical valuation study.
low positive Learning from the successes and failures of early artificial... firm valuation / investor returns conditional on AI integration and data assets
By integrating psychological trust factors with cognitive capability optimisation, this model offers actionable insights for knowledge management practitioners implementing AI‑augmented decision systems while advancing theoretical understanding of human–AI collaboration effectiveness.
Integrative theoretical claim based on combining constructs from psychological trust research and cognitive/capability literature via systematic synthesis; no empirical evaluation reported in the abstract.
low positive Optimising Human– AI Decision Performance: A Trust and Cap... actionability for practitioners / advancement of theoretical understanding / ove...