The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (1902 claims)

Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 369 105 58 432 972
Governance & Regulation 365 171 113 54 713
Research Productivity 229 95 33 294 655
Organizational Efficiency 354 82 58 34 531
Technology Adoption Rate 277 115 63 27 486
Firm Productivity 273 33 68 10 389
AI Safety & Ethics 112 177 43 24 358
Output Quality 228 61 23 25 337
Market Structure 105 118 81 14 323
Decision Quality 154 68 33 17 275
Employment Level 68 32 74 8 184
Fiscal & Macroeconomic 74 52 32 21 183
Skill Acquisition 85 31 38 9 163
Firm Revenue 96 30 22 148
Innovation Output 100 11 20 11 143
Consumer Welfare 66 29 35 7 137
Regulatory Compliance 51 61 13 3 128
Inequality Measures 24 66 31 4 125
Task Allocation 64 6 28 6 104
Error Rate 42 47 6 95
Training Effectiveness 55 12 10 16 93
Worker Satisfaction 42 32 11 6 91
Task Completion Time 71 5 3 1 80
Wages & Compensation 38 13 19 4 74
Team Performance 41 8 15 7 72
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 17 15 9 5 46
Job Displacement 5 28 12 45
Social Protection 18 8 6 1 33
Developer Productivity 25 1 2 1 29
Worker Turnover 10 12 3 25
Creative Output 15 5 3 1 24
Skill Obsolescence 3 18 2 23
Labor Share of Income 7 4 9 20
Clear
Skills Training Remove filter
These infrastructural and access constraints create unequal starting points that can amplify later disparities in labor-market preparedness.
Inference drawn from observed survey disparities in access, hands-on training, and preparedness; the study did not directly measure labor-market outcomes but links preparedness to potential labor-market effects in discussion.
low negative Exploring Student and Educator Challenges in AI Competency D... implied labor-market preparedness (not directly measured in this study)
Top-down AI guidance from institutions is common, while grassroots input from educators and students is often missing, which reduces policy relevance and uptake.
Survey items and thematic coding indicating the origin and participatory nature of institutional AI guidelines; comparative prevalence reported in open and closed responses.
low negative Exploring Student and Educator Challenges in AI Competency D... degree of grassroots input or participatory design in institutional AI policy fo...
Organizational compliance, governance, and transaction costs shape which AI uses are feasible, producing heterogeneity in adoption across firms; trust and accountability frictions can slow adoption even when productivity gains exist.
Workshop participants (n=15) reported compliance and governance considerations; authors infer broader organizational heterogeneity and friction effects from these qualitative data.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... adoption heterogeneity across firms; adoption speed/timing affected by governanc...
Designers’ expressed concerns about skill development suggest potential long-term effects on human capital accumulation; adoption that reduces learning opportunities could lower future wages or employability.
Participants' concerns captured in qualitative workshops (n=15); claim is an extrapolation to labor-market outcomes rather than direct measurement in the study.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... human capital accumulation; future wages; employability (hypothesized)
Legacy systems and siloed incentives create switching frictions that slow diffusion of AI-enabled ISP; early adopters may achieve sustained cost and service advantages and vendors bundling technology with change management could capture large rents.
Authors' argument informed by case observations of switching costs and vendor roles; no causal market-level evidence provided.
low negative Optimizing integrated supply planning in logistics: Bridging... adoption rate, market concentration, vendor rents
Returns to AI investments may exhibit increasing returns to scale, reinforcing winner‑take‑most dynamics unless offset by platformization or open‑source diffusion.
Economic scenario reasoning on capital intensity and platform effects; no empirical calibration or econometric evidence provided.
low negative How AI Will Transform the Daily Life of a Techie within 5 Ye... return on AI investment by firm size (evidence of increasing returns to scale) a...
Job insecurity rises when FDI is short‑term, footloose, or concentrated in capital‑intensive extractive projects.
Conceptual arguments and empirical examples in the review linking investment temporariness and capital intensity to higher job instability; empirical evidence less comprehensive and context-specific.
low negative Foreign Direct Investment, Labor Markets, and Income Distrib... job security, job tenure, employment volatility
Legal liability and cyber-insurance markets will need to adapt as machine-generated code becomes pervasive, with pricing internalizing risk from inadequate verification processes.
Speculative legal/economic implication discussed in the paper; no actuarial or legal-case data provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... insurance pricing changes; liability claims tied to machine-generated code
Individual developers or firms may underinvest in verification because defect accumulation imposes external costs on downstream actors, creating market failures that can justify standards, certifications, or regulation mandating interlocks or minimum verification practices.
Policy and market-failure argument based on externalities presented conceptually; no modeling or empirical evidence of such externalities provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... degree of underinvestment in verification; incidence of downstream costs/externa...
Short-run productivity gains from generative AI may be offset by longer-run increases in maintenance, security breaches, and reliability costs if verification lags.
Economic reasoning and forward-looking implications discussed in the paper; no empirical cost-benefit or longitudinal data presented.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... net productivity over time; maintenance/security costs versus short-term product...
Small, unverified errors, insecure patterns, and brittle interactions accumulate over time (latent accumulation), increasing operational fragility and long-run maintenance costs.
Theoretical argument and illustrative examples in the paper; no longitudinal defect accumulation studies or empirical cost analysis provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... rate of latent defect accumulation; long-run maintenance and reliability costs
Time pressure and productivity incentives lead developers to accept plausible AI outputs without full validation, a behavioral/institutional failure mode called the 'micro-coercion of speed' that effectively reverses the burden of proof.
Behavioral diagnosis and incentive analysis presented conceptually in the paper; no behavioral experiments, surveys, or observational data reported.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... developer acceptance rate of AI outputs without full validation / shift in burde...
Hallucination and error risk introduce potential liabilities in client engagements and may change contracting, insurance, and pricing practices in consulting services.
Derived from practitioner concerns reported in interviews and authors' normative discussion; no contractual or insurance-market data presented.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... liability exposure; contracting/insurance practices; pricing adjustments
Effective deployment requires governance, verification processes, and liability management to manage hallucination risk, creating adoption costs that may advantage larger firms and affect market concentration and pricing power.
Argument based on interviews about necessary organizational safeguards and the resource requirements to implement them; speculative market-structure implications are not empirically tested in the paper.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... adoption costs; firm-level resource burden; changes in market concentration/pric...
Widespread GenAI use may accelerate skill obsolescence for routine competencies and increase the premium on monitoring, critical evaluation, and AI‑integration skills, shifting investment toward retraining and upskilling.
Projection based on qualitative interviews and the authors' economic interpretation of TGAIF; no longitudinal or wage/skill data provided.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... skill obsolescence rates; demand for monitoring/evaluation/AI-integration skills...
Economic rents and advantages may accrue to agents who control large datasets, computing resources, and organizational processes that effectively integrate AI as a co-pilot, potentially increasing market concentration among AI providers.
Economic theory on scale economies and platform effects combined with observed industry patterns; reviewed literature provides conceptual arguments and case examples rather than broad empirical market-structure measurement.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... market concentration measures; returns to data/compute ownership (not fully meas...
Generative AI poses substitution risk for entry-level or routine cognitive work focused on generation or drafting without evaluative responsibility.
Task-based analyses and case studies indicating automation potential for routine generation tasks; empirical demonstrations of AI-produced drafts/outputs that could replace such work, but longer-run displacement evidence is limited.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... task automatability; employment/demand for routine-generation roles (largely unm...
There is a risk of deskilling through excessive reliance on AI, implying a need for continuous training and certification to preserve human judgment.
Qualitative interview evidence and observed concerns about overreliance; authors recommend training/governance based on identified risks; no direct longitudinal measurement of deskilling provided in summary.
low negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... human skill levels (deskilling risk); need for training/certification
Recommendation algorithms and widespread automated advice can induce herding or increase common exposures across retail investor portfolios, with potential macroprudential implications.
Theoretical discussion supported by examples from retail trading episodes and algorithmic amplification literature referenced in the review (conceptual and anecdotal evidence; limited systematic empirical quantification).
low negative Women's Investment Behaviour and Technology: Exploring the I... portfolio correlation across users, asset demand concentration, market volatilit...
Exposure to AI and platform work produces psychosocial effects for workers, including increased job insecurity, stress, and changing task content in surviving occupations.
Surveys, qualitative case studies, and workplace studies summarized in the review reporting worker‑reported insecurity and stress; the review also highlights inconsistent measurement and limited systematic evidence on psychosocial outcomes.
low negative The Impact of AI Machine Learning on Human Labor in the Work... job insecurity, stress, psychosocial wellbeing, and perceived changes in task co...
Delegation of oversight and reallocation of monitoring tasks due to AI integration changes transaction costs and affects organizational design and governance needs (e.g., more verification/audit effort or specialist oversight roles).
Based on participants' reported shifts in who performed monitoring/oversight tasks in the 40 interviews and the authors' interpretation of those shifts in organizational/economic terms.
low neutral AI in project teams: how trust calibration reconfigures team... transaction/monitoring costs and governance arrangements
This study represents the first attempt to conduct a comprehensive evaluation of artificial intelligence (AI) and its influence on job displacement based on the existing body of literature.
Author assertion in the paper; the excerpt provides no external verification (no citation of prior reviews/meta-analyses to justify the 'first attempt' claim).
low null result A Study on Work-Life Balance of Women Employees in the IT Se... comprehensiveness of literature-based evaluation of AI's influence on job displa...
We currently lack an understanding of how political parties perceive the potential impact AI has on employment, the role of regulations in protecting workers from AI-related job losses, and the importance of AI educational and training programs.
Statement of a literature/knowledge gap motivating the study (assertion by the authors; no empirical basis provided in the excerpt).
low null result Political Ideology, Artificial Intelligence (AI), and Labor ... existing knowledge about political party perceptions of AI's impact on employmen...
This study is the first systematic presentation of factual data describing employment outcomes of Russian university AI graduates.
Authors' stated novelty claim in the paper (asserted uniqueness of systematic institutional-level employment outcome data for Russian AI graduates).
low null result Employment og Graduates of Educational Programs in the Field... Novelty / uniqueness of compiled institutional-level dataset on employment outco...
Expect rising demand and wage premia for managers with hybrid capabilities (systems thinking + computational literacy), with a risk of widening returns to managerial skill heterogeneity.
Theoretical implication from predicted complementarities and task reallocation; prescriptive economic inference without empirical labor-market evidence in the paper.
low positive Comparative analysis of strategic vs. computational thinking... labor demand, wage premia, and distributional widening across managerial skill t...
Managers’ time will be reallocated toward hybrid tasks (interpretation, oversight, ethical deliberation), increasing returns to combined strategic and computational skills.
Predictive inference from the role reconfiguration analysis and task-complementarity argument; forward-looking theoretical forecast (no empirical time-use data).
low positive Comparative analysis of strategic vs. computational thinking... managerial time allocation (share devoted to hybrid tasks) and returns/wage prem...
Standards for provenance, labeling of AI-generated content, and interoperable evidence formats would lower verification costs and create beneficial network effects.
Policy recommendation derived from identified verification frictions and the study's analysis of data/model governance needs.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... verification cost and interoperability/network effects
There is growing market demand for AI-assisted fact-checking tools, creating opportunities for software, monitoring services, and labeled datasets.
Analytic implication drawn from findings about increasing AI use and needs for automation/labeling; based on interviews and market inference in the study.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... market demand for AI tools and labeled datasets
Hybrid agency implies complementarity between GenAI and managerial/knowledge‑worker skills (curation, evaluation, coordination), potentially increasing returns to those skills while automating routine cognitive tasks—consistent with skill‑biased technological change.
Synthesis of recurring themes linking GenAI capabilities with managerial skill topics in the thematic clusters; positioned as an implication for labour demand and skill composition rather than an empirically tested effect.
low positive Generative AI and the algorithmic workplace: a bibliometric ... expected changes in returns to managerial/knowledge‑worker skills and automation...
Humans who configure and teach agents gain understanding and skills themselves — learning-by-teaching generates human capital accumulation endogenous to agent deployment (bidirectional scaffolding).
Qualitative, naturalistic observations and comparative documentation of users configuring/teaching agents during the one-month study; no randomized assignment or pre/post quantitative skill testing reported.
low positive When Openclaw Agents Learn from Each Other: Insights from Em... human skill accumulation / understanding from configuring/teaching agents
Collaborative VR features can change team workflows (remote, synchronous inspection sessions), potentially lowering coordination costs across geographically distributed teams.
Paper lists collaborative multi-user sessions as a planned capability and posits organizational effects; no user studies or measurements of coordination cost savings presented.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... coordination costs / team workflow efficiency in distributed teams
Public funding for shared VR-capable data-exploration infrastructure could yield high leverage by improving returns on large observational investments.
Policy recommendation deriving from the platform and ROI arguments in the paper; no cost-benefit analysis or quantified ROI provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... policy leverage (ROI) from funding shared VR infrastructure
Using iDaVIE increases the usable fraction of large observational datasets by improving QC and annotation throughput, thereby raising returns to telescope investments and downstream AI efforts.
This is an inferred implication in the paper (returns-to-scale/platform effects) based on improved QC/annotation throughput; no empirical measurement of usable-fraction increases provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... usable fraction of observational datasets and downstream value for AI/modeling
Higher-quality labels produced via immersive inspection can reduce label noise and lower required training-data sizes for a target ML performance level.
Paper presents this as an implication/expected outcome based on improved annotation quality from immersive inspection; no empirical ML training experiments or quantitative reductions reported.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... label noise level and required training-data size for target model performance
iDaVIE demonstrably reduces cognitive load for multidimensional-data tasks compared with 2D-slice inspection.
Paper asserts reduced cognitive load and faster, more intuitive exploration as an aim and reported outcome; no formal user-study metrics, sample size, or statistical analysis provided.
low positive iDaVIE v1.0: A virtual reality tool for interactive analysis... cognitive load (mental effort) for multidimensional-data inspection
Tools that improve detection or quantification may reduce downstream costs from missed diagnoses or unnecessary follow-ups, improving cost-effectiveness in some scenarios.
Economic modeling and limited observational analyses that extrapolate diagnostic improvements to downstream resource use; direct empirical cost-effectiveness studies are scarce.
low positive Human-AI interaction and collaboration in radiology: from co... downstream healthcare utilization (additional tests, treatments), cost per diagn...
The metacognitive reliability metric can reduce adoption risk for purchasers by providing transparent error-risk assessments and enabling performance-based autonomy thresholds.
Conceptual claim supported by the existence of an empirical confidence metric from the recursive meta-model and discussion of procurement/decision-making implications; not empirically tested with purchasers or procurement outcomes.
low positive Human Autonomy Teaming and AI Metacognition in Maritime Thre... adoption risk (qualitative or procurement decision proxies)
HACL/CS supports human trust and situational awareness.
Human factors measured with trust and situational awareness questionnaires in the simulation; summary reports supportive effects on trust and situational awareness but lacks sample-size/statistical detail.
low positive Human Autonomy Teaming and AI Metacognition in Maritime Thre... self-reported trust and situational awareness scores
Organizational norms and UX influence adoption rates and diffusion of AI: social calibration processes at the team level matter for adoption beyond individual cost–benefit calculations.
Reported by interviewees (N=40) as factors shaping whether and how teams incorporated AI into routines; integrated into theoretical implications for diffusion modeling.
low positive AI in project teams: how trust calibration reconfigures team... AI adoption/diffusion rates at team/organization level
Well-calibrated trust tends to encourage AI being used as a complement to human labor (augmentation), increasing effective productivity; miscalibration (over- or under-trust) can lead to productivity losses.
Inferential claim drawn from interviewees' accounts of when teams appropriately relied on AI (augmentation) versus when inappropriate reliance or avoidance occurred; supported by thematic interpretation rather than quantitative measurement.
low positive AI in project teams: how trust calibration reconfigures team... productive use of AI (complementarity vs substitution) and effective productivit...
Policymakers should support standards for auditability, human‑in‑the‑loop thresholds and training subsidies to reduce coordination failures and make the social benefits of AI adoption more widely shared.
Normative policy recommendation derived from the paper’s analysis of risks, governance needs and distributional concerns; not empirically validated within the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... adoption of standards; breadth of social benefits; coordination failure reductio...
Organisations will invest more in training for AI‑related sensemaking, trust calibration and governance competencies; returns to such training should be evaluated relative to investments in model quality.
Prescriptive inference from the framework and human‑capital theory; supported by referenced literature but not empirically tested in this paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... training investment levels; returns on training; comparative returns vs model in...
Explicit comparative‑advantage allocation will shift the composition of tasks across humans and AI, altering demand for routine versus non‑routine skills and potentially increasing demand for high‑level judgement, oversight and sensemaking skills.
Projected labour‑market implication based on theoretical reasoning and prior literature on task‑based skill demand; not empirically estimated in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... task composition; demand for routine vs non‑routine skills; demand for oversight...
Operationalising the four symbiarchic practices through updated HR systems lets firms capture AI‑enabled productivity gains without eroding trust, ethics or employee well‑being.
Normative claim based on theoretical synthesis and managerial prescription; no empirical testing or field data presented in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... AI‑enabled productivity gains; employee trust; ethical outcomes; employee well‑b...
Public data sharing, reproducibility standards, and shared benchmarks could raise the floor of AI utility across the industry.
Policy implication grounded in arguments about data quality, coverage, and generalizability from the narrative review; speculative recommendation rather than evidence-backed empirical claim.
low positive Learning from the successes and failures of early artificial... baseline AI performance/utility across firms (industry-wide)
There is potential for consolidation as firms acquire data, talent, or validated AI-driven assets.
Industry-structure implication drawn from economics of complementary assets and observed M&A activity patterns; presented as a likely trend rather than demonstrated empirically in the paper.
low positive Learning from the successes and failures of early artificial... M&A activity targeting AI capabilities, data assets, or relevant talent
AI startups that demonstrate validated, reproducible wet-lab outcomes and access to high-quality data are more likely to command premium valuations.
Argument from observed market behavior and economics of complementary assets presented in the narrative; no systematic valuation analysis included.
low positive Learning from the successes and failures of early artificial... startup valuation premium tied to validated wet-lab results and data access
Investors should recalibrate expectations: greater value accrues to firms that integrate AI with experimental pipelines and proprietary data assets rather than firms that only possess AI capability.
Economics-focused implications drawn from thematic analysis of heterogeneity in firm outcomes and integration requirements; market-practice inference rather than empirical valuation study.
low positive Learning from the successes and failures of early artificial... firm valuation / investor returns conditional on AI integration and data assets
AI tools complement sensory expertise and design thinking, shifting skill demand toward interdisciplinary competencies (e.g., computational rheology, psychophysics, cultural analytics).
Reasoned inference from technology literature and skill-complementarity theory; literature synthesis but no labor-market empirical analysis provided.
low positive At the table with Wittgenstein: How language shapes taste an... demand for interdisciplinary skills in food R&D and complementarity between AI t...
The research establishes the theory of performance management by developing operational measurement solutions for companies going through workplace redesign due to AI.
Authors claim theoretical contribution and provision of operational measurement solutions based on the proposed three-dimensional model and the empirical patterns observed in the 2022–2024 LinkedIn and Indeed datasets; no external validation or implementation evidence reported in the summary.
low positive Reconstruction of knowledge worker performance evaluation sy... operational performance-measurement solutions and theoretical framing for perfor...