The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (2608 claims)

Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 609 159 77 736 1615
Governance & Regulation 664 329 160 99 1273
Organizational Efficiency 624 143 105 70 949
Technology Adoption Rate 502 176 98 78 861
Research Productivity 348 109 48 322 836
Output Quality 391 120 44 40 595
Firm Productivity 385 46 85 17 539
Decision Quality 275 143 62 34 521
AI Safety & Ethics 183 241 59 30 517
Market Structure 152 154 109 20 440
Task Allocation 158 50 56 26 295
Innovation Output 178 23 38 17 257
Skill Acquisition 137 52 50 13 252
Fiscal & Macroeconomic 120 64 38 23 252
Employment Level 93 46 96 12 249
Firm Revenue 130 43 26 3 202
Consumer Welfare 99 51 40 11 201
Inequality Measures 36 105 40 6 187
Task Completion Time 134 18 6 5 163
Worker Satisfaction 79 54 16 11 160
Error Rate 64 78 8 1 151
Regulatory Compliance 69 64 14 3 150
Training Effectiveness 81 15 13 18 129
Wages & Compensation 70 25 22 6 123
Team Performance 74 16 21 9 121
Automation Exposure 41 48 19 9 120
Job Displacement 11 71 16 1 99
Developer Productivity 71 14 9 3 98
Hiring & Recruitment 49 7 8 3 67
Social Protection 26 14 8 2 50
Creative Output 26 14 6 2 49
Skill Obsolescence 5 37 5 1 48
Labor Share of Income 12 13 12 37
Worker Turnover 11 12 3 26
Industry 1 1
Clear
Skills Training Remove filter
Artificial intelligence and robotic technologies are fundamentally reshaping labour markets and pose multifaceted challenges to workers engaged in routine and low-skilled tasks.
Narrative review of domestic and international scholarly literature over the past decade (literature review / synthesis).
high negative From Technological Substitution to Institutional Response: A... risks to routine and low-skilled workers (labor market disruption / challenges)
Structural barriers, workforce biases, and digital skill gaps affect women’s participation in AI-enabled sectors.
Claim derived from the paper's synthesis of literature (peer-reviewed studies, policy analyses, preprints) identifying common barriers; the abstract does not report quantitative meta-analysis or specific sample sizes.
high negative Artificial Intelligence and GenderedEmployment: Reviewing Op... drivers of women's participation in AI-enabled sectors (barriers and gaps)
Routine-intensive sectors exhibit higher susceptibility to automation.
Synthesis result reported in the paper based on the systematic review of sector-specific literature (no numeric aggregation or sample size provided in the abstract).
high negative AI and the Future of Job Profiles: A systematic Review of Se... susceptibility to automation
Scalable AI tutoring for procedural skill learning requires structured knowledge representations, yet constructing these representations remains a labor-intensive bottleneck.
Background/claim made in the paper's introduction framing the problem; no specific quantitative evidence reported in the abstract.
high negative Developing Models of Procedural Skills using an AI-assisted ... effort required to construct structured knowledge representations
The study is framed based on Job Demands-Resources (JD-R) theory, positing that HAI-C task complexity is a job demand and AI self-efficacy/humble leadership act as resources that can mitigate negative effects on engagement.
Introduction states JD-R theory as the theoretical basis and describes job demands (HAI-C task complexity) and job/personal resources (humble leadership, AI self-efficacy) in the hypothesized model.
high negative How does human-AI collaboration task complexity affect emplo... theoretical framing / hypothesized relationships
HAI-C tech-learning anxiety reduces employees' work engagement (serves as the mediator between HAI-C task complexity and work engagement).
Mediation analysis via hierarchical regression and bootstrapping on the three-wave survey sample of 497 employees; reported in Results as the mediating mechanism.
Human-AI collaboration task complexity (HAI-C task complexity) negatively affects employees' work engagement by amplifying their HAI-C tech-learning anxiety.
Three-wave longitudinal survey of matched data from 497 employees; mediation analysis using hierarchical regression and bootstrapping reported in the Results section.
Users push back against agent outputs -- through corrections, failure reports, and interruptions -- in 44% of all turns.
Turn-level coding of user behavior in the SWE-chat dataset: proportion of conversational turns containing correction/complaint/interrupt signals, computed across >63,000 user prompts and sessions.
high negative SWE-chat: Coding Agent Interactions From Real Users in the W... rate of user pushback per interaction turn
Agent-written code introduces more security vulnerabilities than code authored by humans.
Comparative analysis of security vulnerabilities attributed to agent-authored code versus human-authored code within the SWE-chat dataset (method details not specified in excerpt).
high negative SWE-chat: Coding Agent Interactions From Real Users in the W... security vulnerabilities introduced by agent-written code versus human-written c...
Just 44% of all agent-produced code survives into user commits.
Empirical measurement of code provenance and survival within the SWE-chat dataset: proportion of agent-produced code that becomes part of subsequent user commits across sessions.
high negative SWE-chat: Coding Agent Interactions From Real Users in the W... survival/usefulness of agent-produced code (proportion incorporated into commits...
Despite rapidly improving capabilities, coding agents remain inefficient in natural settings.
Authors' summary claim supported by dataset-derived metrics such as agent code survival rate (44%) and user pushback (44% of turns); observational analysis of SWE-chat.
high negative SWE-chat: Coding Agent Interactions From Real Users in the W... overall agent efficiency in natural developer workflows (qualitative synthesis)
The policy and research challenge posed by platform-mediated automation is not merely job quantity (technological unemployment) but institutional continuity — how societies reproduce practical competence when platforms optimize for efficiency rather than formation.
Normative and conceptual claim developed through literature synthesis (institutional economics, platform governance, workforce development); presented as an analytical reframing rather than an empirically tested hypothesis.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... institutional continuity and human capital reproduction (quality of workforce fo...
Entry-level roles have historically functioned as apprenticeships in which workers acquire tacit knowledge and critical judgment; if platforms curtail these formative occupational layers, organizations may lack future workers capable of exercising contextual reasoning required to manage complex systems.
Institutional economics and workforce development literature cited in the paper; conceptual synthesis without original empirical measurement reported.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... human capital formation (tacit knowledge acquisition and contextual reasoning ca...
Platform-mediated automation risks hollowing out labor structures from both directions: eroding repetitive, junior roles from below and automating supervisory coordination functions from above.
Theoretical argument synthesizing institutional economics and platform literature; articulated as a conceptual risk rather than demonstrated with original empirical data.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... structural change in occupational layers (hollowing out of junior and supervisor...
Algorithmic systems are displacing routine tasks across both low-wage entry-level work and middle-management functions.
Stated in paper's argumentation; supported by a literature-based review drawing on platform governance literature and recent research on AI-enhanced automation (no original empirical sample or quantitative study reported).
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... displacement of routine tasks (across entry-level and middle-management roles)
As multimodal AI achieves human-parity understanding of speech and gesture, [the keyboard's] necessity dissolves.
Theoretical claim supported by multidisciplinary review (history, neuroscience, technology, organizational studies); no quantified empirical test reported.
high negative The Instrumental Dissolution of Typing: Why AI Challenges th... necessity/usage of keyboard as default input
There was a nonsignificant absolute retest performance reduction in the AI condition and a larger retest performance decrement in the AI condition (i.e., retention decreased more after using Copilot).
Comparison of retest (one-week) performance across conditions reported in results; authors report a nonsignificant reduction and larger decrement for the AI/Copilot condition (n=22).
high negative Fast and Forgettable: A Controlled Study of Novices' Perform... retest performance (learning retention) after one week
Thin training coverage fosters anxiety about substitution and slows diffusion of AI tools.
Reported associations from surveys of mid-level managers and technical staff, interviews, and document analysis across cases; thematic coding identified links between limited training, worker anxiety, and slower diffusion. (Sample size not reported.)
high negative Overcoming Resistance to Change: Artificial Intelligence in ... worker anxiety and speed of diffusion/adoption
Agency in software engineering is primarily constrained by organizational policies rather than individual preferences.
Authors' synthesis of qualitative results across the ACTA/Delphi and task/review phases indicating organizational policy factors were cited as primary constraints.
high negative From Junior to Senior: Allocating Agency and Navigating Prof... Primary source of constraint on developer agency (organizational policy vs indiv...
Underreliance on AI might deprive software developers of potential gains in productivity and quality.
Stated in the paper and motivated by themes from twenty-two developer interviews indicating missed benefits when developers underuse LLM tools.
high negative Towards an Appropriate Level of Reliance on AI: A Preliminar... productivity and output quality
Overreliance on AI may lead to long-term negative consequences (e.g., atrophy of critical thinking skills).
Paper explicitly states this risk and grounds the discussion in findings from twenty-two developer interviews (qualitative evidence and participant-reported concerns).
high negative Towards an Appropriate Level of Reliance on AI: A Preliminar... atrophy of critical thinking skills / skill degradation
AI can exacerbate occupational polarization, digital exclusion, and discriminatory outcomes when models are trained on biased data or deployed without transparency and accountability.
Thematic synthesis across included studies identifying mechanisms (biased training data, lack of transparency/accountability) linked to negative distributional outcomes (occupational polarization, digital exclusion, discrimination).
high negative Artificial Intelligence in the Labor Market: Evidence on Wor... distributional and equity outcomes (polarization, exclusion, discrimination)
Small and medium-sized practices face challenges of skill gaps and resource constraints that hinder adoption of technology and data analytics.
Consistent findings across included studies highlighting barriers in small and medium-sized practices (SMPs).
high negative The Use of Technology and Data Analytics in Modern Auditing:... ability to adopt and implement technology/data analytics
Large language models remain confined to linguistic simulation rather than grounded understanding.
Conceptual assertion in the paper arguing limits of current models; no empirical tests or measurements reported.
high negative Governing Reflective Human-AI Collaboration: A Framework for... grounded_understanding (absence thereof)
Human decision makers may fail to execute optimal follow-up actions, potentially reducing overall performance.
Motivating argument in the paper (conceptual observation about human suboptimal policies in sequential decision-making).
high negative Improving Human Performance with Value-Aware Interventions: ... overall decision-making performance (expected return/value)
The opacity, fluency, and low-friction interaction patterns of LLMs obscure the boundary between human and machine contribution, leading users to infer competence from outputs rather than from the processes that generate them.
Theoretical argument grounded in prior literature on automation bias and cognitive offloading; presented as explanatory mechanism in the paper rather than an empirically tested causal estimate.
high negative The LLM Fallacy: Misattribution in AI-Assisted Cognitive Wor... user inference of competence (output-based vs process-based attribution)
The paper introduces the 'LLM fallacy,' a cognitive attribution error in which individuals misinterpret LLM-assisted outputs as evidence of their own independent competence, producing a systematic divergence between perceived and actual capability.
Conceptual/theoretical claim and formal definition offered in the paper; no empirical validation reported in the abstract.
high negative The LLM Fallacy: Misattribution in AI-Assisted Cognitive Wor... divergence between perceived competence and actual competence when using LLM out...
Infrastructure constraints, particularly in developing countries, limit AI adoption in auditing.
Thematic analysis of reviewed articles noting infrastructure limitations (e.g., ICT infrastructure) in developing-country contexts.
high negative Implementing Artificial Intelligence in Auditing: A Systemat... infrastructure constraints affecting AI adoption
Limitations in auditor competencies (skills and training) hinder effective AI adoption in auditing.
Thematic findings across the sample of articles report auditor competency gaps as a challenge to AI implementation.
high negative Implementing Artificial Intelligence in Auditing: A Systemat... auditor competencies / skill gaps
Ethical and data privacy concerns are persistent challenges to AI implementation in auditing.
Recurring theme in the reviewed literature identified via thematic analysis; papers cite ethics and privacy as obstacles.
high negative Implementing Artificial Intelligence in Auditing: A Systemat... ethical and data privacy concerns as barriers
Several challenges persist for AI adoption in auditing, including high technology investment costs.
Thematic analysis of barriers reported across the 15 articles highlighting cost as a recurrent challenge.
high negative Implementing Artificial Intelligence in Auditing: A Systemat... barrier: technology investment costs to AI adoption
Asymptomatic effects of AI use evolved into chronic harms such as skill atrophy and identity commoditization among workers.
Reported longitudinal findings from the study indicating progression from asymptomatic (subtle) effects to chronic harms; abstract lists harms but provides no quantification or sample details.
high negative From Future of Work to Future of Workers: Addressing Asympto... skill atrophy and worker identity commoditization
Initial operational gains from AI use masked a phenomenon called 'intuition rust' — a gradual dulling of expert judgment.
Empirical observation reported from the year-long longitudinal study of cancer specialists (phenomenon named and described; abstract provides no quantitative measures or sample size).
high negative From Future of Work to Future of Workers: Addressing Asympto... expert judgment (intuition/clinical reasoning)
Low-skill roles in packaging, sorting, and basic assembly face a high risk of automation.
Paper's findings/prediction derived from task-level classification (routine/repetitive tasks) applied to jobs in Nagpur's medium enterprises; no reported sample size or quantified risk metrics in the excerpt.
high negative PREDICTING THE FUTURE OF JOBS IN NAGPUR DISTRICT MIDC: THE R... risk of automation for specific low-skill job categories (packaging, sorting, ba...
The study's findings are subject to design limitations including an AM/PM session confound, differential attrition, and LLM grading sensitivity to document length.
Authors' reported limitations section citing specific threats to internal validity and measurement (session timing confound, differential attrition across conditions, and grading biases of the LLM used to evaluate documents).
high negative Scaffolding Human-AI Collaboration: A Field Experiment on Be... threats to validity (confounds and measurement sensitivity)
The behavioral scaffolding intervention was associated with substantially lower document production.
Same field experiment (N=388); the behavioral scaffolding required joint AI use within pairs and was compared to unstructured use, with reported reductions in document production in the behavioral condition.
high negative Scaffolding Human-AI Collaboration: A Field Experiment on Be... document production (quantity of documents produced)
A behavioral scaffolding intervention (a structured protocol requiring joint AI use within pairs) was associated with lower document quality relative to unstructured use.
Field experiment with 388 employees at a Fortune 500 retailer; random/experimental assignment to scaffolding conditions while all participants had access to the same AI tool; comparison reported between behavioral scaffolding condition and unstructured use.
Latent-outcome estimation faces a within-study noncomparability challenge: different indicators within a study may have different and possibly nonlinear relationships with the same latent outcome, making them not directly comparable.
Theoretical exposition in the paper describing heterogenous indicator-to-latent mappings and potential nonlinearity; illustrated with examples (no empirical sample size).
high negative Nonparametric Identification and Estimation of Causal Effect... comparability of different indicators for the same latent outcome within a study
Latent-outcome estimation faces a cross-study noncomparability challenge: different measurement systems across studies may cause estimators to target different empirical quantities even when the underlying latent treatment effect is the same.
Conceptual and theoretical argumentation in the paper describing identification issues across studies due to differing measurement systems; supported by examples and discussion (no empirical sample size).
high negative Nonparametric Identification and Estimation of Causal Effect... comparability of estimated latent treatment effects across studies
Rote learning will become obsolete in favor of contextual application.
Paper's forward-looking prediction based on synthesis of adult learning theory and workforce development literature; no empirical sample size or quantified trend data provided.
high negative The Future of Education in an AI-Driven World: Preparing Org... decline/obsolescence of rote learning and increase in contextual application
These advancements have raised concerns regarding workforce redundancy, particularly for routine and low-skilled jobs.
Synthesis of concerns documented in the reviewed literature and observed sectoral trends (literature review; qualitative synthesis).
high negative IMPACT OF ARTIFICIAL INTELLIGENCE ON EMPLOYMENT IN THE COMME... risk of worker displacement in routine and low-skilled roles
Foundation-model usage can increase compute-related emissions.
Conceptual/environmental concern highlighted in the paper about the carbon footprint of heavy model use and persistent storage; no quantified emissions analysis or lifecycle assessment presented.
high negative Remote-Capable Knowledge Work Should Default to AI-Enabled F... compute-related (carbon) emissions associated with foundation-model usage
These systems can cause skill atrophy.
Theoretical risk articulated in the paper that reliance on AI assistance may degrade human skills over time; no longitudinal skill-measurement or experimental evidence provided.
high negative Remote-Capable Knowledge Work Should Default to AI-Enabled F... degradation or atrophy of worker skills
The same foundation-model systems can also intensify surveillance.
Cautionary claim in the paper noting the surveillance risk of durable, queryable traces and integrated tooling; presented as a conceptual risk rather than empirically measured increase in surveillance.
high negative Remote-Capable Knowledge Work Should Default to AI-Enabled F... increase in workplace surveillance capability/use
Job insecurity emerges as a critical mediating factor influencing employee attitudes and behavioural responses to generative AI, including upskilling intentions and resistance to technological change.
Review-level synthesis identifying job insecurity reported in included studies as mediating relationships between AI adoption and employee attitudes/behaviours (e.g., upskilling, resistance).
high negative Generative AI in the Workplace: A Systematic Review of Produ... upskilling intentions and resistance to technological change (mediated by job in...
Employees express concerns about role displacement (job loss or role changes) associated with generative AI adoption.
Reported across multiple studies included in the review; the review summarises these concerns as part of mixed employee perceptions.
high negative Generative AI in the Workplace: A Systematic Review of Produ... perceived risk of role displacement / job loss
These positive perceptions coexist with employee concerns about skill obsolescence related to generative AI.
Synthesis of studies included in the review documenting worker concerns about skills becoming obsolete due to AI-driven changes.
high negative Generative AI in the Workplace: A Systematic Review of Produ... concerns about skill obsolescence
The explanatory interface suppresses the natural development of both cognitive trust and emotional trust.
Longitudinal/within-experiment measures of cognitive and emotional trust reported in the RCT; authors state that explanatory interface suppressed the natural development of these trust dimensions in the 120-participant experiment.
high negative How AI-Assisted Decision-Making Paradigms and Explainability... cognitive trust and emotional trust development
The explanatory interface exerts a negative effect on learned trust.
Randomized controlled experiment measuring learned trust; authors report a negative (statistically significant) effect of explanatory interface on learned trust in their sample of 120 pre-service teachers.
The improvement in task performance due to the explanatory interface is confined to the task execution stage and does not transfer to subsequent independent tasks.
Experimental measurement of immediate (during-assisted) task performance and subsequent independent task performance; authors report improvement only during task execution and no transfer effect to later independent tasks in their RCT with 120 participants.
high negative How AI-Assisted Decision-Making Paradigms and Explainability... performance transfer to subsequent independent tasks