The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (2954 claims)

Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 369 105 58 432 972
Governance & Regulation 365 171 113 54 713
Research Productivity 229 95 33 294 655
Organizational Efficiency 354 82 58 34 531
Technology Adoption Rate 277 115 63 27 486
Firm Productivity 273 33 68 10 389
AI Safety & Ethics 112 177 43 24 358
Output Quality 228 61 23 25 337
Market Structure 105 118 81 14 323
Decision Quality 154 68 33 17 275
Employment Level 68 32 74 8 184
Fiscal & Macroeconomic 74 52 32 21 183
Skill Acquisition 85 31 38 9 163
Firm Revenue 96 30 22 148
Innovation Output 100 11 20 11 143
Consumer Welfare 66 29 35 7 137
Regulatory Compliance 51 61 13 3 128
Inequality Measures 24 66 31 4 125
Task Allocation 64 6 28 6 104
Error Rate 42 47 6 95
Training Effectiveness 55 12 10 16 93
Worker Satisfaction 42 32 11 6 91
Task Completion Time 71 5 3 1 80
Wages & Compensation 38 13 19 4 74
Team Performance 41 8 15 7 72
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 17 15 9 5 46
Job Displacement 5 28 12 45
Social Protection 18 8 6 1 33
Developer Productivity 25 1 2 1 29
Worker Turnover 10 12 3 25
Creative Output 15 5 3 1 24
Skill Obsolescence 3 18 2 23
Labor Share of Income 7 4 9 20
Clear
Human Ai Collab Remove filter
The market for HR analytics platforms and tailored AI services is expanding, with potential for vendor lock-in effects and platform concentration.
Market implication synthesized in the review from literature noting growing demand for HR AI tools; largely inferential rather than empirically proven within the reviewed studies.
low mixed Data-Driven Strategies in Human Resource Management: The Rol... market size for HR AI tools, market concentration, lock-in indicators
Automation of administrative HR tasks may reduce demand for lower-skilled HR roles while increasing wages and demand for analytics-capable workers, contributing to within-firm wage reallocation.
Review implication synthesizing literature trends on automation and skill demand; not based on causal longitudinal evidence (review highlights evidence gaps).
low mixed Data-Driven Strategies in Human Resource Management: The Rol... employment levels by HR skill category, wage changes by skill
Heterogeneous adoption of data-driven HRM may widen productivity dispersion across firms and affect market competition.
Implication drawn in the review based on heterogeneous adoption patterns discussed in included studies and economic interpretation of productivity effects.
low mixed Data-Driven Strategies in Human Resource Management: The Rol... productivity dispersion across firms, market competition measures
Principal stratification analysis suggests the training’s effect on scores operated primarily by expanding the set of LLM users (an adoption channel) rather than substantially improving per-user productivity among those who would already use the LLM.
Mechanism decomposition using principal stratification applied to the randomized trial data (n = 164); analysis indicates a larger contribution from the adoption margin than from within-user productivity gains, though estimates have wide confidence intervals.
low mixed Training for Technology: Adoption and Productive Use of Gene... Mechanism components: adoption rate and per-user effectiveness (score conditiona...
Systemic risks from misaligned optimisation (narrow objectives, externalities) warrant oversight mechanisms (AI steering committees, escalation paths) and potentially sectoral regulation of decision-critical algorithms.
Policy-prescriptive claim based on conceptual identification of optimisation externalities and accountability gaps; no sectoral case studies or empirical risk quantification in the paper.
low negative Comparative analysis of strategic vs. computational thinking... systemic risk exposure and effectiveness of oversight/regulatory mechanisms
Measurement friction from the results-actionability gap creates a hidden cost: teams can detect problems but cannot cheaply translate findings into improvements, reducing the speed and ROI of LLM investments.
Authors' implication drawn from interview evidence about the effort required for remediation and lack of direct translation from evaluations to fixes; presented as an economic implication rather than directly measured quantity.
low negative Results-Actionability Gap: Understanding How Practitioners E... inferred effect on ROI and speed of product improvement
Risk of platform shutdown (platform mortality) shapes user behavior by reducing incentives to invest time/effort configuring agents, creating stranded-asset-like risks.
Qualitative observations and economic reasoning linking user reports/behaviors to perceived platform risk during the one-month observational period; no formal economic measurement or causal identification.
low negative When Openclaw Agents Learn from Each Other: Insights from Em... user investment in configuring agents / adoption incentives under platform shutd...
If verified, explainable GLAI is priced higher due to compliance costs, access-to-justice gaps may widen as lower-cost but riskier offerings persist or services become more expensive.
Distributional reasoning linking higher compliance costs to price increases and access effects; supported by illustrative examples, no empirical price or access data.
low negative Why Avoid Generative Legal AI Systems? Hallucination, Overre... access-to-justice metrics correlated with pricing of verified vs. unverified GLA...
Routine, unrestrained adoption of GLAI without enforceable mechanisms for effective human review threatens judicial independence and rights protections.
Normative and legal argumentation supported by conceptual analysis and illustrative scenarios. No empirical causal evidence; projection based on theoretical risk pathways.
low negative Why Avoid Generative Legal AI Systems? Hallucination, Overre... level of threat to judicial independence and protection of rights (institutional...
Insurers will price systemic-tail risks differently from routine failure risk, potentially increasing premiums for high-autonomy deployments or requiring minimum oversight modes for coverage.
Analytical argument about liability, risk pooling, and insurance practices; no empirical insurance-pricing data supplied.
low negative Resilience Meets Autonomy: Governing Embodied AI in Critical... insurance pricing and coverage conditions for high-autonomy deployments (premium...
There is a risk of deskilling, especially for trainees receiving reduced diagnostic practice when AI automates routine tasks.
Conceptual arguments supported by qualitative reports and limited observational findings; empirical longitudinal evidence quantifying deskilling is sparse.
low negative Human-AI interaction and collaboration in radiology: from co... trainee diagnostic performance over time, case exposure counts, measures of reta...
Erosion of informal communication and tacit coordination driven by AI integration can create negative externalities on team efficiency that are not captured by short-run metrics.
Derived from interview narratives describing loss of ad hoc communications and tacit knowledge exchange after AI adoption; interpreted as producing costs not reflected in immediate measurable outputs.
low negative AI in project teams: how trust calibration reconfigures team... team efficiency and unmeasured coordination/tacit work
Uneven adoption of symbiarchic HR practices across firms could concentrate productivity gains and rents in firms or occupations that successfully integrate AI while preserving human judgement, potentially widening within‑ and between‑firm inequality.
Projected distributional implication based on economic theory and the paper’s framework; presented as a hypothesis for empirical testing rather than as an observed result.
low negative Symbiarchic leadership: leading integrated human and AI cybe... within‑ and between‑firm inequality; distribution of productivity rents
Demanding oversight of multiple AI agents drives increased task-switching for workers.
Asserted in the paper as part of the mechanism linking AI use to cognitive overload, based on organizational observations and theory; no empirical task-switching frequency or time-use data provided in the excerpt.
low negative When AI Assistance Becomes Cognitive Overload: Understanding... task-switching frequency / oversight burden
Preliminary evidence that inappropriate reliance on AI outputs is worse for complex information needs (complex answers).
Post-hoc/stratified analysis in the user study examining the effect of the complexity of the information need on reliance/error-detection; described as preliminary in the paper.
low negative To Believe or Not To Believe: Comparing Supporting Informati... error-detection rate and reliance stratified by complexity of question/answer
More granular and auditable credentials may shift signaling dynamics and risk credential inflation; regulators should monitor credential proliferation and market value.
Conceptual warning in paper (theoretical); no empirical credential-market study included.
low negative Curriculum engineering: organisation, orientation, and manag... number and granularity of credentials issued, employer valuation of credentials,...
Overreliance on GenAI CDS may lead to deskilling of clinicians, eroding judgment over time and increasing systemic vulnerability.
The paper cites theoretical risk and references limited longitudinal concerns; empirical longitudinal studies demonstrating deskilling are scarce per the paper’s stated evidence gaps.
low negative GenAI and clinical decision making in general practice clinician diagnostic skill over time; reliance/override rates; error rates when ...
Organizational compliance, governance, and transaction costs shape which AI uses are feasible, producing heterogeneity in adoption across firms; trust and accountability frictions can slow adoption even when productivity gains exist.
Workshop participants (n=15) reported compliance and governance considerations; authors infer broader organizational heterogeneity and friction effects from these qualitative data.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... adoption heterogeneity across firms; adoption speed/timing affected by governanc...
Designers’ expressed concerns about skill development suggest potential long-term effects on human capital accumulation; adoption that reduces learning opportunities could lower future wages or employability.
Participants' concerns captured in qualitative workshops (n=15); claim is an extrapolation to labor-market outcomes rather than direct measurement in the study.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... human capital accumulation; future wages; employability (hypothesized)
Legacy systems and siloed incentives create switching frictions that slow diffusion of AI-enabled ISP; early adopters may achieve sustained cost and service advantages and vendors bundling technology with change management could capture large rents.
Authors' argument informed by case observations of switching costs and vendor roles; no causal market-level evidence provided.
low negative Optimizing integrated supply planning in logistics: Bridging... adoption rate, market concentration, vendor rents
Returns to AI investments may exhibit increasing returns to scale, reinforcing winner‑take‑most dynamics unless offset by platformization or open‑source diffusion.
Economic scenario reasoning on capital intensity and platform effects; no empirical calibration or econometric evidence provided.
low negative How AI Will Transform the Daily Life of a Techie within 5 Ye... return on AI investment by firm size (evidence of increasing returns to scale) a...
Legal liability and cyber-insurance markets will need to adapt as machine-generated code becomes pervasive, with pricing internalizing risk from inadequate verification processes.
Speculative legal/economic implication discussed in the paper; no actuarial or legal-case data provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... insurance pricing changes; liability claims tied to machine-generated code
Individual developers or firms may underinvest in verification because defect accumulation imposes external costs on downstream actors, creating market failures that can justify standards, certifications, or regulation mandating interlocks or minimum verification practices.
Policy and market-failure argument based on externalities presented conceptually; no modeling or empirical evidence of such externalities provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... degree of underinvestment in verification; incidence of downstream costs/externa...
Short-run productivity gains from generative AI may be offset by longer-run increases in maintenance, security breaches, and reliability costs if verification lags.
Economic reasoning and forward-looking implications discussed in the paper; no empirical cost-benefit or longitudinal data presented.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... net productivity over time; maintenance/security costs versus short-term product...
Small, unverified errors, insecure patterns, and brittle interactions accumulate over time (latent accumulation), increasing operational fragility and long-run maintenance costs.
Theoretical argument and illustrative examples in the paper; no longitudinal defect accumulation studies or empirical cost analysis provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... rate of latent defect accumulation; long-run maintenance and reliability costs
Time pressure and productivity incentives lead developers to accept plausible AI outputs without full validation, a behavioral/institutional failure mode called the 'micro-coercion of speed' that effectively reverses the burden of proof.
Behavioral diagnosis and incentive analysis presented conceptually in the paper; no behavioral experiments, surveys, or observational data reported.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... developer acceptance rate of AI outputs without full validation / shift in burde...
Hallucination and error risk introduce potential liabilities in client engagements and may change contracting, insurance, and pricing practices in consulting services.
Derived from practitioner concerns reported in interviews and authors' normative discussion; no contractual or insurance-market data presented.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... liability exposure; contracting/insurance practices; pricing adjustments
Effective deployment requires governance, verification processes, and liability management to manage hallucination risk, creating adoption costs that may advantage larger firms and affect market concentration and pricing power.
Argument based on interviews about necessary organizational safeguards and the resource requirements to implement them; speculative market-structure implications are not empirically tested in the paper.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... adoption costs; firm-level resource burden; changes in market concentration/pric...
Widespread GenAI use may accelerate skill obsolescence for routine competencies and increase the premium on monitoring, critical evaluation, and AI‑integration skills, shifting investment toward retraining and upskilling.
Projection based on qualitative interviews and the authors' economic interpretation of TGAIF; no longitudinal or wage/skill data provided.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... skill obsolescence rates; demand for monitoring/evaluation/AI-integration skills...
Uncertainty about long-run agentic behavior increases option value and downside risk of investing in agentic systems, which may raise discount rates and required returns.
Economic argument applying risk/return logic to agentic uncertainty; no quantitative empirical evidence provided.
low negative Visioning Human-Agentic AI Teaming: Continuity, Tension, and... investment valuation metrics (discount rates, required returns) for agentic syst...
Economic rents and advantages may accrue to agents who control large datasets, computing resources, and organizational processes that effectively integrate AI as a co-pilot, potentially increasing market concentration among AI providers.
Economic theory on scale economies and platform effects combined with observed industry patterns; reviewed literature provides conceptual arguments and case examples rather than broad empirical market-structure measurement.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... market concentration measures; returns to data/compute ownership (not fully meas...
Generative AI poses substitution risk for entry-level or routine cognitive work focused on generation or drafting without evaluative responsibility.
Task-based analyses and case studies indicating automation potential for routine generation tasks; empirical demonstrations of AI-produced drafts/outputs that could replace such work, but longer-run displacement evidence is limited.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... task automatability; employment/demand for routine-generation roles (largely unm...
Upfront integration and recurring governance costs mean smaller firms may face higher relative costs — potentially increasing scale advantages for larger incumbents.
Deployment case studies and cost reports indicating significant fixed integration and governance costs; inference to market structure is speculative.
low negative The Effectiveness of ChatGPT in Customer Service and Communi... relative upfront and ongoing costs; indicators of scale advantages or market con...
There is a risk of deskilling through excessive reliance on AI, implying a need for continuous training and certification to preserve human judgment.
Qualitative interview evidence and observed concerns about overreliance; authors recommend training/governance based on identified risks; no direct longitudinal measurement of deskilling provided in summary.
low negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... human skill levels (deskilling risk); need for training/certification
Recommendation algorithms and widespread automated advice can induce herding or increase common exposures across retail investor portfolios, with potential macroprudential implications.
Theoretical discussion supported by examples from retail trading episodes and algorithmic amplification literature referenced in the review (conceptual and anecdotal evidence; limited systematic empirical quantification).
low negative Women's Investment Behaviour and Technology: Exploring the I... portfolio correlation across users, asset demand concentration, market volatilit...
Exposure to AI and platform work produces psychosocial effects for workers, including increased job insecurity, stress, and changing task content in surviving occupations.
Surveys, qualitative case studies, and workplace studies summarized in the review reporting worker‑reported insecurity and stress; the review also highlights inconsistent measurement and limited systematic evidence on psychosocial outcomes.
low negative The Impact of AI Machine Learning on Human Labor in the Work... job insecurity, stress, psychosocial wellbeing, and perceived changes in task co...
Regulators and standard-setters who value transparency and auditability will need to account for the gap between evaluation results and actionable fixes; firms may require incentives or rules to ensure evaluation leads to remediation, not just documentation.
Authors' policy implication derived from the study's finding of a results-actionability gap and discussion of auditability concerns; speculative recommendation rather than empirical finding.
low neutral Results-Actionability Gap: Understanding How Practitioners E... policy/regulatory effectiveness regarding evaluation leading to remediation (spe...
Delegation of oversight and reallocation of monitoring tasks due to AI integration changes transaction costs and affects organizational design and governance needs (e.g., more verification/audit effort or specialist oversight roles).
Based on participants' reported shifts in who performed monitoring/oversight tasks in the 40 interviews and the authors' interpretation of those shifts in organizational/economic terms.
low neutral AI in project teams: how trust calibration reconfigures team... transaction/monitoring costs and governance arrangements
Expect rising demand and wage premia for managers with hybrid capabilities (systems thinking + computational literacy), with a risk of widening returns to managerial skill heterogeneity.
Theoretical implication from predicted complementarities and task reallocation; prescriptive economic inference without empirical labor-market evidence in the paper.
low positive Comparative analysis of strategic vs. computational thinking... labor demand, wage premia, and distributional widening across managerial skill t...
Managers’ time will be reallocated toward hybrid tasks (interpretation, oversight, ethical deliberation), increasing returns to combined strategic and computational skills.
Predictive inference from the role reconfiguration analysis and task-complementarity argument; forward-looking theoretical forecast (no empirical time-use data).
low positive Comparative analysis of strategic vs. computational thinking... managerial time allocation (share devoted to hybrid tasks) and returns/wage prem...
Standards for provenance, labeling of AI-generated content, and interoperable evidence formats would lower verification costs and create beneficial network effects.
Policy recommendation derived from identified verification frictions and the study's analysis of data/model governance needs.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... verification cost and interoperability/network effects
There is growing market demand for AI-assisted fact-checking tools, creating opportunities for software, monitoring services, and labeled datasets.
Analytic implication drawn from findings about increasing AI use and needs for automation/labeling; based on interviews and market inference in the study.
low positive Fact-Checking Platforms in the Middle East: A Comparative St... market demand for AI tools and labeled datasets
Regulators should consider guidelines on AI monitoring, algorithmic fairness in performance evaluations, and protections to prevent hybrid‑induced career penalties.
Policy recommendation based on conceptual assessment of risks identified in literature synthesis; not an empirical claim—no policy evaluation data provided.
low positive The Sociology of Remote Work and Organisational Culture: How... existence/applicability of regulatory guidelines; protections against career pen...
Hybrid agency implies complementarity between GenAI and managerial/knowledge‑worker skills (curation, evaluation, coordination), potentially increasing returns to those skills while automating routine cognitive tasks—consistent with skill‑biased technological change.
Synthesis of recurring themes linking GenAI capabilities with managerial skill topics in the thematic clusters; positioned as an implication for labour demand and skill composition rather than an empirically tested effect.
low positive Generative AI and the algorithmic workplace: a bibliometric ... expected changes in returns to managerial/knowledge‑worker skills and automation...
There is demand for tooling that bridges evaluation outputs to actionable fixes (e.g., failure-mode libraries, standardized remediation templates, evaluation-to-priority mapping), signaling economic opportunities for third-party tools and consulting services.
Authors' inference based on the documented results-actionability gap and participants' descriptions of pain points; presented as a market implication rather than direct market measurement.
low positive Results-Actionability Gap: Understanding How Practitioners E... inferred market demand for evaluation-to-action tooling/services
Firms that invest in instrumentation, cross-functional processes, and remediation levers capture more value from LLMs; organizations with better evaluation-to-action pipelines will obtain higher productivity gains and market edge.
Authors' inference from observed heterogeneity among teams in the interviews and comparison of practices in teams that reported more success converting evaluations into changes.
low positive Results-Actionability Gap: Understanding How Practitioners E... relative productivity/value capture tied to evaluation-to-action capability (inf...
Policy instruments (law and markets) should be designed to remain institutionally and procedurally responsive to ethical claims that resist full codification (e.g., through participatory governance, oversight mechanisms, equitable redress, care-centered procurement standards).
Normative policy prescriptions derived from the Levinasian diagnosis and case illustrations; proposed measures are normative and not empirically evaluated within the paper.
low positive Examining ethical challenges in human–robot interaction usin... responsiveness of policy and market instruments to non-codifiable ethical claims...
Integrating Object-Oriented Ontology (OOO) and the material turn enables attention to nonhuman actors and assemblages without collapsing them into human-centered instrumentalism.
Theoretical synthesis of OOO/material-turn literature and argument that this synthesis offers analytic resources for socio-technical assemblages; illustrated conceptually in domains.
low positive Examining ethical challenges in human–robot interaction usin... conceptual adequacy of analytic lens for nonhuman actors and assemblages (qualit...
Humans who configure and teach agents gain understanding and skills themselves — learning-by-teaching generates human capital accumulation endogenous to agent deployment (bidirectional scaffolding).
Qualitative, naturalistic observations and comparative documentation of users configuring/teaching agents during the one-month study; no randomized assignment or pre/post quantitative skill testing reported.
low positive When Openclaw Agents Learn from Each Other: Insights from Em... human skill accumulation / understanding from configuring/teaching agents
Regulators may prefer systems that support contestability and audit trails and could mandate argumentation-style explainability in certain sectors.
Speculative policy prediction; no regulatory statements or empirical policy adoption evidence cited.
low positive Argumentative Human-AI Decision-Making: Toward AI Agents Tha... regulatory adoption rate of contestability/audit-trail requirements