The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (3470 claims)

Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 609 159 77 736 1615
Governance & Regulation 664 329 160 99 1273
Organizational Efficiency 624 143 105 70 949
Technology Adoption Rate 502 176 98 78 861
Research Productivity 348 109 48 322 836
Output Quality 391 120 44 40 595
Firm Productivity 385 46 85 17 539
Decision Quality 275 143 62 34 521
AI Safety & Ethics 183 241 59 30 517
Market Structure 152 154 109 20 440
Task Allocation 158 50 56 26 295
Innovation Output 178 23 38 17 257
Skill Acquisition 137 52 50 13 252
Fiscal & Macroeconomic 120 64 38 23 252
Employment Level 93 46 96 12 249
Firm Revenue 130 43 26 3 202
Consumer Welfare 99 51 40 11 201
Inequality Measures 36 105 40 6 187
Task Completion Time 134 18 6 5 163
Worker Satisfaction 79 54 16 11 160
Error Rate 64 78 8 1 151
Regulatory Compliance 69 64 14 3 150
Training Effectiveness 81 15 13 18 129
Wages & Compensation 70 25 22 6 123
Team Performance 74 16 21 9 121
Automation Exposure 41 48 19 9 120
Job Displacement 11 71 16 1 99
Developer Productivity 71 14 9 3 98
Hiring & Recruitment 49 7 8 3 67
Social Protection 26 14 8 2 50
Creative Output 26 14 6 2 49
Skill Obsolescence 5 37 5 1 48
Labor Share of Income 12 13 12 37
Worker Turnover 11 12 3 26
Industry 1 1
Clear
Org Design Remove filter
There are significant AI adoption challenges in education and industry that affect employability and role transformation.
Synthesized evidence from industry reports and empirical studies discussed in the review highlighting barriers to adoption in education and industry.
The transformation toward algorithmic enterprises raises critical concerns regarding agency, accountability, data monopolization, and algorithmic bias.
Presented as a principal concern in the paper's conceptual discussion and interdisciplinary critique; based on analysis of governance and ethical literature rather than new empirical evidence in the abstract.
high negative Algorithmic Enterprises: Rethinking Firm Strategy in the Age... risks to agency, accountability, market power (data monopolization), and algorit...
Industrial firms face a dual challenge: (1) the development and deployment of digital technologies and (2) the proliferation and integration of the corresponding skills portfolios.
Conceptual framing and literature synthesis presented in the paper (identification by authors); not tied to a specific quantitative sample in the provided text.
high negative Industry 4.0 Inc.—Mergers and acquisitions and the digital t... ability to develop and deploy digital technologies and integrate skills portfoli...
The study is framed based on Job Demands-Resources (JD-R) theory, positing that HAI-C task complexity is a job demand and AI self-efficacy/humble leadership act as resources that can mitigate negative effects on engagement.
Introduction states JD-R theory as the theoretical basis and describes job demands (HAI-C task complexity) and job/personal resources (humble leadership, AI self-efficacy) in the hypothesized model.
high negative How does human-AI collaboration task complexity affect emplo... theoretical framing / hypothesized relationships
HAI-C tech-learning anxiety reduces employees' work engagement (serves as the mediator between HAI-C task complexity and work engagement).
Mediation analysis via hierarchical regression and bootstrapping on the three-wave survey sample of 497 employees; reported in Results as the mediating mechanism.
Human-AI collaboration task complexity (HAI-C task complexity) negatively affects employees' work engagement by amplifying their HAI-C tech-learning anxiety.
Three-wave longitudinal survey of matched data from 497 employees; mediation analysis using hierarchical regression and bootstrapping reported in the Results section.
Important boundary conditions include data maturity, process integration, governance discipline, and the degree of functional trust between finance and operating units.
List of boundary conditions reported in the paper based on documentary case analysis and synthesis with literature.
high negative Research on the Impact of Generative AI on the Quality of Ma... constraints on GenAI impact on management accounting decision quality
GenAI does not improve management accounting decision quality primarily by replacing managerial judgment.
Interpretive finding based on documentary analysis of disclosures from the three case firms and relevant literature; presented as a summary conclusion in the paper.
high negative Research on the Impact of Generative AI on the Quality of Ma... management accounting decision quality
Beyond technical barriers there are organizational ones: a persistent AI literacy gap, cultural heterogeneity, and governance structures that have not yet caught up with agentic capabilities.
Interview data (over 30) reporting organizational challenges including limited AI literacy, diverse cultural attitudes across organizations, and lagging governance relative to agentic AI capabilities.
high negative Agentic AI in Engineering and Manufacturing: Industry Perspe... organizational readiness factors (AI literacy, culture, governance alignment)
Adoption is constrained less by model capability than by fragmented and machine-unfriendly data, stringent security and regulatory requirements, and limited API-accessible legacy toolchains.
Stakeholder interviews (over 30) reporting barriers to deployment; qualitative synthesis identifies data fragmentation, security/regulatory requirements, and legacy toolchain access as primary constraints.
high negative Agentic AI in Engineering and Manufacturing: Industry Perspe... barriers to AI adoption in engineering/manufacturing
The value alignment problem for artificial intelligence (AI) is often framed as a purely technical or normative challenge, sometimes focused on hypothetical future systems.
Author's literature-based observation and critique in the paper's introduction (conceptual argument; no empirical sample reported).
high negative Relative Principals, Pluralistic Alignment, and the Structur... framing_of_problem_in_literature
Regulated deployment imposes four load-bearing systems properties — deterministic replay, auditable rationale, multi-tenant isolation, statelessness for horizontal scale — and stateful architectures violate them by construction.
Conceptual/architectural argument presented in the paper (theoretical analysis), not an empirical measurement in the abstract.
high negative Stateless Decision Memory for Enterprise AI Agents compatibility of stateful architectures with regulatory/system properties
The policy and research challenge posed by platform-mediated automation is not merely job quantity (technological unemployment) but institutional continuity — how societies reproduce practical competence when platforms optimize for efficiency rather than formation.
Normative and conceptual claim developed through literature synthesis (institutional economics, platform governance, workforce development); presented as an analytical reframing rather than an empirically tested hypothesis.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... institutional continuity and human capital reproduction (quality of workforce fo...
Entry-level roles have historically functioned as apprenticeships in which workers acquire tacit knowledge and critical judgment; if platforms curtail these formative occupational layers, organizations may lack future workers capable of exercising contextual reasoning required to manage complex systems.
Institutional economics and workforce development literature cited in the paper; conceptual synthesis without original empirical measurement reported.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... human capital formation (tacit knowledge acquisition and contextual reasoning ca...
Platform-mediated automation risks hollowing out labor structures from both directions: eroding repetitive, junior roles from below and automating supervisory coordination functions from above.
Theoretical argument synthesizing institutional economics and platform literature; articulated as a conceptual risk rather than demonstrated with original empirical data.
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... structural change in occupational layers (hollowing out of junior and supervisor...
Algorithmic systems are displacing routine tasks across both low-wage entry-level work and middle-management functions.
Stated in paper's argumentation; supported by a literature-based review drawing on platform governance literature and recent research on AI-enhanced automation (no original empirical sample or quantitative study reported).
high negative When Platforms Replace the Pipeline: AI, Labor Erosion, and ... displacement of routine tasks (across entry-level and middle-management roles)
Training data scarcity is an emerging challenge for organizations that aim to train proprietary LLMs.
Paper highlights training data scarcity as a challenge in its analysis and discussion sections (qualitative observation).
high negative Buy Or Build? A Practitioner’s Framework for Large Language ... feasibility of training proprietary LLMs (availability of training data)
The infrastructure for cross-user agent collaboration is entirely absent, let alone the governance mechanisms needed to secure it.
Authoritative claim in paper framing the research gap; presented as observational/argumentative (no empirical audit reported).
high negative ClawNet: Human-Symbiotic Agent Network for Cross-User Autono... availability of cross-user collaboration infrastructure and governance mechanism...
Current AI agent frameworks have made remarkable progress in automating individual tasks, yet all existing systems serve a single user.
Statement in paper's introduction/positioning; conceptual survey-style claim (no empirical study or systematic benchmark reported).
high negative ClawNet: Human-Symbiotic Agent Network for Cross-User Autono... automation scope (single-user vs multi-user)
As multimodal AI achieves human-parity understanding of speech and gesture, [the keyboard's] necessity dissolves.
Theoretical claim supported by multidisciplinary review (history, neuroscience, technology, organizational studies); no quantified empirical test reported.
high negative The Instrumental Dissolution of Typing: Why AI Challenges th... necessity/usage of keyboard as default input
Current session-based context handling (sessions ending, context windows filling, memory APIs returning flat facts) produces intelligence that is powerful per session but amnesiac across time.
Descriptive diagnostic argument in the paper; no empirical measurement reported in this text.
high negative The Continuity Layer: Why Intelligence Needs an Architecture... temporal persistence of model 'understanding' (memory/continuity)
The paper identifies governance challenges such as accountability gaps, digital sovereignty risks, ethical pluralism, and strategic weaponization arising from embedding AI in diplomatic practice.
Conceptual and normative analysis section of the paper outlining risks and governance challenges; illustrated by examples and argumentation.
high negative Strategic Cognition and Artificial Diplomacy: Designing Huma... presence of governance risks (accountability gaps, digital sovereignty, ethical ...
Thin training coverage fosters anxiety about substitution and slows diffusion of AI tools.
Reported associations from surveys of mid-level managers and technical staff, interviews, and document analysis across cases; thematic coding identified links between limited training, worker anxiety, and slower diffusion. (Sample size not reported.)
high negative Overcoming Resistance to Change: Artificial Intelligence in ... worker anxiety and speed of diffusion/adoption
Agency in software engineering is primarily constrained by organizational policies rather than individual preferences.
Authors' synthesis of qualitative results across the ACTA/Delphi and task/review phases indicating organizational policy factors were cited as primary constraints.
high negative From Junior to Senior: Allocating Agency and Navigating Prof... Primary source of constraint on developer agency (organizational policy vs indiv...
The study identified significant implementation challenges including algorithmic bias, digital divide concerns, data privacy risks, and low technology readiness among HR teams in Tier 2 cities.
Synthesis of qualitative case study findings from 4 organizations plus survey responses (N=150) reporting barriers and risks encountered during adoption.
high negative A Study on the Effectiveness of Technology-Driven Recruitmen... implementation challenges / risks
This condition of authorship uncertainty reshapes how teams attribute ideas, negotiate accountability, and coordinate collective reasoning.
Theoretical claim based on conceptual analysis in the paper; no empirical method or sample described in the abstract.
high negative Who Gets Credit? Operationalizing AI Disclosure as Epistemic... idea attribution, accountability negotiation, collective reasoning / coordinatio...
As generative AI becomes an ambient presence in collaborative work, a new social ambiguity emerges around authorship and responsibility.
Conceptual argument presented in the paper (theoretical/observational claim). No empirical method or sample size reported in the abstract.
high negative Who Gets Credit? Operationalizing AI Disclosure as Epistemic... authorship uncertainty / attribution of responsibility
Large language models remain confined to linguistic simulation rather than grounded understanding.
Conceptual assertion in the paper arguing limits of current models; no empirical tests or measurements reported.
high negative Governing Reflective Human-AI Collaboration: A Framework for... grounded_understanding (absence thereof)
Fluency is not reliability: without structures that stabilise both human and model reasoning, AI cannot be trusted or governed where it matters most.
Central thesis/claim of the paper; normative argument synthesising the paper's observations and proposals rather than an empirically tested finding provided here.
high negative The Missing Knowledge Layer in AI: A Framework for Stable Hu... trustworthiness/governability of AI in high-stakes contexts
Humans often mistake fluency for reliability: when a model responds smoothly, users tend to trust it, even when both model and user are drifting together.
Behavioral/psychological assertion in the paper referencing human interaction patterns with fluent outputs; no experimental data or sample size reported in this paper excerpt.
high negative The Missing Knowledge Layer in AI: A Framework for Stable Hu... user trust in model outputs
LLMs produce fluent outputs even when their internal reasoning has drifted; a confident answer can conceal uncertainty, speculation, or inconsistency, and small changes in phrasing can lead to different conclusions.
Conceptual/observational claim presented in the paper; no original empirical test or sample size reported here.
high negative The Missing Knowledge Layer in AI: A Framework for Stable Hu... reliability/consistency of model outputs (decision quality)
Stronger reasoning capabilities do not prevent LLMs from defecting in single-shot social dilemmas (i.e., models defect with or without reasoning enabled).
Authors' experiments that explicitly compared model behavior with reasoning enabled vs disabled in single-shot social dilemmas; details not provided in the excerpt.
high negative CoopEval: Benchmarking Cooperation-Sustaining Mechanisms and... cooperation/defection rates conditional on reasoning capability being enabled
Repetition-induced cooperation deteriorates drastically when co-players vary.
Authors' experimental observation comparing repeated-game cooperation under fixed vs varying co-players in their study; no quantitative metrics or sample sizes provided in the excerpt.
high negative CoopEval: Benchmarking Cooperation-Sustaining Mechanisms and... cooperation level under repeated interactions when co-players vary
Our experiments show that recent models — with or without reasoning enabled — consistently defect in single-shot social dilemmas.
Authors' experimental results comparing recent LLMs in single-shot social dilemma games, with reasoning enabled vs disabled; specific models, number of games, and sample sizes are not provided in the excerpt.
high negative CoopEval: Benchmarking Cooperation-Sustaining Mechanisms and... rate of defection (vs cooperation) in single-shot social dilemmas
Recent works report that LLMs with stronger reasoning capabilities behave less cooperatively in mixed-motive games such as the prisoner's dilemma and public goods settings.
Statement referencing prior literature (recent works) summarized in the paper's introduction/background; no specific dataset or sample size given in the excerpt.
high negative CoopEval: Benchmarking Cooperation-Sustaining Mechanisms and... cooperative behavior in mixed-motive games (e.g., prisoner's dilemma, public goo...
Efficiency (e.g., minimizing time and cost with AI-only planning) does not equal effectiveness: optimizing for efficiency can erode team cognition and reduce decision quality.
Synthesis of experimental quantitative results (time/cost vs. risk capture and rework) and qualitative assessment indicating that AI-driven efficiency can come at the expense of risk awareness and planning robustness.
high negative Cognitive Offloading in Agile Teams: How Artificial Intellig... trade-off between efficiency and decision quality / team cognition
Human-only planning incurs substantial overhead.
Same controlled experiment reporting that human-only planning produced higher time and cost overheads relative to AI-assisted approaches.
high negative Cognitive Offloading in Agile Teams: How Artificial Intellig... planning overhead (time/cost)
AI-only planning increases rework due to unstated assumptions.
Experiment measured rework rates and accompanying qualitative analysis attributing increased rework in the AI-only condition to unstated assumptions made by algorithmic planning.
AI-only planning significantly degrades risk capture rates.
Same controlled three-condition experiment on a live client deliverable; paper reports measures/qualitative indicators of risk capture rates and states degradation for AI-only condition.
Existing competition-aware CFL and incentive-design approaches reward organizations based on marginal training contributions but fail to account for the costs of strengthening competitors.
Literature critique and comparison in the paper; theoretical discussion rather than a reported empirical trial or sample.
high negative Cooperate to Compete: Strategic Data Generation and Incentiv... adequacy_of_incentive_design (accounting for competitor-strengthening costs)
Non-IID data amplifies this coopetition dilemma by producing asymmetric learning gains across organizations and undermining sustained participation.
Conceptual claim supported by the paper's theoretical modeling and later experiments (described as 'non-IID data' experiments); no numeric sample size given in abstract.
high negative Cooperate to Compete: Strategic Data Generation and Incentiv... asymmetry_in_learning_gains / sustained_participation
Cross-silo federated learning (CFL) deployments in data-sensitive domains are inherently coopetitive: organizations cooperate during model training while competing in downstream markets, so training contributions can inadvertently strengthen rivals.
Conceptual argument and literature motivation presented in the paper's introduction; no empirical sample size reported.
high negative Cooperate to Compete: Strategic Data Generation and Incentiv... strengthening_of_rivals / participation incentives
The study proposes an integrative conceptual model and research propositions highlighting cross-functional challenges in governance, organizational capabilities, socio-technical alignment, and responsible implementation.
Statement in abstract that the authors developed a conceptual model and research propositions based on their review and identified cross‑functional challenges.
high negative The implementation of artificial intelligence in organizatio... identification_of_challenges (governance, capabilities, socio-technical_alignmen...
Most AI tooling targets that fraction [the ~10% of the workday spent writing code].
Assertion made in the paper (abstract) as an observed mismatch between where AI tooling focuses and overall developer work activities.
high negative To Copilot and Beyond: 22 AI Systems Developers Want Built focus of AI tooling relative to developer time allocation
Environmental demands place an upper bound on the degree of heterogeneity required in a distributed production system.
Theoretical claim derived from the Distributed Production System framework and discussed in the paper; supported by conceptual argument and model constraints rather than empirical data; no sample size reported.
high negative The Principle of Maximum Heterogeneity Optimises Productivit... required degree of heterogeneity (upper bound) given environmental demands
The study's findings are subject to design limitations including an AM/PM session confound, differential attrition, and LLM grading sensitivity to document length.
Authors' reported limitations section citing specific threats to internal validity and measurement (session timing confound, differential attrition across conditions, and grading biases of the LLM used to evaluate documents).
high negative Scaffolding Human-AI Collaboration: A Field Experiment on Be... threats to validity (confounds and measurement sensitivity)
The behavioral scaffolding intervention was associated with substantially lower document production.
Same field experiment (N=388); the behavioral scaffolding required joint AI use within pairs and was compared to unstructured use, with reported reductions in document production in the behavioral condition.
high negative Scaffolding Human-AI Collaboration: A Field Experiment on Be... document production (quantity of documents produced)
A behavioral scaffolding intervention (a structured protocol requiring joint AI use within pairs) was associated with lower document quality relative to unstructured use.
Field experiment with 388 employees at a Fortune 500 retailer; random/experimental assignment to scaffolding conditions while all participants had access to the same AI tool; comparison reported between behavioral scaffolding condition and unstructured use.
Rote learning will become obsolete in favor of contextual application.
Paper's forward-looking prediction based on synthesis of adult learning theory and workforce development literature; no empirical sample size or quantified trend data provided.
high negative The Future of Education in an AI-Driven World: Preparing Org... decline/obsolescence of rote learning and increase in contextual application
Enterprise sales organizations are systematically hampered by what this paper terms 'Revenue Friction'—the accumulative productivity loss caused by fragmented, human-mediated data entry across disconnected CRM, ERP, and quoting systems.
Statement/definition presented in the paper excerpt. No empirical method, sample size, or quantitative evidence reported in the provided text.
high negative From CRM to Cognition: Autonomous Revenue Operations Systems... accumulative productivity loss (termed 'Revenue Friction') resulting from fragme...