The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (4793 claims)

Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 402 112 67 480 1076
Governance & Regulation 402 192 122 62 790
Research Productivity 249 98 34 311 697
Organizational Efficiency 395 95 70 40 603
Technology Adoption Rate 321 126 73 39 564
Firm Productivity 306 39 70 12 432
Output Quality 256 66 25 28 375
AI Safety & Ethics 116 177 44 24 363
Market Structure 107 128 85 14 339
Decision Quality 177 76 38 20 315
Fiscal & Macroeconomic 89 58 33 22 209
Employment Level 77 34 80 9 202
Skill Acquisition 92 33 40 9 174
Innovation Output 120 12 23 12 168
Firm Revenue 98 34 22 154
Consumer Welfare 73 31 37 7 148
Task Allocation 84 16 33 7 140
Inequality Measures 25 77 32 5 139
Regulatory Compliance 54 63 13 3 133
Error Rate 44 51 6 101
Task Completion Time 88 5 4 3 100
Training Effectiveness 58 12 12 16 99
Worker Satisfaction 47 32 11 7 97
Wages & Compensation 53 15 20 5 93
Team Performance 47 12 15 7 82
Automation Exposure 24 22 9 6 62
Job Displacement 6 38 13 57
Hiring & Recruitment 41 4 6 3 54
Developer Productivity 34 4 3 1 42
Social Protection 22 10 6 2 40
Creative Output 16 7 5 1 29
Labor Share of Income 12 5 9 26
Skill Obsolescence 3 20 2 25
Worker Turnover 10 12 3 25
Clear
Productivity Remove filter
AI adoption raises a risk of digital colonialism: foreign control of data, platforms, and value capture may divert economic gains away from local actors.
Conceptual analysis drawing on policy documents and empirical literature about data flows, platform economics, and international investment; no original quantitative measurement in this paper.
medium negative Towards Responsible Artificial Intelligence Adoption: Emergi... data ownership, revenue capture by foreign firms, local value capture, concentra...
Increased monitoring and algorithmic management raise concerns about worker autonomy and privacy and will prompt regulatory responses (data protection, algorithmic transparency) that shape adoption costs and trajectories.
Recurring concerns reported across included studies and the review's policy implication section; grounded in qualitative and normative discussions within the literature.
medium negative Data-Driven Strategies in Human Resource Management: The Rol... worker autonomy/privacy incidents, regulatory actions, adoption costs
Over-standardisation of curricula can create mismatches between certified competencies and firm-specific needs.
Stated in Risks: the paper warns that overly standardized curricula may not fit firm-specific requirements. This is a conceptual caution, not supported by within-paper empirical comparisons.
medium negative Curriculum engineering: organisation, orientation, and manag... alignment between certified competencies and firm-specific job demands (skill-ma...
High fixed costs may concentrate training capacity among a few providers, risking reduced competition.
Listed under Risks to Watch: the paper warns that high fixed costs could concentrate capacity. This is a theoretical market-concentration risk; no empirical market analysis is provided.
medium negative Curriculum engineering: organisation, orientation, and manag... market concentration (Herfindahl index), number of active training providers
Upfront and maintenance costs are substantial; economic evaluation should compare these costs to downstream benefits such as placement rates and productivity gains.
Paper recommends economic evaluation, lists cost-per-curriculum and other cost metrics; presented as advice rather than results. No empirical cost–benefit data provided.
medium negative Curriculum engineering: organisation, orientation, and manag... cost-per-curriculum, ROI metrics, placement rates, productivity measures
Complexity and lock-in to specific standards may raise barriers to innovation and increase switching costs.
Discussed in Regulation and compliance economics and Risks: claims that standardisation and embedded processes could produce vendor/standard lock-in. This is a theoretical risk flagged by the authors, not supported by empirical data in the paper.
medium negative Curriculum engineering: organisation, orientation, and manag... switching costs, rate of innovation adoption, vendor dependence indicators
Model risk, bias, and privacy concerns impose negative externalities (e.g., systemic risk in supply chains, discrimination), motivating governance standards, auditing, and possibly regulation.
Documentation in standards, practitioner reports, and conceptual literature within the 2020–2025 review describing incidents, risks, and calls for governance/regulation.
medium negative Integrating Artificial Intelligence and Enterprise Resource ... externality indicators (e.g., cross-firm contagion incidents, measured discrimin...
Firms will need to invest in new control technologies, governance structures, and personnel (AI auditors, red teams), increasing the total cost of GenAI adoption.
Economic reasoning and implications section; no empirical cost estimates or survey data; projection based on anticipated control needs.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... total cost of GenAI adoption including ongoing control and governance expenditur...
Malicious insiders, external actors (vendors, consultants, customers), shadow AI (unsanctioned consumer-grade GenAI use), and supply-chain/third-party prompt templates are plausible attack vectors for prompt fraud.
Threat taxonomy and scenario mapping with case-style examples; conceptual identification of actors rather than documented incident attribution.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... range of plausible adversary vectors capable of injecting malicious prompts
Poor logging, weak prompt governance, and over-reliance on machine-generated artifacts increase organizational vulnerability to prompt fraud.
Control gap analysis and prescriptive argumentation; examples of weak controls used to illustrate exploitability; no empirical measurement of effect sizes.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... organizational vulnerability/risk exposure to prompt fraud given control quality
Because prompt fraud operates at the linguistic/procedural surface rather than the network/technical surface, existing control frameworks are ill-prepared to address this new attack surface.
Control gap analysis comparing conventional internal controls to the linguistic attack surface; conceptual rather than empirical evaluation.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... adequacy of existing internal control frameworks to mitigate prompt-driven risks
Upfront governance costs (policy, tooling, staff) become a key part of adoption cost and affect ROI calculations and payback periods for automation investments.
Economic reasoning and implications discussed in the paper; no empirical cost data provided—recommendation based on practitioner experience and theoretical cost accounting.
medium negative Governed Hyperautomation for CRM and ERP: A Reference Patter... adoption costs, ROI, payback periods (economic outcomes, not empirically measure...
Traditional automation governance is often ad hoc, underestimates security and compliance risks, and does not scale safely for mission-critical enterprise systems.
Synthesis of industry best practices and practitioner-sourced lessons (qualitative observations and case illustrations). No systematic survey or quantitative incidence rates provided.
medium negative Governed Hyperautomation for CRM and ERP: A Reference Patter... quality of governance practices; prevalence of security/compliance risk awarenes...
Prompt fraud reduces the marginal cost of producing convincing fraudulent artifacts, which may increase fraud frequency and expected losses absent mitigations.
Economic reasoning and conceptual modeling of incentives; no empirical estimates of frequency or losses included.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... expected frequency of fraud and expected losses under unchanged mitigation effor...
Lack of prompt provenance, versioning, and validation practices increases organizational exposure to prompt fraud.
Conceptual analysis and recommended controls (provenance/versioning) drawn from audit-framework comparisons and threat modeling.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... existence of prompt-provenance/versioning/validation practices and associated ri...
There is insufficient logging/traceability of prompts, responses, and model versions in many workflows, creating a control weakness for detecting prompt fraud.
Observations from literature/regulatory review and the paper's threat/control mapping; asserted as a common operational gap (no systematic measurement).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... presence/quality of prompt/response/model-version logging and its sufficiency fo...
Shadow AI — unsanctioned, decentralized use of GenAI tools — amplifies prompt-fraud risk by bypassing central controls and audit trails.
Conceptual analysis and organizational risk reasoning; references to common practices of unsanctioned tool use (no empirical prevalence data).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... increase in unmonitored prompt activity and corresponding reduction in detectabi...
External actors can commit prompt fraud via customer-facing systems or social-engineering prompt chains.
Conceptual threat scenarios and mapping of attack surfaces (customer-facing interfaces, input channels); illustrative examples provided.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... risk of prompt-fraud initiated through external-facing inputs or social-engineer...
Internal actors manipulating prompts within authorized AI workflows are a realistic and important threat vector for prompt fraud.
Threat modeling and scenario-based analysis highlighting insiders with authorized access who can craft prompts.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... risk or incidence of prompt-fraud events originating from internal actors
Prompt fraud can defeat controls that rely on plausibility, standard formatting, or human review that trusts model-like language.
Threat mapping and literature on automation bias; illustrative vignettes showing how machine-like outputs mimic authoritative formats.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... effectiveness of plausibility/format/human-review-based controls in identifying ...
Prompt fraud lowers the entry cost of producing convincing fraudulent artifacts, increasing the ease with which attackers can create plausible forgeries.
Economic reasoning and conceptual analysis based on GenAI behavior and illustrative scenarios (no empirical cost or frequency data).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... marginal cost (effort/resources) required to produce convincing fraudulent artif...
Prompt fraud — the intentional manipulation of natural-language prompts to cause generative AI systems to produce misleading, fabricated, or deceptive artifacts that bypass internal controls — constitutes a novel, low-cost fraud vector that traditional IT- and process-focused controls are ill-equipped to detect or prevent.
Conceptual analysis and threat modeling grounded in literature/regulatory review and illustrative vignettes; no systematic empirical incidence data provided.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... ability of existing IT/process controls to detect or prevent fraud produced via ...
Secure infrastructure (including SECaaS-provided tools) affects the availability and trustworthiness of AI training data and models; breaches reduce returns to AI R&D via direct losses and reduced trust.
Conceptual linkage supported by case studies of data/model theft and technical literature on secure enclaves, differential privacy, federated learning; no broad quantitative estimate provided.
medium negative Security- as- a- service: enhancing cloud security through m... incidence of data/model breaches, economic returns to AI R&D
Security externalities (one firm's breach raising ecosystem risk) complicate private incentives and may justify policy interventions such as standards or mandatory reporting.
Economic theory on externalities, case studies showing spillovers from breaches, and policy analyses recommending interventions.
medium negative Security- as- a- service: enhancing cloud security through m... spillover risk, incentive alignment, justification for regulation
Concentration among large cloud/SECaaS providers can create market power, platform dependency, and affect competition in AI markets.
Market-structure theory, observed concentration patterns in industry reports, and qualitative case studies; no causal estimates provided in the chapter.
medium negative Security- as- a- service: enhancing cloud security through m... market power indicators, competition measures in AI markets
Latency and integration frictions can limit the suitability of SECaaS for specialized workloads, including some AI pipelines.
Technical evaluations and benchmarks that measure latency/resource overhead; reports and case studies noting integration challenges for high-throughput or low-latency workloads.
medium negative Security- as- a- service: enhancing cloud security through m... latency, integration overhead, suitability for AI workloads
Reliance on a small set of major cloud/SECaaS providers creates vendor lock-in, concentration risk, and systemic vulnerability if a major provider is compromised.
Market-structure discussions, observed provider outages and incidents (case studies), and theoretical arguments about concentration; no single causally identified empirical estimate provided.
medium negative Security- as- a- service: enhancing cloud security through m... market concentration, systemic risk, dependency risk
Without improvements in robustness, consistency, and neuroscientific validity of explanations, clinical uptake will be constrained, slowing commercialization and reducing returns for developers focused only on performance.
Synthesis and forward-looking argument linking methodological deficits documented in the literature to likely reduced market adoption; no direct empirical market impact measurement provided.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... clinical uptake, commercialization pace, developer returns
Weak or inconsistent explanations increase regulatory and medico-legal risk; standardized, validated XAI can lower compliance costs and liability exposure.
Logical inference connecting explanation reliability to regulatory scrutiny and liability concerns, presented as an implication in the review (no direct empirical legal analysis provided).
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... regulatory/compliance and legal risk
Preprocessing pipelines (filtering, artifact removal such as ICA, re-referencing, segmentation) materially affect XAI outputs.
Review cites multiple studies and methodological notes showing explanation maps vary with preprocessing choices; effect reported qualitatively across papers.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... sensitivity of explanation outputs to preprocessing steps
There is a scarcity of human/clinical validation studies testing whether explanations improve clinician decision-making or align with clinical reasoning.
Observation from literature survey: few reviewed works include clinician studies or longitudinal/clinical impact evaluations.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... presence/absence of human/clinical validation
Identified methodological limitations include sensitivity of explanations to hyperparameters and preprocessing choices, inconsistent explanations across similar inputs, and poor correlation with known neurophysiology.
Synthesis of reported failure modes and limitations from multiple EEG-XAI studies reviewed in the paper.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... stability/consistency of explanations and alignment with neurophysiological know...
Most studies focus on qualitative visualizations (e.g., heatmaps) rather than quantitative, reproducible metrics for explanation quality; few evaluate neuroscientific validity or clinical usefulness, and robustness to noise and preprocessing is often untested.
Review-level assessment of evaluation practices across papers, noting prevalence of visual inspection and scarcity of standardized quantitative metrics or clinical validation.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... evaluation rigor: qualitative vs quantitative; assessment of robustness and clin...
Current explainability methods for EEG frequently lack robustness, consistency, and alignment with neuroscientific knowledge, limiting their trustworthiness and practical utility.
Aggregate observations from reviewed EEG-XAI studies noting inconsistent attributions, sensitivity to analysis choices, and few studies that validate explanations against neuroscientific markers or clinical endpoints.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... robustness/consistency/neuroscientific validity of explanations (trustworthiness...
Divergent governance regimes increase the risk of data localization, interoperability frictions, and regulatory fragmentation — raising costs for multinational AI development and limiting global model generalizability.
Policy‑level comparative inference from contrasting national approaches identified in the document analysis and related literature on cross‑border data governance; no direct measurement of costs or model generalizability in the paper.
medium negative Balancing openness and security in scientific data governanc... data localization, interoperability frictions, regulatory fragmentation, costs t...
State‑led coordination can rapidly mobilize resources and scale national champions, altering competitive dynamics and potentially creating winner‑take‑most outcomes.
Theoretical inference from document evidence of state mobilization and developmentalist goals in Chinese texts, combined with literature on state coordination and industrial scaling (no empirical competition measures in the paper).
medium negative Balancing openness and security in scientific data governanc... market concentration / competitive dynamics (winner‑take‑most)
Resource-rich labs and firms are likely to adopt LLM orchestration faster, which could widen gaps in research capacity between institutions and countries unless mitigated by policy choices.
Equity and diffusion argument based on resource requirements (compute, data, validation); no adoption-rate data or cross-institution comparisons provided.
medium negative ChatMicroscopy: A Perspective Review of Large Language Model... adoption rates across institutions, disparities in research capacity
There is potential for 'winner-take-most' market outcomes if a few players combine superior models, instrument control software, and exclusive datasets.
Economics reasoning about network effects and data concentration; no empirical market concentration metrics specific to microscopy provided.
medium negative ChatMicroscopy: A Perspective Review of Large Language Model... market concentration and distribution of market share among firms
Upfront investments required for compute, data labeling, validation, and safety testing may raise entry costs and favor incumbents.
Economic logic about fixed costs and scale advantages; no measured entry-cost or firm-dynamics data provided.
medium negative ChatMicroscopy: A Perspective Review of Large Language Model... entry costs and competitive dynamics (incumbent advantage)
There is a risk of deskilling for some technical roles, creating implications for training and workforce development.
Theoretical reasoning about automation-induced deskilling; no empirical study or measured skill changes provided.
medium negative ChatMicroscopy: A Perspective Review of Large Language Model... level of technical skill required for routine roles and training needs
Optional LLM access without training was associated with shorter written answers compared with no LLM access.
Measured answer length in the randomized trial (n = 164); comparison between untrained optional-access arm and no-access arm showed shorter answers in the untrained-access group.
medium negative Training for Technology: Adoption and Productive Use of Gene... Answer length (measured length of exam answers)
AI reshapes local labor markets by automating routine tasks.
Micro-level analysis of occupations and task content using granular online job-posting data (decomposition of occupational and task changes); panel and IV analyses link higher AI exposure to declines in routine-task employment shares.
medium negative Artificial intelligence, greening of occupational structure ... Share/level of routine-task employment (occupation/task measures from job postin...
Short-run displacement risks from AI adoption create distributional concerns that warrant active labor market policies (retraining, wage insurance) and portable social protections.
Worker-level evidence of short-run employment losses in routine occupations, particularly in emerging economies, and literature synthesis on displacement effects motivating policy recommendations.
medium negative S-TCO: A Sustainable Teacher Context Ontology for Educationa... short-run employment changes in vulnerable occupations and implied welfare/distr...
Human-in-the-loop controls formalize supervisory labor and create persistent oversight costs even after automation scales.
Pattern design and governance lifecycle recommendations highlighting human checkpoints; qualitative reasoning without measurement of oversight hours or costs.
medium negative Governed Hyperautomation for CRM and ERP: A Reference Patter... ongoing human oversight hours/costs per automated transaction
AI-enabled platforms can increase market concentration and platform power, creating competition and data-governance risks and uneven distributional effects across regions and worker skill levels.
Observational platform-concentration indicators and distributional analyses in the case material; scenario and sensitivity checks on distributional outcomes under alternative adoption/policy regimes.
medium negative Artificial Intelligence–Enabled E-Commerce Systems and Autom... market concentration measures (e.g., platform market share), distributional outc...
AI substitutes for and displaces many routine and low-skill occupations, increasing automation risk for those roles.
Multiple empirical studies in the reviewed sample document higher automation/substitution risk and observed employment declines in routine/low-skill tasks and occupations.
medium negative The role of generative artificial intelligence on labor mark... employment levels in routine and low-skill occupations
AI adoption can reinforce winner‑take‑most market dynamics and increase market concentration due to data‑ and AI‑driven advantages.
Theoretical arguments and industry analyses on platform markets and data economies; empirical market‑structure studies and descriptive evidence cited in the review; the claim is derived from synthesis rather than a single causal identification design.
medium negative The Impact of AI Machine Learning on Human Labor in the Work... market concentration measures and firm market shares (competition outcomes)
Impacts of AI on labor are uneven globally: developing regions face larger risks due to digital infrastructure gaps, limited reskilling capacity, and weaker social protections.
Cross‑country comparative analyses, policy and industry reports highlighting infrastructure and institutional differences, and sectoral case studies; review notes geographic bias toward advanced economies in the empirical literature, making some cross‑region inference provisional.
medium negative The Impact of AI Machine Learning on Human Labor in the Work... vulnerability to job displacement, capacity for reskilling, and distributional i...
There is widespread displacement of routine and lower‑skilled tasks associated with AI and automation.
Task‑based analyses decomposing occupations into automatable vs augmentable tasks, econometric studies correlating measures of automation/AI exposure with declines in employment and/or hours in routine occupations, and industry reports documenting automation of routine tasks; evidence is largely from cross‑country and country‑specific empirical work summarized in the review.
medium negative The Impact of AI Machine Learning on Human Labor in the Work... employment levels and task content in routine and lower‑skilled occupations
Traditional macro indicators (GDP, income, unemployment) explain less than 5% of the state- and county-level variation in skills-based exposure.
Statistical analysis/regressions relating the Iceberg Index to standard macro indicators at state and county levels (reported explained variance R^2 < 0.05); sample includes all U.S. states and ~3,000 counties.
medium negative The Iceberg Index: Measuring Workforce Exposure in the AI Ec... percent variance explained (R^2) in the Iceberg Index by traditional macro indic...