The Commonplace
Home Dashboard Papers Evidence Syntheses Digests 🎲

Evidence (5539 claims)

Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 402 112 67 480 1076
Governance & Regulation 402 192 122 62 790
Research Productivity 249 98 34 311 697
Organizational Efficiency 395 95 70 40 603
Technology Adoption Rate 321 126 73 39 564
Firm Productivity 306 39 70 12 432
Output Quality 256 66 25 28 375
AI Safety & Ethics 116 177 44 24 363
Market Structure 107 128 85 14 339
Decision Quality 177 76 38 20 315
Fiscal & Macroeconomic 89 58 33 22 209
Employment Level 77 34 80 9 202
Skill Acquisition 92 33 40 9 174
Innovation Output 120 12 23 12 168
Firm Revenue 98 34 22 154
Consumer Welfare 73 31 37 7 148
Task Allocation 84 16 33 7 140
Inequality Measures 25 77 32 5 139
Regulatory Compliance 54 63 13 3 133
Error Rate 44 51 6 101
Task Completion Time 88 5 4 3 100
Training Effectiveness 58 12 12 16 99
Worker Satisfaction 47 32 11 7 97
Wages & Compensation 53 15 20 5 93
Team Performance 47 12 15 7 82
Automation Exposure 24 22 9 6 62
Job Displacement 6 38 13 57
Hiring & Recruitment 41 4 6 3 54
Developer Productivity 34 4 3 1 42
Social Protection 22 10 6 2 40
Creative Output 16 7 5 1 29
Labor Share of Income 12 5 9 26
Skill Obsolescence 3 20 2 25
Worker Turnover 10 12 3 25
Clear
Adoption Remove filter
Existing extrapolation‑based projection systems understate AI’s nonlinear, spillover, and augmentation effects and miss differential impacts across occupations, industries, regions, and demographic groups.
Theoretical argument and literature-based reasoning in the paper; no quantitative demonstration comparing extrapolation systems to the proposed approach.
medium negative Enhancing BLS Methodologies for Projecting AI's Impact on Em... magnitude and distribution of AI effects (nonlinearity, spillovers, augmentation...
Traditional BLS projection methods are insufficient for forecasting labor market changes driven by rapid AI adoption.
Conceptual critique and argumentation in the paper; no empirical evaluation or comparative forecast error statistics provided.
medium negative Enhancing BLS Methodologies for Projecting AI's Impact on Em... forecasting accuracy / ability to capture AI-driven labor market changes
Rapid post-2020 advances in AI (LLMs and multimodal models) have already rendered some pre-2020 profession-level conclusions obsolete by 2025.
Argument based on observed acceleration in AI capabilities after 2020 (LLMs, multimodal systems) discussed in the paper; evidence is temporal comparison of the state of capabilities and the applicability of older exposure indices rather than a single empirical re-test of all prior predictions.
medium negative Recent Methodologies on AI and Labour - a Desk Review validity/applicability of pre-2020 profession-level forecasts in 2025
Generative AI introduces risks such as model hallucinations and potential erosion of human skills over time.
Practitioner interview reports and authors' interpretive synthesis; qualitative evidence from consulting firms describing hallucination incidents and concerns about reduced skill practice. No longitudinal or quantitative measurement reported.
medium negative Where Automation Meets Augmentation: Balancing the Double-Ed... hallucination/error risk; consultant skill retention/skill erosion
Conversely, lack of standards or failed validation can create regulatory setbacks, reputational risk, and stranded R&D spending.
Case reports and regulatory analysis in the narrative review describing negative outcomes from failed validation or non-aligned AI tools (qualitative evidence).
medium negative Artificial Intelligence in Drug Discovery and Development: R... incidence of regulatory setbacks, reputational damage, amount of stranded/wasted...
Regulators can operationalize 'human oversight' through auditable handover architectures like DAR, but this will increase compliance and record-keeping costs for firms and public bodies.
Policy implication argued in the paper: coupling Reversal Register and hysteresis parameters to regulatory enforcement; no empirical cost estimates provided.
medium negative Human–AI Handovers: A Dynamic Authority Reversal Framework f... compliance_costs; recordkeeping_burden; regulator_enforceability
Compliance with GDPR/CCPA and auditing for bias/harms imposes non-trivial technical and legal costs; implementing federated learning and DP increases engineering complexity and compute cost.
Paper's policy and cost discussion; cites increased engineering complexity and compute demands for privacy-preserving deployments but does not present quantified cost estimates.
medium negative Personalized Content Selection in Marketing Using BERT and G... engineering complexity metrics, compute/resource costs, legal/compliance expendi...
Firms need complementary investments (data pipelines, monitoring tools, feedback loops, human oversight systems) which materially affect the economics of adoption.
Industry case studies and practitioner reports synthesized in the review describing necessary complementary investments; no quantified investment sample or ROI analysis provided here.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... required investment levels, effect on adoption economics and ROI
Regulatory attention is likely to focus on transparency, liability for factual errors, data privacy, and nondiscrimination; compliance and auditing will add to adoption costs.
Policy and regulatory analyses aggregated in the review and references to ongoing regulatory discussions; no primary regulatory impact study conducted in this paper.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... regulatory compliance requirements, related adoption costs, and scope of regulat...
Generative AI currently lacks genuine empathy and relational capabilities necessary for high-stakes or sensitive interactions.
Conceptual analyses and practitioner case examples aggregated in the review; limited direct quantitative measurement cited in this brief review.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... empathy/relational effectiveness in sensitive interactions, customer satisfactio...
Generative models exhibit contextual misunderstandings and cannot reliably infer nuanced customer intent in all cases.
Synthesis of empirical studies and practitioner observations documenting misinterpretation and intent-detection failures; no new testing reported in this review.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... accuracy of intent detection and rate of context-related misunderstandings
Integration complexity (data access, context continuity, privacy/security, workflow alignment) raises implementation costs and time-to-value.
Deployment case studies and vendor reports documenting engineering effort, data plumbing, compliance work, and multi-month integration timelines; no aggregated cost meta-analysis provided.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... implementation cost; time-to-value (time until measurable benefits)
Lack of genuine empathy and emotional intelligence undermines performance on complex or emotionally charged interactions.
Qualitative assessments and noisy measurement from pilot studies and customer feedback in complex cases; limited experimental validation and heterogeneous metrics.
medium negative The Effectiveness of ChatGPT in Customer Service and Communi... customer satisfaction/trust in emotionally charged interactions; resolution qual...
Time/resource costs for re-running analyses and lack of computational environment capture (e.g., Docker/conda containers) increase the difficulty of reproducing results.
Empirical notes from reproduction attempts about compute/time burdens and survey/interview responses highlighting absence of containerized or captured environments as an obstacle.
medium negative On the Computational Reproducibility of Human-Computer Inter... reported burden (time/compute) and absence of environment capture as barriers to...
Environment and dependency issues (library versions, platform differences) are common reproducibility problems.
Failures in running analysis code attributed to dependency/version mismatches and authors' reports; discussion of lack of environment capture (containers/notebooks) as a contributing factor.
medium negative On the Computational Reproducibility of Human-Computer Inter... frequency of environment/dependency issues causing irreproducibility
Unspecified preprocessing steps, parameter settings, or random seeds often prevent exact reproduction of reported results.
Reproduction attempts where outputs differed due to undocumented preprocessing/parameters and corroborating survey/interview accounts from original authors.
medium negative On the Computational Reproducibility of Human-Computer Inter... occurrence of undocumented preprocessing/parameter choices as a barrier to repro...
Incomplete, non-runnable, or poorly documented analysis code is a frequent obstacle to reproducibility.
Empirical attempts to run shared analysis artifacts (scripts, code) and authors' self-reports from surveys/interviews identifying code quality and documentation problems.
medium negative On the Computational Reproducibility of Human-Computer Inter... incidence of code-related failures preventing reproduction (non-runnable or poor...
A common barrier to reproducing results is missing or incomplete data, or data not accessible in the exact form used in the paper.
Observed failure modes from empirical reproduction attempts combined with survey and interview responses from paper authors reporting data availability and completeness issues.
medium negative On the Computational Reproducibility of Human-Computer Inter... frequency/prevalence of missing or incomplete data as a cause of irreproducibili...
AI illiteracy (lack of understanding of AI capabilities/limits) impedes adoption and appropriate use of AI tools in finance.
Survey and interview data reporting lower adoption/intended use among respondents with limited self-reported AI understanding; supplemented by qualitative explanations; sample described as finance professionals across multinational institutions (size unspecified).
medium negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... adoption rates; appropriate use of AI tools
Excessive reliance on algorithmic suggestions can erode human judgment and create systemic risks.
Interview reports and, where available, operational/risk metrics indicating overreliance patterns; authors note systemic-risk implications based on combined qualitative and quantitative observations (no causal identification reported).
medium negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... quality of human judgment; systemic risk
Cognitive biases and inappropriate trust (both overtrust and distrust) distort decision outcomes and limit the benefits of AI-assisted decision-making.
Qualitative interview evidence describing instances of cognitive bias and misplaced trust; some quantitative indicators of decision distortion and risk where operational performance/risk metrics were available; sample: finance professionals across multinational institutions (detailed metrics not specified).
medium negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... decision quality/distortion; systemic risk indicators
Market dominance by global platforms can stifle local entrants and distort competition; policies should address market power and data monopolies.
Review of platform economics and competition policy literature; policy argumentation rather than new empirical competition analysis in this paper.
medium negative Towards Responsible Artificial Intelligence Adoption: Emergi... market concentration indices, entry/exit rates of local firms, measures of compe...
If local data ownership, capacity and governance are weak, economic gains from AI risk accruing to foreign firms and exacerbating income and wealth concentration.
Conceptual synthesis referencing empirical studies on platform rents and data monetization; no original economic distribution analysis presented.
medium negative Towards Responsible Artificial Intelligence Adoption: Emergi... distribution of AI-related revenues, market share of foreign vs local firms, mea...
AI and automation can displace labour—particularly routine tasks—heightening the need for retraining, active labour policies and social protection.
Review of literature on automation and labour markets combined with normative inference for African contexts; no primary labour market data presented.
medium negative Towards Responsible Artificial Intelligence Adoption: Emergi... job displacement rates, changes in task composition, employment levels in routin...
AI adoption raises a risk of digital colonialism: foreign control of data, platforms, and value capture may divert economic gains away from local actors.
Conceptual analysis drawing on policy documents and empirical literature about data flows, platform economics, and international investment; no original quantitative measurement in this paper.
medium negative Towards Responsible Artificial Intelligence Adoption: Emergi... data ownership, revenue capture by foreign firms, local value capture, concentra...
Increased monitoring and algorithmic management raise concerns about worker autonomy and privacy and will prompt regulatory responses (data protection, algorithmic transparency) that shape adoption costs and trajectories.
Recurring concerns reported across included studies and the review's policy implication section; grounded in qualitative and normative discussions within the literature.
medium negative Data-Driven Strategies in Human Resource Management: The Rol... worker autonomy/privacy incidents, regulatory actions, adoption costs
Over-standardisation of curricula can create mismatches between certified competencies and firm-specific needs.
Stated in Risks: the paper warns that overly standardized curricula may not fit firm-specific requirements. This is a conceptual caution, not supported by within-paper empirical comparisons.
medium negative Curriculum engineering: organisation, orientation, and manag... alignment between certified competencies and firm-specific job demands (skill-ma...
High fixed costs may concentrate training capacity among a few providers, risking reduced competition.
Listed under Risks to Watch: the paper warns that high fixed costs could concentrate capacity. This is a theoretical market-concentration risk; no empirical market analysis is provided.
medium negative Curriculum engineering: organisation, orientation, and manag... market concentration (Herfindahl index), number of active training providers
Upfront and maintenance costs are substantial; economic evaluation should compare these costs to downstream benefits such as placement rates and productivity gains.
Paper recommends economic evaluation, lists cost-per-curriculum and other cost metrics; presented as advice rather than results. No empirical cost–benefit data provided.
medium negative Curriculum engineering: organisation, orientation, and manag... cost-per-curriculum, ROI metrics, placement rates, productivity measures
Complexity and lock-in to specific standards may raise barriers to innovation and increase switching costs.
Discussed in Regulation and compliance economics and Risks: claims that standardisation and embedded processes could produce vendor/standard lock-in. This is a theoretical risk flagged by the authors, not supported by empirical data in the paper.
medium negative Curriculum engineering: organisation, orientation, and manag... switching costs, rate of innovation adoption, vendor dependence indicators
Biased training data or objective functions in AI models could perpetuate gender disparities by offering different products or risk scores to men and women.
Review of AI fairness literature and examples of algorithmic disparate impacts summarized in the paper (conceptual and case evidence; not an empirical test tied specifically to fintech products in the review).
medium negative Women's Investment Behaviour and Technology: Exploring the I... differences in product recommendations, risk scoring disparities, disparate outc...
Upfront governance costs (policy, tooling, staff) become a key part of adoption cost and affect ROI calculations and payback periods for automation investments.
Economic reasoning and implications discussed in the paper; no empirical cost data provided—recommendation based on practitioner experience and theoretical cost accounting.
medium negative Governed Hyperautomation for CRM and ERP: A Reference Patter... adoption costs, ROI, payback periods (economic outcomes, not empirically measure...
Traditional automation governance is often ad hoc, underestimates security and compliance risks, and does not scale safely for mission-critical enterprise systems.
Synthesis of industry best practices and practitioner-sourced lessons (qualitative observations and case illustrations). No systematic survey or quantitative incidence rates provided.
medium negative Governed Hyperautomation for CRM and ERP: A Reference Patter... quality of governance practices; prevalence of security/compliance risk awarenes...
Prompt fraud reduces the marginal cost of producing convincing fraudulent artifacts, which may increase fraud frequency and expected losses absent mitigations.
Economic reasoning and conceptual modeling of incentives; no empirical estimates of frequency or losses included.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... expected frequency of fraud and expected losses under unchanged mitigation effor...
Lack of prompt provenance, versioning, and validation practices increases organizational exposure to prompt fraud.
Conceptual analysis and recommended controls (provenance/versioning) drawn from audit-framework comparisons and threat modeling.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... existence of prompt-provenance/versioning/validation practices and associated ri...
There is insufficient logging/traceability of prompts, responses, and model versions in many workflows, creating a control weakness for detecting prompt fraud.
Observations from literature/regulatory review and the paper's threat/control mapping; asserted as a common operational gap (no systematic measurement).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... presence/quality of prompt/response/model-version logging and its sufficiency fo...
Shadow AI — unsanctioned, decentralized use of GenAI tools — amplifies prompt-fraud risk by bypassing central controls and audit trails.
Conceptual analysis and organizational risk reasoning; references to common practices of unsanctioned tool use (no empirical prevalence data).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... increase in unmonitored prompt activity and corresponding reduction in detectabi...
External actors can commit prompt fraud via customer-facing systems or social-engineering prompt chains.
Conceptual threat scenarios and mapping of attack surfaces (customer-facing interfaces, input channels); illustrative examples provided.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... risk of prompt-fraud initiated through external-facing inputs or social-engineer...
Internal actors manipulating prompts within authorized AI workflows are a realistic and important threat vector for prompt fraud.
Threat modeling and scenario-based analysis highlighting insiders with authorized access who can craft prompts.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... risk or incidence of prompt-fraud events originating from internal actors
Prompt fraud can defeat controls that rely on plausibility, standard formatting, or human review that trusts model-like language.
Threat mapping and literature on automation bias; illustrative vignettes showing how machine-like outputs mimic authoritative formats.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... effectiveness of plausibility/format/human-review-based controls in identifying ...
Prompt fraud lowers the entry cost of producing convincing fraudulent artifacts, increasing the ease with which attackers can create plausible forgeries.
Economic reasoning and conceptual analysis based on GenAI behavior and illustrative scenarios (no empirical cost or frequency data).
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... marginal cost (effort/resources) required to produce convincing fraudulent artif...
Prompt fraud — the intentional manipulation of natural-language prompts to cause generative AI systems to produce misleading, fabricated, or deceptive artifacts that bypass internal controls — constitutes a novel, low-cost fraud vector that traditional IT- and process-focused controls are ill-equipped to detect or prevent.
Conceptual analysis and threat modeling grounded in literature/regulatory review and illustrative vignettes; no systematic empirical incidence data provided.
medium negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... ability of existing IT/process controls to detect or prevent fraud produced via ...
Secure infrastructure (including SECaaS-provided tools) affects the availability and trustworthiness of AI training data and models; breaches reduce returns to AI R&D via direct losses and reduced trust.
Conceptual linkage supported by case studies of data/model theft and technical literature on secure enclaves, differential privacy, federated learning; no broad quantitative estimate provided.
medium negative Security- as- a- service: enhancing cloud security through m... incidence of data/model breaches, economic returns to AI R&D
Security externalities (one firm's breach raising ecosystem risk) complicate private incentives and may justify policy interventions such as standards or mandatory reporting.
Economic theory on externalities, case studies showing spillovers from breaches, and policy analyses recommending interventions.
medium negative Security- as- a- service: enhancing cloud security through m... spillover risk, incentive alignment, justification for regulation
Concentration among large cloud/SECaaS providers can create market power, platform dependency, and affect competition in AI markets.
Market-structure theory, observed concentration patterns in industry reports, and qualitative case studies; no causal estimates provided in the chapter.
medium negative Security- as- a- service: enhancing cloud security through m... market power indicators, competition measures in AI markets
Latency and integration frictions can limit the suitability of SECaaS for specialized workloads, including some AI pipelines.
Technical evaluations and benchmarks that measure latency/resource overhead; reports and case studies noting integration challenges for high-throughput or low-latency workloads.
medium negative Security- as- a- service: enhancing cloud security through m... latency, integration overhead, suitability for AI workloads
Reliance on a small set of major cloud/SECaaS providers creates vendor lock-in, concentration risk, and systemic vulnerability if a major provider is compromised.
Market-structure discussions, observed provider outages and incidents (case studies), and theoretical arguments about concentration; no single causally identified empirical estimate provided.
medium negative Security- as- a- service: enhancing cloud security through m... market concentration, systemic risk, dependency risk
Without improvements in robustness, consistency, and neuroscientific validity of explanations, clinical uptake will be constrained, slowing commercialization and reducing returns for developers focused only on performance.
Synthesis and forward-looking argument linking methodological deficits documented in the literature to likely reduced market adoption; no direct empirical market impact measurement provided.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... clinical uptake, commercialization pace, developer returns
Weak or inconsistent explanations increase regulatory and medico-legal risk; standardized, validated XAI can lower compliance costs and liability exposure.
Logical inference connecting explanation reliability to regulatory scrutiny and liability concerns, presented as an implication in the review (no direct empirical legal analysis provided).
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... regulatory/compliance and legal risk
Preprocessing pipelines (filtering, artifact removal such as ICA, re-referencing, segmentation) materially affect XAI outputs.
Review cites multiple studies and methodological notes showing explanation maps vary with preprocessing choices; effect reported qualitatively across papers.
medium negative Explainable Artificial Intelligence (XAI) for EEG Analysis: ... sensitivity of explanation outputs to preprocessing steps