The Commonplace
Home Dashboard Papers Evidence Digests 🎲

Evidence (2340 claims)

Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims

Evidence Matrix

Claim counts by outcome category and direction of finding.

Outcome Positive Negative Mixed Null Total
Other 378 106 59 455 1007
Governance & Regulation 379 176 116 58 739
Research Productivity 240 96 34 294 668
Organizational Efficiency 370 82 63 35 553
Technology Adoption Rate 296 118 66 29 513
Firm Productivity 277 34 68 10 394
AI Safety & Ethics 117 177 44 24 364
Output Quality 244 61 23 26 354
Market Structure 107 123 85 14 334
Decision Quality 168 74 37 19 301
Fiscal & Macroeconomic 75 52 32 21 187
Employment Level 70 32 74 8 186
Skill Acquisition 89 32 39 9 169
Firm Revenue 96 34 22 152
Innovation Output 106 12 21 11 151
Consumer Welfare 70 30 37 7 144
Regulatory Compliance 52 61 13 3 129
Inequality Measures 24 68 31 4 127
Task Allocation 75 11 29 6 121
Training Effectiveness 55 12 12 16 96
Error Rate 42 48 6 96
Worker Satisfaction 45 32 11 6 94
Task Completion Time 78 5 4 2 89
Wages & Compensation 46 13 19 5 83
Team Performance 44 9 15 7 76
Hiring & Recruitment 39 4 6 3 52
Automation Exposure 18 17 9 5 50
Job Displacement 5 31 12 48
Social Protection 21 10 6 2 39
Developer Productivity 29 3 3 1 36
Worker Turnover 10 12 3 25
Skill Obsolescence 3 19 2 24
Creative Output 15 5 3 1 24
Labor Share of Income 10 4 9 23
Clear
Org Design Remove filter
Policy implication: policymakers seeking to balance openness and security should consider layered, adaptive instruments that can be tuned by sector or actor; economic analysis can help identify where centralized coordination yields scale economies versus where decentralized rights‑based approaches preserve competition and trust.
Normative policy recommendation extrapolated from the paper's comparative findings and theoretical framing; not tested empirically in the paper.
low mixed Balancing openness and security in scientific data governanc... policy design effectiveness (layered/adaptive instruments), trade‑offs between s...
Demand for labor may shift from routine instrument operation and image processing toward higher-level tasks (experiment design, oversight, interpretation), and LLMs may amplify productivity of skilled scientists, potentially increasing wage premia for those who supervise AI-guided workflows.
Labor-economics reasoning and analogy to prior automation effects; no empirical labor-market or wage data presented specific to microscopy.
low mixed ChatMicroscopy: A Perspective Review of Large Language Model... labor demand composition, distribution of wages, skill premium
Widespread adoption of formal governance could lower systemic risk from enterprise AI failures, whereas heterogeneous adoption may create winners and losers based on governance quality.
Conceptual systems-level argument and comparative-case reasoning; no quantitative systemic-risk modeling or empirical evidence provided.
low mixed Governed Hyperautomation for CRM and ERP: A Reference Patter... systemic risk of enterprise AI failures and competitive market outcomes
Greater automation of routine ERP/CRM tasks will displace some operational roles while increasing demand for governance, oversight, and AI-engineering skills, shifting labor toward higher-skill, higher-wage tasks.
Theoretical labor-market implication derived from the pattern's effects on task automation and governance needs; based on qualitative synthesis, not empirical labor-market analysis.
low mixed Governed Hyperautomation for CRM and ERP: A Reference Patter... changes in labor demand by skill level, displacement of routine roles, increased...
Risk-adjusted total cost of ownership (TCO) may fall if governance prevents costly incidents (e.g., compliance fines, data breaches), despite higher upfront costs.
Conceptual economic argument supported by qualitative examples and best-practice reasoning; no empirical ROI or incident-rate data presented.
low mixed Governed Hyperautomation for CRM and ERP: A Reference Patter... risk-adjusted TCO and incident-related cost savings
Systemic risks from misaligned optimisation (narrow objectives, externalities) warrant oversight mechanisms (AI steering committees, escalation paths) and potentially sectoral regulation of decision-critical algorithms.
Policy-prescriptive claim based on conceptual identification of optimisation externalities and accountability gaps; no sectoral case studies or empirical risk quantification in the paper.
low negative Comparative analysis of strategic vs. computational thinking... systemic risk exposure and effectiveness of oversight/regulatory mechanisms
Measurement friction from the results-actionability gap creates a hidden cost: teams can detect problems but cannot cheaply translate findings into improvements, reducing the speed and ROI of LLM investments.
Authors' implication drawn from interview evidence about the effort required for remediation and lack of direct translation from evaluations to fixes; presented as an economic implication rather than directly measured quantity.
low negative Results-Actionability Gap: Understanding How Practitioners E... inferred effect on ROI and speed of product improvement
Erosion of informal communication and tacit coordination driven by AI integration can create negative externalities on team efficiency that are not captured by short-run metrics.
Derived from interview narratives describing loss of ad hoc communications and tacit knowledge exchange after AI adoption; interpreted as producing costs not reflected in immediate measurable outputs.
low negative AI in project teams: how trust calibration reconfigures team... team efficiency and unmeasured coordination/tacit work
Uneven adoption of symbiarchic HR practices across firms could concentrate productivity gains and rents in firms or occupations that successfully integrate AI while preserving human judgement, potentially widening within‑ and between‑firm inequality.
Projected distributional implication based on economic theory and the paper’s framework; presented as a hypothesis for empirical testing rather than as an observed result.
low negative Symbiarchic leadership: leading integrated human and AI cybe... within‑ and between‑firm inequality; distribution of productivity rents
Demanding oversight of multiple AI agents drives increased task-switching for workers.
Asserted in the paper as part of the mechanism linking AI use to cognitive overload, based on organizational observations and theory; no empirical task-switching frequency or time-use data provided in the excerpt.
low negative When AI Assistance Becomes Cognitive Overload: Understanding... task-switching frequency / oversight burden
Concerns that foundation model providers and downstream firms may capture excessive consumer surplus motivate regulatory interventions analyzed in the paper.
Motivation and literature/regulatory context presented in the paper; not an empirical finding but a stated rationale for the policy analysis.
low negative The Economics of AI Supply Chain Regulation consumer surplus (potential capture by firms)
The problem of characterizing equilibria in finite-player continuous-time games with endogenous signals has resisted exact analysis for four decades.
Historical claim asserted in the paper's introduction/motivation referencing prior literature gaps (longstanding difficulty in dealing with infinite belief hierarchies in dynamic games with endogenous signals).
low negative Forecasting and Manipulating the Forecasts of Others historical status of the literature (difficulty/resistance to exact analysis)
Organizational compliance, governance, and transaction costs shape which AI uses are feasible, producing heterogeneity in adoption across firms; trust and accountability frictions can slow adoption even when productivity gains exist.
Workshop participants (n=15) reported compliance and governance considerations; authors infer broader organizational heterogeneity and friction effects from these qualitative data.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... adoption heterogeneity across firms; adoption speed/timing affected by governanc...
Designers’ expressed concerns about skill development suggest potential long-term effects on human capital accumulation; adoption that reduces learning opportunities could lower future wages or employability.
Participants' concerns captured in qualitative workshops (n=15); claim is an extrapolation to labor-market outcomes rather than direct measurement in the study.
low negative The Values of Value in AI Adoption: Rethinking Efficiency in... human capital accumulation; future wages; employability (hypothesized)
Legacy systems and siloed incentives create switching frictions that slow diffusion of AI-enabled ISP; early adopters may achieve sustained cost and service advantages and vendors bundling technology with change management could capture large rents.
Authors' argument informed by case observations of switching costs and vendor roles; no causal market-level evidence provided.
low negative Optimizing integrated supply planning in logistics: Bridging... adoption rate, market concentration, vendor rents
Returns to AI investments may exhibit increasing returns to scale, reinforcing winner‑take‑most dynamics unless offset by platformization or open‑source diffusion.
Economic scenario reasoning on capital intensity and platform effects; no empirical calibration or econometric evidence provided.
low negative How AI Will Transform the Daily Life of a Techie within 5 Ye... return on AI investment by firm size (evidence of increasing returns to scale) a...
Private governance and firm-level solutions (internal standards, bargaining with unions) may proliferate, but these can entrench firm-specific norms and increase market power asymmetries.
Conceptual argument drawing on governance and industrial organization literature; no empirical measurement of prevalence or market-power effects included.
low negative AI governance under the second Trump administration: implica... prevalence of private governance; firm-specific norms; market power asymmetries
Legal liability and cyber-insurance markets will need to adapt as machine-generated code becomes pervasive, with pricing internalizing risk from inadequate verification processes.
Speculative legal/economic implication discussed in the paper; no actuarial or legal-case data provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... insurance pricing changes; liability claims tied to machine-generated code
Individual developers or firms may underinvest in verification because defect accumulation imposes external costs on downstream actors, creating market failures that can justify standards, certifications, or regulation mandating interlocks or minimum verification practices.
Policy and market-failure argument based on externalities presented conceptually; no modeling or empirical evidence of such externalities provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... degree of underinvestment in verification; incidence of downstream costs/externa...
Short-run productivity gains from generative AI may be offset by longer-run increases in maintenance, security breaches, and reliability costs if verification lags.
Economic reasoning and forward-looking implications discussed in the paper; no empirical cost-benefit or longitudinal data presented.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... net productivity over time; maintenance/security costs versus short-term product...
Small, unverified errors, insecure patterns, and brittle interactions accumulate over time (latent accumulation), increasing operational fragility and long-run maintenance costs.
Theoretical argument and illustrative examples in the paper; no longitudinal defect accumulation studies or empirical cost analysis provided.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... rate of latent defect accumulation; long-run maintenance and reliability costs
Time pressure and productivity incentives lead developers to accept plausible AI outputs without full validation, a behavioral/institutional failure mode called the 'micro-coercion of speed' that effectively reverses the burden of proof.
Behavioral diagnosis and incentive analysis presented conceptually in the paper; no behavioral experiments, surveys, or observational data reported.
low negative Overton Framework v1.0: Cognitive Interlocks for Integrity i... developer acceptance rate of AI outputs without full validation / shift in burde...
Hallucination and error risk introduce potential liabilities in client engagements and may change contracting, insurance, and pricing practices in consulting services.
Derived from practitioner concerns reported in interviews and authors' normative discussion; no contractual or insurance-market data presented.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... liability exposure; contracting/insurance practices; pricing adjustments
Effective deployment requires governance, verification processes, and liability management to manage hallucination risk, creating adoption costs that may advantage larger firms and affect market concentration and pricing power.
Argument based on interviews about necessary organizational safeguards and the resource requirements to implement them; speculative market-structure implications are not empirically tested in the paper.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... adoption costs; firm-level resource burden; changes in market concentration/pric...
Widespread GenAI use may accelerate skill obsolescence for routine competencies and increase the premium on monitoring, critical evaluation, and AI‑integration skills, shifting investment toward retraining and upskilling.
Projection based on qualitative interviews and the authors' economic interpretation of TGAIF; no longitudinal or wage/skill data provided.
low negative Where Automation Meets Augmentation: Balancing the Double-Ed... skill obsolescence rates; demand for monitoring/evaluation/AI-integration skills...
Uncertainty about long-run agentic behavior increases option value and downside risk of investing in agentic systems, which may raise discount rates and required returns.
Economic argument applying risk/return logic to agentic uncertainty; no quantitative empirical evidence provided.
low negative Visioning Human-Agentic AI Teaming: Continuity, Tension, and... investment valuation metrics (discount rates, required returns) for agentic syst...
Economic rents and advantages may accrue to agents who control large datasets, computing resources, and organizational processes that effectively integrate AI as a co-pilot, potentially increasing market concentration among AI providers.
Economic theory on scale economies and platform effects combined with observed industry patterns; reviewed literature provides conceptual arguments and case examples rather than broad empirical market-structure measurement.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... market concentration measures; returns to data/compute ownership (not fully meas...
Generative AI poses substitution risk for entry-level or routine cognitive work focused on generation or drafting without evaluative responsibility.
Task-based analyses and case studies indicating automation potential for routine generation tasks; empirical demonstrations of AI-produced drafts/outputs that could replace such work, but longer-run displacement evidence is limited.
low negative ChatGPT as an Innovative Tool for Idea Generation and Proble... task automatability; employment/demand for routine-generation roles (largely unm...
There is a risk of deskilling through excessive reliance on AI, implying a need for continuous training and certification to preserve human judgment.
Qualitative interview evidence and observed concerns about overreliance; authors recommend training/governance based on identified risks; no direct longitudinal measurement of deskilling provided in summary.
low negative Human-AI Synergy in Financial Decision-Making: Exploring Tru... human skill levels (deskilling risk); need for training/certification
Insurance markets may price AI-specific fraud risk, raising premiums or creating new products (AI-fraud insurance).
Speculative economic implication suggested by the authors; no market data or insurer statements cited.
low negative Prompt Engineering or Prompt Fraud? Governance Challenges fo... changes in insurance pricing or product offerings attributable to AI-specific fr...
Vendors offering integrated governed hyperautomation stacks may capture premium pricing and increase switching costs, potentially widening adoption gaps between large incumbents and SMEs.
Market-structure and competitive dynamics discussed theoretically in the Implications section; no market-share or pricing data provided.
low negative Governed Hyperautomation for CRM and ERP: A Reference Patter... vendor pricing premiums; switching costs; differential adoption by firm size (ma...
Regulators and standard-setters who value transparency and auditability will need to account for the gap between evaluation results and actionable fixes; firms may require incentives or rules to ensure evaluation leads to remediation, not just documentation.
Authors' policy implication derived from the study's finding of a results-actionability gap and discussion of auditability concerns; speculative recommendation rather than empirical finding.
low neutral Results-Actionability Gap: Understanding How Practitioners E... policy/regulatory effectiveness regarding evaluation leading to remediation (spe...
Delegation of oversight and reallocation of monitoring tasks due to AI integration changes transaction costs and affects organizational design and governance needs (e.g., more verification/audit effort or specialist oversight roles).
Based on participants' reported shifts in who performed monitoring/oversight tasks in the 40 interviews and the authors' interpretation of those shifts in organizational/economic terms.
low neutral AI in project teams: how trust calibration reconfigures team... transaction/monitoring costs and governance arrangements
Observable firm-level and economy-wide moments—changes in spans of control, manager share of payroll, incidence of new tasks, employment growth, and shifts in the wage distribution—can be used to test the model's predictions.
Model-implied empirical identification strategy and suggested measurable moments in the paper's discussion/implications section (theoretical prediction, not an empirical test).
low null result AI as Coordination-Compressing Capital: Task Reallocation, O... empirical testable moments (spans of control, manager payroll share, new-task in...
Expect rising demand and wage premia for managers with hybrid capabilities (systems thinking + computational literacy), with a risk of widening returns to managerial skill heterogeneity.
Theoretical implication from predicted complementarities and task reallocation; prescriptive economic inference without empirical labor-market evidence in the paper.
low positive Comparative analysis of strategic vs. computational thinking... labor demand, wage premia, and distributional widening across managerial skill t...
Managers’ time will be reallocated toward hybrid tasks (interpretation, oversight, ethical deliberation), increasing returns to combined strategic and computational skills.
Predictive inference from the role reconfiguration analysis and task-complementarity argument; forward-looking theoretical forecast (no empirical time-use data).
low positive Comparative analysis of strategic vs. computational thinking... managerial time allocation (share devoted to hybrid tasks) and returns/wage prem...
Regulators should consider guidelines on AI monitoring, algorithmic fairness in performance evaluations, and protections to prevent hybrid‑induced career penalties.
Policy recommendation based on conceptual assessment of risks identified in literature synthesis; not an empirical claim—no policy evaluation data provided.
low positive The Sociology of Remote Work and Organisational Culture: How... existence/applicability of regulatory guidelines; protections against career pen...
Hybrid agency implies complementarity between GenAI and managerial/knowledge‑worker skills (curation, evaluation, coordination), potentially increasing returns to those skills while automating routine cognitive tasks—consistent with skill‑biased technological change.
Synthesis of recurring themes linking GenAI capabilities with managerial skill topics in the thematic clusters; positioned as an implication for labour demand and skill composition rather than an empirically tested effect.
low positive Generative AI and the algorithmic workplace: a bibliometric ... expected changes in returns to managerial/knowledge‑worker skills and automation...
There is demand for tooling that bridges evaluation outputs to actionable fixes (e.g., failure-mode libraries, standardized remediation templates, evaluation-to-priority mapping), signaling economic opportunities for third-party tools and consulting services.
Authors' inference based on the documented results-actionability gap and participants' descriptions of pain points; presented as a market implication rather than direct market measurement.
low positive Results-Actionability Gap: Understanding How Practitioners E... inferred market demand for evaluation-to-action tooling/services
Firms that invest in instrumentation, cross-functional processes, and remediation levers capture more value from LLMs; organizations with better evaluation-to-action pipelines will obtain higher productivity gains and market edge.
Authors' inference from observed heterogeneity among teams in the interviews and comparison of practices in teams that reported more success converting evaluations into changes.
low positive Results-Actionability Gap: Understanding How Practitioners E... relative productivity/value capture tied to evaluation-to-action capability (inf...
Structured errors (SERF) enable automated recovery, reducing human-in-the-loop remediation and the marginal cost of scaling agent fleets.
Reasoned implication from the design of SERF; proposed as an expected operational benefit rather than demonstrated quantitative result in the summary.
low positive Bridging Protocol and Production: Design Patterns for Deploy... human remediation hours per incident; MTTR; automated recovery success rate
Adaptive budgeting (ATBA) can reduce wasted latency and cost by optimizing timeouts and retries across tool chains, improving throughput and reducing per-interaction resource spend.
Algorithmic claim supported by theoretical framing and proposed reproducible benchmarks; no concrete field-level cost/throughput numbers provided in the summary.
low positive Bridging Protocol and Production: Design Patterns for Deploy... per-interaction latency/cost, throughput, retry rates under ATBA vs. baseline
Improved identity propagation (via CABP) reduces risk and compliance costs by lowering misattributed actions and improving audit trails, thereby reducing expected liability and incident-resolution overhead.
Analytical / economic argument in the implications section; no reported quantitative field results in the summary to directly measure cost reduction.
low positive Bridging Protocol and Production: Design Patterns for Deploy... incidence of misattributed actions; audit trail completeness; incident-resolutio...
Organizational norms and UX influence adoption rates and diffusion of AI: social calibration processes at the team level matter for adoption beyond individual cost–benefit calculations.
Reported by interviewees (N=40) as factors shaping whether and how teams incorporated AI into routines; integrated into theoretical implications for diffusion modeling.
low positive AI in project teams: how trust calibration reconfigures team... AI adoption/diffusion rates at team/organization level
Well-calibrated trust tends to encourage AI being used as a complement to human labor (augmentation), increasing effective productivity; miscalibration (over- or under-trust) can lead to productivity losses.
Inferential claim drawn from interviewees' accounts of when teams appropriately relied on AI (augmentation) versus when inappropriate reliance or avoidance occurred; supported by thematic interpretation rather than quantitative measurement.
low positive AI in project teams: how trust calibration reconfigures team... productive use of AI (complementarity vs substitution) and effective productivit...
Policymakers should support standards for auditability, human‑in‑the‑loop thresholds and training subsidies to reduce coordination failures and make the social benefits of AI adoption more widely shared.
Normative policy recommendation derived from the paper’s analysis of risks, governance needs and distributional concerns; not empirically validated within the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... adoption of standards; breadth of social benefits; coordination failure reductio...
Organisations will invest more in training for AI‑related sensemaking, trust calibration and governance competencies; returns to such training should be evaluated relative to investments in model quality.
Prescriptive inference from the framework and human‑capital theory; supported by referenced literature but not empirically tested in this paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... training investment levels; returns on training; comparative returns vs model in...
Explicit comparative‑advantage allocation will shift the composition of tasks across humans and AI, altering demand for routine versus non‑routine skills and potentially increasing demand for high‑level judgement, oversight and sensemaking skills.
Projected labour‑market implication based on theoretical reasoning and prior literature on task‑based skill demand; not empirically estimated in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... task composition; demand for routine vs non‑routine skills; demand for oversight...
Operationalising the four symbiarchic practices through updated HR systems lets firms capture AI‑enabled productivity gains without eroding trust, ethics or employee well‑being.
Normative claim based on theoretical synthesis and managerial prescription; no empirical testing or field data presented in the paper.
low positive Symbiarchic leadership: leading integrated human and AI cybe... AI‑enabled productivity gains; employee trust; ethical outcomes; employee well‑b...
Public data sharing, reproducibility standards, and shared benchmarks could raise the floor of AI utility across the industry.
Policy implication grounded in arguments about data quality, coverage, and generalizability from the narrative review; speculative recommendation rather than evidence-backed empirical claim.
low positive Learning from the successes and failures of early artificial... baseline AI performance/utility across firms (industry-wide)