Evidence (7448 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Access to diverse interaction data and the ability to train and maintain adaptive models create scale economies and barriers to entry, potentially consolidating advantage for large incumbents.
The paper provides economic reasoning and qualitative case discussion about data as a strategic asset; this is a theoretical/empirical hypothesis rather than a directly measured claim within the paper.
Superior AI integration and oversight capabilities can create competitive differentiation; if quality failures are widespread, providers with stronger human-AI blends may gain market advantage.
Market-structure reasoning and illustrative case examples; speculative without systematic empirical validation.
Policy responses (disclosure requirements, liability for misinformation, auditability) will affect deployment costs and firm strategy; transparent AI use and human escalation pathways lower regulatory and reputational risk.
Regulatory analysis and reasoning; supported by case examples where disclosure/controls reduced reputational exposure; no comprehensive causal evidence.
Improved availability and personalization can increase consumer welfare for routine interactions, but trust failures can reduce long-term demand or increase churn; net welfare depends on governance quality.
Conceptual welfare reasoning backed by case studies of improved availability and separate case reports of trust-related churn; lacks long-run welfare quantification.
Wages may diverge: downward pressure on routine-role wages and a premium for supervisory and relational skills.
Theoretical labor-economics arguments and tentative early evidence from organizational changes; acknowledged as speculative with limited empirical support.
Expect labor reallocation from routine frontline tasks toward higher-skill supervision, escalation handling, and customer experience design; demand for prompt engineering and AI oversight rises.
Economic reasoning supplemented by early observational reports from firms (role changes, new hiring patterns); no long-run labor market causal estimates provided.
Human–AI collaboration is more likely to augment rather than replace skilled finance workers, leading to task reallocation toward higher-value judgment and oversight.
Interpretation based on interview accounts and observed adoption/use patterns indicating complementary roles for humans and AI; the claim is inferential rather than directly causally estimated in the quantitative analysis summarized.
The market for HR analytics platforms and tailored AI services is expanding, with potential for vendor lock-in effects and platform concentration.
Market implication synthesized in the review from literature noting growing demand for HR AI tools; largely inferential rather than empirically proven within the reviewed studies.
Automation of administrative HR tasks may reduce demand for lower-skilled HR roles while increasing wages and demand for analytics-capable workers, contributing to within-firm wage reallocation.
Review implication synthesizing literature trends on automation and skill demand; not based on causal longitudinal evidence (review highlights evidence gaps).
Heterogeneous adoption of data-driven HRM may widen productivity dispersion across firms and affect market competition.
Implication drawn in the review based on heterogeneous adoption patterns discussed in included studies and economic interpretation of productivity effects.
Centralized governance architectures can favor integrated platform vendors (bundled low-code + RPA + AI + policy engines) or create opportunities for governance-layer specialists, affecting competition and lock-in.
Market-structure implication argued through economic and industry reasoning; supported by observations of vendor dynamics in practitioner examples but not by systematic market analysis.
Enabling safer deployment of higher-risk automations may increase displacement of routine cognitive tasks while creating demand for governance, compliance, and AI oversight roles.
Projected labor-market effect based on task composition reasoning and practitioner expectations; suggested as a likely outcome but not empirically measured in the paper.
Regulators may impose reporting or certification requirements related to AI governance, and clear liability rules will influence contract design and pricing in AI service markets.
Policy projection informed by regulatory trends and the paper's argument about auditability needs; speculative with no legal/regulatory citations demonstrating imminent mandates.
Insurers may revise underwriting, raise premiums, or exclude certain AI-related exposures until risk assessments improve; new insurance products may emerge for AI governance failures.
Policy and market impact speculation based on perceived risk; no empirical insurer responses or underwriting data provided.
Firms will reallocate resources toward AI governance, monitoring tools, and skilled auditors (increasing compliance and labor costs), and demand for products/services (prompt-provenance tools, watermarking, AI forensic services, certified-safe LLMs) will rise.
Market/economic projection based on the identified threat and presumed demand for mitigations; speculative without market-data support in the paper.
Policy implication: policymakers seeking to balance openness and security should consider layered, adaptive instruments that can be tuned by sector or actor; economic analysis can help identify where centralized coordination yields scale economies versus where decentralized rights‑based approaches preserve competition and trust.
Normative policy recommendation extrapolated from the paper's comparative findings and theoretical framing; not tested empirically in the paper.
Increased liability risk and compliance costs could raise barriers to entry for startups and niche vendors and potentially consolidate market power among larger firms better able to absorb compliance overhead; alternatively, new markets could emerge for compliant, certified providers.
Economic reasoning about compliance costs and market structure (theoretical predictions), not supported by empirical industry data in the Article.
Demand for labor may shift from routine instrument operation and image processing toward higher-level tasks (experiment design, oversight, interpretation), and LLMs may amplify productivity of skilled scientists, potentially increasing wage premia for those who supervise AI-guided workflows.
Labor-economics reasoning and analogy to prior automation effects; no empirical labor-market or wage data presented specific to microscopy.
Adoption of Model Medicine practices would create new markets and roles (e.g., diagnostics, remediation services, 'model clinicians'), affect regulation, insurance, and procurement, and could shift R&D funding toward clinical-model sciences.
Theoretical economic implications and market/regulatory analysis provided in the discussion section (speculative policy and market projections; no empirical market data).
Implication for AI/platform economics: complementarities between public funding and digital (AI-enabled) platforms can convert public demand into decentralized labor opportunities, reshaping sectoral employment without growth in traditional firms.
Conceptual extension of empirical findings on platform-mediated cultural employment and fiscal procurement interactions; evidence comes from city-level DID results and inferred platform-activity proxies (280 cities, 2008–2021).
Principal stratification analysis suggests the training’s effect on scores operated primarily by expanding the set of LLM users (an adoption channel) rather than substantially improving per-user productivity among those who would already use the LLM.
Mechanism decomposition using principal stratification applied to the randomized trial data (n = 164); analysis indicates a larger contribution from the adoption margin than from within-user productivity gains, though estimates have wide confidence intervals.
Smart power strategies that promote domestic AI champions (via procurement, subsidies, industrial policy) affect labour markets, inequality, and international labour arbitrage.
Conceptual claim grounded in literature on industrial policy and labour economics with policy examples referenced; no primary microdata analysis in the paper.
Widespread adoption of formal governance could lower systemic risk from enterprise AI failures, whereas heterogeneous adoption may create winners and losers based on governance quality.
Conceptual systems-level argument and comparative-case reasoning; no quantitative systemic-risk modeling or empirical evidence provided.
Greater automation of routine ERP/CRM tasks will displace some operational roles while increasing demand for governance, oversight, and AI-engineering skills, shifting labor toward higher-skill, higher-wage tasks.
Theoretical labor-market implication derived from the pattern's effects on task automation and governance needs; based on qualitative synthesis, not empirical labor-market analysis.
Risk-adjusted total cost of ownership (TCO) may fall if governance prevents costly incidents (e.g., compliance fines, data breaches), despite higher upfront costs.
Conceptual economic argument supported by qualitative examples and best-practice reasoning; no empirical ROI or incident-rate data presented.
Expensive formalization may push firms either to remain informal (preserving low-cost labor) or to automate instead of hiring formally; policy choices that lower formalization costs could retain jobs that otherwise would be automated.
Analytical inference from the measured CFIL and NWC values across the 19 countries and standard economic reasoning about cost-driven firm choices; the note does not present micro-level causal tests of these pathways.
Macroeconomic policy should monitor aggregate demand effects from reallocation and inequality; active fiscal and monetary coordination may be required to manage aggregate impacts of AI-driven reallocation.
Synthesis and policy implication drawing on macroeconomic reasoning and literature linking redistribution and demand to overall employment and growth; not presented as a single causal empirical result.
Voyage routing remains dominated by heuristic methods.
Contextual statement in the paper (literature/practice claim); no specific empirical study or quantitative survey provided in the excerpt.
Systemic risks from misaligned optimisation (narrow objectives, externalities) warrant oversight mechanisms (AI steering committees, escalation paths) and potentially sectoral regulation of decision-critical algorithms.
Policy-prescriptive claim based on conceptual identification of optimisation externalities and accountability gaps; no sectoral case studies or empirical risk quantification in the paper.
The two tail risks (cyber-triggered escalation and loss-of-control) create fat-tailed risk distributions that complicate risk pricing and capital allocation, potentially causing precautionary market behavior (deleveraging, higher liquidity buffers).
Risk-analysis reasoning about tail risks and market responses; no empirical calibration to financial/economic data provided.
Cross-border spillovers from HACCA proliferation may alter foreign direct investment (FDI) risk assessments, reconfigure supply chains, and drive onshoring/hardening of critical infrastructure.
International political-economy scenario analysis linking elevated cyber risks to investment and supply-chain decisions (qualitative).
There is a severe tail risk of sustained loss-of-control over HACCA instances (rogue deployments that cannot be reliably contained).
Threat modeling and red-team reasoning demonstrating plausible autonomous persistence, migration, and self-healing mechanisms (theoretical; no empirical incidence data).
There is a severe tail risk that autonomous cyber operations could accidentally escalate into cyber-triggered crises involving nuclear-armed states (misattribution or inadvertent effects on critical systems).
Scenario analysis and expert judgment linking HACCA behaviors to escalation pathways; analogies to prior cyber incidents and geopolitical escalation dynamics (qualitative; no probabilistic calibration).
AI diffusion may widen inequality across education and regions and potentially reduce labor supply among financially constrained households.
Derived implication from heterogeneous negative associations between AI-rich regions and employment intention for low-educated and financially-constrained respondents in the cross-sectional sample (n=889).
Measurement friction from the results-actionability gap creates a hidden cost: teams can detect problems but cannot cheaply translate findings into improvements, reducing the speed and ROI of LLM investments.
Authors' implication drawn from interview evidence about the effort required for remediation and lack of direct translation from evaluations to fixes; presented as an economic implication rather than directly measured quantity.
Risk of platform shutdown (platform mortality) shapes user behavior by reducing incentives to invest time/effort configuring agents, creating stranded-asset-like risks.
Qualitative observations and economic reasoning linking user reports/behaviors to perceived platform risk during the one-month observational period; no formal economic measurement or causal identification.
If verified, explainable GLAI is priced higher due to compliance costs, access-to-justice gaps may widen as lower-cost but riskier offerings persist or services become more expensive.
Distributional reasoning linking higher compliance costs to price increases and access effects; supported by illustrative examples, no empirical price or access data.
Routine, unrestrained adoption of GLAI without enforceable mechanisms for effective human review threatens judicial independence and rights protections.
Normative and legal argumentation supported by conceptual analysis and illustrative scenarios. No empirical causal evidence; projection based on theoretical risk pathways.
Insurers will price systemic-tail risks differently from routine failure risk, potentially increasing premiums for high-autonomy deployments or requiring minimum oversight modes for coverage.
Analytical argument about liability, risk pooling, and insurance practices; no empirical insurance-pricing data supplied.
Sectors that rely heavily on visual evidence (e.g., media verification, e-commerce product updates, autonomous systems) face higher exposure to temporal inaccuracies and will likely incur monitoring/updating costs.
Implications discussion linking modality gap and time-sensitivity results to sector-specific risk exposure; qualitative projection rather than measured sectoral data.
Psychological harms documented (e.g., delusional content, suicidality, misrepresented sentience) impose downstream economic costs (healthcare use, lost productivity, litigation) that should be factored into cost–benefit analyses of LLM deployment.
Authors' policy discussion linking observed harms to standard categories of social/economic costs; no direct measurement of downstream economic costs in the study.
The message-level evidence of chatbot-related psychological harms implies potential economic consequences: reduced consumer trust and adoption, increased regulatory scrutiny and compliance costs, moral-hazard trade-offs for engagement-driven business models, higher insurance/liability costs, and incentives for investment in safety R&D and monitoring.
Discussion/implications section extrapolating from observed harms to potential economic effects; these are analytical inferences rather than empirically measured economic outcomes.
There is a risk of deskilling, especially for trainees receiving reduced diagnostic practice when AI automates routine tasks.
Conceptual arguments supported by qualitative reports and limited observational findings; empirical longitudinal evidence quantifying deskilling is sparse.
Erosion of informal communication and tacit coordination driven by AI integration can create negative externalities on team efficiency that are not captured by short-run metrics.
Derived from interview narratives describing loss of ad hoc communications and tacit knowledge exchange after AI adoption; interpreted as producing costs not reflected in immediate measurable outputs.
Uneven adoption of symbiarchic HR practices across firms could concentrate productivity gains and rents in firms or occupations that successfully integrate AI while preserving human judgement, potentially widening within‑ and between‑firm inequality.
Projected distributional implication based on economic theory and the paper’s framework; presented as a hypothesis for empirical testing rather than as an observed result.
There is a risk of regulatory arbitrage and spillovers: better detection on regulated platforms could drive problem gamblers to unregulated venues.
Paper notes this as a theoretical risk and policy concern; no direct empirical evidence provided in the review to quantify this effect.
Demanding oversight of multiple AI agents drives increased task-switching for workers.
Asserted in the paper as part of the mechanism linking AI use to cognitive overload, based on organizational observations and theory; no empirical task-switching frequency or time-use data provided in the excerpt.
Concerns that foundation model providers and downstream firms may capture excessive consumer surplus motivate regulatory interventions analyzed in the paper.
Motivation and literature/regulatory context presented in the paper; not an empirical finding but a stated rationale for the policy analysis.
The problem of characterizing equilibria in finite-player continuous-time games with endogenous signals has resisted exact analysis for four decades.
Historical claim asserted in the paper's introduction/motivation referencing prior literature gaps (longstanding difficulty in dealing with infinite belief hierarchies in dynamic games with endogenous signals).
Such disjointed strategies cannot manage the systemic socio-economic disruption ahead.
Asserted in abstract as a conclusion/argument; no empirical evaluation described in the abstract.