Evidence (2340 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Org Design
Remove filter
Embedding governance reduces downside risks (compliance fines, data breaches), improving expected net returns of automation investments and lowering the adoption threshold for risk-averse firms.
Conceptual cost-benefit argument and industry best-practice examples; lacking quantitative measurement of returns or threshold shifts.
Authors propose the 'AI orchestra' concept: future development will involve coordinated ensembles of specialized AI agents (code generation, test generation, dependency analysis, security scanning) orchestrated by humans and higher-level controllers.
Theoretical/conceptual argument by the authors grounded in qualitative findings from Netlight (practitioner reports of multiple tools and coordination frictions); this is a forward-looking synthesis rather than an empirically established fact.
Canvas Design Principles aimed at reducing algorithmic myopia matter for welfare and regulatory concerns: better adaptive behavior reduces mispricing/misattribution risks but raises questions about transparency, accountability, and systemic amplification of shocks.
Policy and governance implication inferred from the claimed reductions in algorithmic myopia and increased adaptivity; study does not report direct welfare/regulatory impact measurements.
Faster, more accurate identification of demand shifts can compress the window for first‑mover advantages, intensify competitive dynamics, and raise the premium on organizational agility and human–AI integration capabilities.
Theoretical implication derived from observed improvements in signal detection (~5.8×) and resilience; not directly measured as market‑level competitive outcomes in the study.
Product teams evaluating LLM-powered features rely on a spectrum of practices—from informal “vibe checks” to organizational meta-work—to cope with LLMs’ unpredictability.
Qualitative interview study with 19 practitioners; thematic coding of transcripts produced descriptions of a range of evaluation practices used by teams.
Improved throughput and lower travel costs can induce additional travel demand (rebound), partially offsetting congestion/emissions gains unless paired with demand-management measures.
Theoretical economic reasoning presented in the paper as a caveat; not directly measured in the simulation experiments (no induced-demand dynamic experiments reported).
There is a social welfare trade‑off between personalization value (higher AAR) and normative/social risk (higher MR); optimal policy and product design should balance these using BenchPreS metrics.
Analytical argument combining empirical findings (trade‑off between AAR and MR) with economic welfare considerations; the paper does not present formal welfare estimates or market experiments.
Organizational heterogeneity in strategic backing and mentoring explains variation in benefits from AI adoption across firms and sectors, contributing to cross-firm productivity dispersion.
Theoretical claim linking organizational moderators to heterogeneous adoption outcomes; proposed as an empirical research direction without data provided.
Managerial and peer mentoring styles (e.g., directive vs. developmental mentoring) influence how affordances are perceived and actualized, affecting learning, trust, and task allocation in human–AI collaboration.
Theoretical argument drawing on mentoring and organizational behavior literatures integrated with AST/AAT; no empirical tests or sample presented.
Organizational forms may shift (e.g., flatter, more modular organizations; increased platform-mediated teams) because easier global coordination changes the cost-benefit calculus for outsourcing and insourcing.
Conceptual mapping from reduced coordination costs to organizational design implications and illustrative examples; no firm-level empirical case studies or panel data presented.
AI-mediated reduction in language frictions could compress wage premia tied to language skills, reduce demand for pure translation/transcription roles, and increase demand for AI-supervisory, verification, and model-prompting roles.
Theoretical labor-market implications and illustrative scenarios linking reduced language frictions to labor supply/demand shifts; no empirical labor-market analysis or sample data included.
The adoption of AI governance programmes by military institutions will have strategic implications.
Hypothesis stated by the author; presented as forward-looking analysis without accompanying empirical modeling, historical analogues, or measured strategic outcomes in the provided text.
Findings have important implications for enterprise strategy and economic policy in early-stage AI adoption environments.
Discussion and policy implications drawn from the paper's theoretical framework and empirical results; not tested empirically within the paper.
Women in Ireland use advanced digital skills at rates broadly comparable to women elsewhere in Europe; Ireland's large gender gap instead reflects particularly high rates of advanced digital task use among men.
Cross-country comparison of female rates of advanced digital task use in ESJS descriptive tables; comparison highlights that Irish female rates are similar to European female averages while Irish male rates are unusually high.
Differences in observable worker and job characteristics (education, field of study, occupation, sector) explain only a minority of the Europe-wide gender gap in advanced digital task use, accounting for around 30% on average.
Decomposition analysis (e.g., Oaxaca–Blinder style) applied to ESJS data to partition the gender gap into explained (observable characteristics) and unexplained components. (Exact sample sizes by subgroup not reported in excerpt.)
Lower barriers to producing design concepts with GenAI could enable more freelancing and entry by non-traditional providers, altering market structure and intensifying competition at the lower end of the value chain.
Speculative implication extrapolated from interview findings and economic reasoning in the paper; not empirically tested within the study.
Demand for designers will likely shift toward individuals combining domain expertise with algorithmic/AI fluency (prompting strategies, tool orchestration), potentially increasing returns to these hybrid skills.
Inference and implication drawn from interview themes about algorithmic thinking and authors' policy/economics discussion; not empirically tested in study.
Adoption of advanced simulation and AI could affect productivity, returns to capital versus labor, trade and outsourcing patterns, and distributional outcomes, with benefits potentially concentrated among large firms.
Theoretical implications and discussion in the paper's AI economics section; framed as suggested areas for future study rather than empirically established effects.
Reported pilot gains, if scaled, could shift firm‑level returns and industry productivity measures, but gains are contingent on coordinated adoption; uneven uptake may produce winner‑takes‑more dynamics among technologically advanced firms.
Inference from pilot results and economic reasoning in the reviewed literature; no large‑scale empirical validation provided in the review.
Topology is the dominant factor for price stability and scalability compared to other swept variables (load, presence of hybrid integrator, governance constraints).
Factor-ablation analysis within the 1,620-run simulation study showing the largest explanatory effect (largest changes in volatility and scalability metrics) attributable to graph topology rather than load, hybrid flag, or governance settings.
Adoption heterogeneity may widen productivity dispersion across firms and contribute to market concentration, since organizations with better data, processes, and training budgets will capture more benefit.
Economic interpretation of literature and survey findings; speculative projection rather than empirical measurement within the study.
Societal acceptance of AI-generated audiovisual media is uncertain and could range from widespread uptake to broad rejection.
Discussion drawing on mixed empirical studies and scenario construction in the review; the paper notes contradictory findings in existing studies but does not provide primary survey data or sample sizes.
If cognitive interlocks are widely adopted, many negative externalities can be internalized and AI-driven productivity gains can be realized more sustainably; absent such controls, equilibrium may drift toward higher error rates and systemic incidents.
Long-run equilibrium argument based on theoretical reasoning and conditional claims; no longitudinal or cross-firm empirical evidence presented.
If AI raises the quality and pace of research, social returns to public research funding could increase, but distributional concerns and negative externalities must be managed to realize aggregate welfare gains.
Welfare implication discussed in the paper. Framed as conditional and theoretical; not empirically quantified in the abstract.
Policy interventions (data governance, transparency, reproducibility standards, ethical guidelines) will shape adoption and externalities (misinformation, misuse, reproducibility crises).
Policy recommendation/implication stated in the paper. This is a normative and predictive claim grounded in governance literature; the abstract does not present empirical evaluation of specific policies.
The effectiveness of generative AI depends critically on human-AI workflows: prompt design, iterative refinement, and human vetting materially affect outcomes.
Qualitative analyses of interaction patterns and experiments manipulating prompting/iteration showing variation in outcomes; many studies report improved outputs after iterative prompting and human-in-the-loop refinement.
Integrated ERP vendors embedding AI could strengthen vendor lock-in, while interoperable AI layers may foster ecosystems and specialized entrants; empirical work is needed to determine market outcomes.
Conceptual discussion and observed vendor behavior in practitioner literature; explicit statement in the paper that empirical analysis is required.
Persistent declines in self-efficacy after passive AI exposure suggest potential for skill atrophy and slower reversion when tasks must be performed without AI.
Inference from observed persistent reductions in self-efficacy post-return in the experiment; skill atrophy and reversion costs not directly measured—this is an implied consequence.
Firms that adopt passive, copy-based AI workflows risk psychological costs that could offset short-run productivity gains from AI.
Inference drawn from experimental findings of reduced efficacy/ownership/meaningfulness under passive use and short-term enjoyment gains; not directly tested for firm-level productivity or turnover—extrapolation from individual-level psychological measures.
Teams often produce evaluation outputs (tests, metrics, user feedback) but lack mechanisms, processes, or technical levers to convert those outputs into actionable engineering or product changes—a novel “results-actionability gap.”
Recurring theme from the 19 practitioner interviews and coding; authors explicitly articulate and label this gap based on participants' reports.
The study confirms several previously documented evaluation challenges with LLMs: model unpredictability, metric mismatch, high human-evaluation costs, and difficulty reproducing failures.
Interview data from 19 practitioners; thematic analysis flagged these recurring problems as reported by participants and aligned with prior literature.
Security of LLM-based MASs functions as an economic externality: failures can impose social costs (misinformation, poor collective decisions), and absent liability or market incentives providers may underinvest in robustness.
Economic reasoning and implication section in the paper—conceptual argument linking the technical vulnerability to economic externality and incentive misalignment. No empirical economic data provided in the summary.
Analytical conditions on stubbornness and influence weights identify when a single adversary can dominate network dynamics (i.e., influence propagation criteria derived from FJ fixed-point analysis).
Mathematical/theoretical analysis of FJ model fixed points and influence propagation in the paper; derivation of conditions relating agent stubbornness and interpersonal trust weights to steady-state influence.
If models frequently leak or misuse preferences in third‑party contexts, users and organizations will discount the value of personalization or demand stronger controls, increasing costs for deploying memory features and reducing consumer surplus.
Economic reasoning and implication drawn from the observed misapplication behavior; no empirical user adoption or market data provided in the study to directly support this claim.
The failure mode (misapplication of preferences to third parties) creates negative externalities (privacy violations, normative harms, misinformation, contractual breaches) that markets and platforms may not internalize without regulation or design changes.
Economic interpretation and argumentation building on the empirical failure mode; these harms are hypothesized implications rather than measured outcomes in the paper.
Uneven organizational supports can concentrate returns to AI in firms and workers that successfully actualize affordances, potentially widening wage and employment disparities; targeted policy and training investments can mitigate these effects.
Theoretical implication from the framework with policy recommendations; no empirical testing or sample reported in the paper.
Research literature synthesis demonstrates 70-75% automation potential.
Quantitative estimate offered by the authors (70-75%) as part of function-by-function analysis; no described empirical evaluation or sample supporting the figure.
Knowledge transmission (teaching/lecturing) shows 75-80% AI substitutability.
Authors' quantitative estimate presented in the analysis (75-80%); the paper does not detail empirical methods or validation samples for this percentage.
Administrative tasks face 75-80% disruption risk from AI.
Paper provides a quantitative estimate (75-80%) as part of its functional disruption assessment; no empirical methodology, dataset, or sample size is described to support the numeric range.
The remaining difference (roughly 70%) is not explained by the factors observed in the data, indicating additional influences not captured in the survey.
Residual (unexplained) component from decomposition analyses on ESJS data.
Heterogeneous trust levels across firms and schools may produce uneven productivity gains and widen performance gaps.
Logical implication and policy discussion in the paper; the cross-sectional study documents relationships between trust and outcomes but does not provide aggregate diffusion or cross-firm longitudinal evidence to confirm unequal sectoral diffusion.
Overreliance on unvetted AI can propagate biases; economic gains from AI therefore require governance, auditing, and accountability mechanisms.
Framed as a risk and policy recommendation in the discussion; not an empirical finding from the cross-sectional survey reported in the summary.
Shrinking acquisition workforce capacity functions as a critical scarce input in defense AI economics; reduced human capital lowers the Department's ability to extract value from AI investments and to internalize externalities, decreasing effective returns to AI procurement.
Institutional trend evidence of workforce reductions combined with economic analysis treating institutional capacity as an input factor. No empirical quantification of returns or elasticity provided—this is analytical inference.
Ambiguous standards increase uncertainty for contracting officers, raising the risk that they will either over-rely on vendor claims or inconsistently enforce requirements, both of which harm procurement integrity.
Policy-text analysis identifying vague criteria combined with qualitative analysis of procurement decision workflows; argument based on measurement and enforcement friction literature. No empirical study of contracting officer behavior provided.
Lower governance barriers and ambiguous procurement criteria (e.g., undefined 'model objectivity') can skew market competition toward suppliers that prioritize rapid iteration and opaque practices over rigorous assurance, harming traceability and quality.
Market-effects reasoning grounded in policy changes (document analysis) and qualitative institutional analysis of measurement/enforcement frictions. No market-share or supplier-behavior data provided.
Mandating permissive contract terms and enabling waivers reduces private incentives for contractors to invest in safety and compliance, creating classical moral-hazard problems in defense AI procurement.
Economic reasoning and principal–agent analysis applied to the documented contractual changes (primary-source policy text). No empirical measurement of contractor investment behavior provided; claim is theoretical/inferential.
A mismatch between expanded waiver authority (Barrier Removal Board) and declining acquisition oversight capacity creates procurement-integrity and systemic risks: faster acquisition concurrent with weakened institutional checks increases likelihood of improper procurement decisions and unchecked deployment of unsafe or unvetted AI models.
Synthesis of primary-source policy analysis, institutional staffing trend evidence, and qualitative risk/scenario assessment using principal–agent and moral-hazard frameworks. This is a conceptual risk projection rather than an empirically derived probability estimate.
Emerging agentic/AGI capabilities introduce new failure modes and governance challenges that standard ML oversight may not cover.
Emerging literature, theoretical analyses, and expert opinion summarized in the synthesis; authors note limited empirical long-term data and characterize this as an emergent risk.
If many firms adopt AI generation without matching verification, aggregate fragility in software-dependent infrastructure could rise, increasing downtime costs and systemic economic risk.
Macro-level risk projection and system fragility argument in the paper; no macroeconomic modeling or empirical scenario analysis provided.
This reversal of the burden of proof creates moral-hazard-like behavior: incentives for speed reduce verification effort.
Theoretical argument built on the micro-coercion mechanism and economic reasoning; no empirical validation provided.