Evidence (11633 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 736 | 1615 |
| Governance & Regulation | 664 | 329 | 160 | 99 | 1273 |
| Organizational Efficiency | 624 | 143 | 105 | 70 | 949 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 348 | 109 | 48 | 322 | 836 |
| Output Quality | 391 | 120 | 44 | 40 | 595 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 275 | 143 | 62 | 34 | 521 |
| AI Safety & Ethics | 183 | 241 | 59 | 30 | 517 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 105 | 40 | 6 | 187 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 78 | 8 | 1 | 151 |
| Regulatory Compliance | 69 | 64 | 14 | 3 | 150 |
| Training Effectiveness | 81 | 15 | 13 | 18 | 129 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
Societal acceptance of AI-generated audiovisual media is uncertain and could range from widespread uptake to broad rejection.
Discussion drawing on mixed empirical studies and scenario construction in the review; the paper notes contradictory findings in existing studies but does not provide primary survey data or sample sizes.
If cognitive interlocks are widely adopted, many negative externalities can be internalized and AI-driven productivity gains can be realized more sustainably; absent such controls, equilibrium may drift toward higher error rates and systemic incidents.
Long-run equilibrium argument based on theoretical reasoning and conditional claims; no longitudinal or cross-firm empirical evidence presented.
If AI raises the quality and pace of research, social returns to public research funding could increase, but distributional concerns and negative externalities must be managed to realize aggregate welfare gains.
Welfare implication discussed in the paper. Framed as conditional and theoretical; not empirically quantified in the abstract.
Policy interventions (data governance, transparency, reproducibility standards, ethical guidelines) will shape adoption and externalities (misinformation, misuse, reproducibility crises).
Policy recommendation/implication stated in the paper. This is a normative and predictive claim grounded in governance literature; the abstract does not present empirical evaluation of specific policies.
Labor demand effects are ambiguous: junior/entry-level demand may be reduced for some tasks while demand for verification and higher-skill roles may rise.
Economic reasoning, early observational signals, and theoretical task-reallocation frameworks; empirical longitudinal evidence is limited or absent.
The effectiveness of generative AI depends critically on human-AI workflows: prompt design, iterative refinement, and human vetting materially affect outcomes.
Qualitative analyses of interaction patterns and experiments manipulating prompting/iteration showing variation in outcomes; many studies report improved outputs after iterative prompting and human-in-the-loop refinement.
CRAEA-style systems could increase household productivity and substitute for some routine in-home human labor, altering demand for certain service roles and increasing demand for higher-skill roles (e.g., maintenance, AI oversight).
Paper's implications/economic analysis and qualitative extrapolation based on observed performance improvements in simulation; no empirical labor-market or deployment data provided to substantiate real-world labor substitution claims.
Integrated ERP vendors embedding AI could strengthen vendor lock-in, while interoperable AI layers may foster ecosystems and specialized entrants; empirical work is needed to determine market outcomes.
Conceptual discussion and observed vendor behavior in practitioner literature; explicit statement in the paper that empirical analysis is required.
Market demand is likely to bifurcate: high-value clinical markets will require rigorous explainability and neuroscientific grounding (higher willingness-to-pay), while research and consumer segments may tolerate black-box models (lower margins).
Market segmentation argument built from differing end-user requirements and tolerance for opaque models; presented as a projected implication rather than an empirically tested market study.
Persistent declines in self-efficacy after passive AI exposure suggest potential for skill atrophy and slower reversion when tasks must be performed without AI.
Inference from observed persistent reductions in self-efficacy post-return in the experiment; skill atrophy and reversion costs not directly measured—this is an implied consequence.
Firms that adopt passive, copy-based AI workflows risk psychological costs that could offset short-run productivity gains from AI.
Inference drawn from experimental findings of reduced efficacy/ownership/meaningfulness under passive use and short-term enjoyment gains; not directly tested for firm-level productivity or turnover—extrapolation from individual-level psychological measures.
Teams often produce evaluation outputs (tests, metrics, user feedback) but lack mechanisms, processes, or technical levers to convert those outputs into actionable engineering or product changes—a novel “results-actionability gap.”
Recurring theme from the 19 practitioner interviews and coding; authors explicitly articulate and label this gap based on participants' reports.
The study confirms several previously documented evaluation challenges with LLMs: model unpredictability, metric mismatch, high human-evaluation costs, and difficulty reproducing failures.
Interview data from 19 practitioners; thematic analysis flagged these recurring problems as reported by participants and aligned with prior literature.
Emergent quality hierarchies among agents imply winner-take-most dynamics in informational value and potential market concentration in agent quality.
Observed formation of quality hierarchies in agent interactions and documented economic interpretation; this is a hypothesis/implication drawn from qualitative patterns rather than measured market outcomes.
Security of LLM-based MASs functions as an economic externality: failures can impose social costs (misinformation, poor collective decisions), and absent liability or market incentives providers may underinvest in robustness.
Economic reasoning and implication section in the paper—conceptual argument linking the technical vulnerability to economic externality and incentive misalignment. No empirical economic data provided in the summary.
Analytical conditions on stubbornness and influence weights identify when a single adversary can dominate network dynamics (i.e., influence propagation criteria derived from FJ fixed-point analysis).
Mathematical/theoretical analysis of FJ model fixed points and influence propagation in the paper; derivation of conditions relating agent stubbornness and interpersonal trust weights to steady-state influence.
Large-scale battlegrounds and competitions increase compute demand and associated costs, with implications for budgets and environmental externalities.
Paper notes that the Battling Track dataset (20M+ trajectories), model training for baselines/competitions, and running a living benchmark imply substantial compute; this is an argued implication rather than measured environmental impact.
Rapid deployment of autonomous learners could accelerate displacement in affected sectors and widen inequality if gains concentrate among capital owners or platform providers.
Socioeconomic risk assessment and projection; conceptual and not empirically quantified in the paper.
Faster, more generalist embodied AI could substitute for routine physical and social tasks, shifting human labor toward oversight, high-level planning, creativity, and flexible social cognition roles.
Labor-market impact hypothesis derived from automation literature; conceptual projection only.
Organizations without access to high-frequency operational data may face increased barriers to entry in latency-sensitive markets, concentrating rents with incumbents who can collect such data.
Paper presents this as an implication of the dataset/value results: proprietary high-frequency data can create competitive advantages. This is a policy/economic implication derived from model performance observations rather than a tested market analysis.
If models frequently leak or misuse preferences in third‑party contexts, users and organizations will discount the value of personalization or demand stronger controls, increasing costs for deploying memory features and reducing consumer surplus.
Economic reasoning and implication drawn from the observed misapplication behavior; no empirical user adoption or market data provided in the study to directly support this claim.
The failure mode (misapplication of preferences to third parties) creates negative externalities (privacy violations, normative harms, misinformation, contractual breaches) that markets and platforms may not internalize without regulation or design changes.
Economic interpretation and argumentation building on the empirical failure mode; these harms are hypothesized implications rather than measured outcomes in the paper.
Widespread adoption of predictive HR tools raises distributional and fairness concerns (algorithmic bias, disparate impacts) and privacy risks that may prompt regulatory responses affecting adoption costs and equilibrium outcomes.
Discussion/implications section raises these risks conceptually; the paper does not empirically measure downstream policy or distributional effects.
Unclear liability frameworks increase perceived and real costs and can slow adoption by hospitals and insurers.
Policy analyses and procurement narratives noting liability uncertainty cited as a barrier to procurement and deployment.
Up-front implementation costs commonly include procurement, integration with PACS/EMR, UI/UX development, regulatory compliance, and staff training; recurring costs include monitoring, data labeling, software updates, and cybersecurity.
Implementation reports, vendor and hospital accounts, and qualitative studies documenting cost categories (specific dollar amounts vary across settings and are rarely published in detail).
Uneven organizational supports can concentrate returns to AI in firms and workers that successfully actualize affordances, potentially widening wage and employment disparities; targeted policy and training investments can mitigate these effects.
Theoretical implication from the framework with policy recommendations; no empirical testing or sample reported in the paper.
These trends (job polarization and differential wage/mobility outcomes) may exacerbate economic disparities across regions.
Interpretation and projection based on the observed trends in the reviewed literature and reports; presented as a risk/implication rather than an empirically tested causal finding in the summary.
Without continuous support for upskilling/reskilling and inclusive policies, AI risks becoming a source of exclusion rather than an enabler of human advancement.
Normative conclusion derived from reviewed literature and thematic interpretation in the qualitative study (literature-based; evidence is secondary and not quantified).
At the national level, AI-related innovations are yet to be transformed into measurable economic gains.
Interpretation based on the observed negative association between AI patent counts and GDP growth from the panel regressions (OLS, FE, Difference and System GMM) and theoretical reasoning about adoption/diffusion lags and complementary requirements; empirical support derives from the same models (sample details not provided).
Research literature synthesis demonstrates 70-75% automation potential.
Quantitative estimate offered by the authors (70-75%) as part of function-by-function analysis; no described empirical evaluation or sample supporting the figure.
Knowledge transmission (teaching/lecturing) shows 75-80% AI substitutability.
Authors' quantitative estimate presented in the analysis (75-80%); the paper does not detail empirical methods or validation samples for this percentage.
Administrative tasks face 75-80% disruption risk from AI.
Paper provides a quantitative estimate (75-80%) as part of its functional disruption assessment; no empirical methodology, dataset, or sample size is described to support the numeric range.
Over 400,000 [individuals] are projected to die before obtaining permanent residency.
Mortality projection applied to the estimated backlog and projected wait times (authors' projection); exact demographic assumptions (age distribution, mortality rates) and method are not provided in the excerpt.
The remaining difference (roughly 70%) is not explained by the factors observed in the data, indicating additional influences not captured in the survey.
Residual (unexplained) component from decomposition analyses on ESJS data.
Demand-dependent pricing in the modeled energy load management setting creates a social dilemma: everyone would benefit from coordination, but in equilibrium agents often choose to incur congestion costs that cooperative turn-taking would avoid.
Theoretical/modeling analysis of consumer agents scheduling appliance use under demand-dependent pricing as described in the paper (analytical argument and/or model simulations). Specific sample sizes or simulation parameters are not given in the abstract.
Aggregation and linkage across data sources can reveal intimate, predictive traits that were not foreseeable to the data subject at the time of sale.
Conceptual argument with references to documented cases and literature on data linkage and inference; relies on illustrative examples rather than original empirical experiments.
Policy-relevant implication (extrapolated): identity heterogeneity implies family- and purpose-driven entrepreneurs may be less likely to pursue AI-enabled innovation after income shocks, suggesting targeted outreach and low-risk entry paths to avoid widening digital divides.
Extrapolation from documented identity-heterogeneous declines in innovation after income shocks (empirical result) to probable patterns in AI adoption; AI adoption is not directly measured in the paper's dataset.
The United States shows a more market-driven (firm-dominated) patenting profile and comparatively weaker integration between AI and robotics patent trajectories.
Country-level and actor-type decomposition for U.S. patent filings (1980–2019), showing higher firm share of patents and weaker long-run association/cointegration between core AI and AI-enhanced robotics series compared with China (as reported in the paper).
There is a risk of a two‑tier market where high‑quality temporal‑preserving enhancements are costly, increasing inequality in experiential welfare and cognitive capital.
Speculative socioeconomic implication based on cost/access arguments and distributional concerns; no inequality modeling or empirical pricing data provided.
Technical expansion without an accompanying theory of lived temporality risks increasing capabilities while degrading the qualitative depth of human experience (presence, attentional flow, felt meaning).
Argumentative claim supported by philosophical analysis and literature synthesis (neurophenomenology, attention economics); no empirical test reported (N/A).
Differential access to higher-quality (paid) versus free GenAI tools and differing ability to engage with the tool could widen inequality among students and institutions.
Authors' implication based on student-reported concerns about limitations of free ChatGPT versions and on heterogeneous gains across disciplines; this is a policy/implication claim not directly measured in the experiment.
High-quality, equitable climate information displays public-good characteristics (nonrival, nonexcludable at scale), so private incentives alone will underprovide geographically representative data and shared infrastructure.
Economic reasoning supported by observed concentration of compute and model development (mapping) and standard public-goods theory; no formal empirical market model estimated in the paper.
Improving photorealism with objective color-fidelity metrics and refinement reduces the need for manual color correction and retouching in downstream workflows.
Paper and summary argue this as an implication: higher-fidelity outputs from CFR/CFM reduce manual editing demand. This is an economic/market implication rather than a directly evidenced experimental result in the paper (no labor-market causal study reported).
The paradigm implies potential market risks including vendor lock-in and concentration if only a few providers control scalable linear-optical samplers.
Conceptual risk analysis in the paper's discussion of economic implications; this is a qualitative argument built on the technical premise that trained models require access to specialized quantum sampling hardware for deployment.
Heterogeneous trust levels across firms and schools may produce uneven productivity gains and widen performance gaps.
Logical implication and policy discussion in the paper; the cross-sectional study documents relationships between trust and outcomes but does not provide aggregate diffusion or cross-firm longitudinal evidence to confirm unequal sectoral diffusion.
Overreliance on unvetted AI can propagate biases; economic gains from AI therefore require governance, auditing, and accountability mechanisms.
Framed as a risk and policy recommendation in the discussion; not an empirical finding from the cross-sectional survey reported in the summary.
If FDI brings capital‑intensive, AI‑enabled production without complementary upskilling, it may exacerbate wage inequality and deepen labor market dualism in SSA.
Theoretical inference and analogy from documented patterns of skill‑biased technological change and FDI-driven inequality in the reviewed literature; empirical evidence specific to AI in SSA is lacking in the review.
Full replacement of physicians would require breakthroughs in robust generalization, embodied capabilities, and legal/regulatory change—currently lacking.
Conceptual inference based on documented limitations (OOD generalization, lack of embodied/sensorimotor capability, unsettled legal/regulatory environment) summarized in the review.
Shrinking acquisition workforce capacity functions as a critical scarce input in defense AI economics; reduced human capital lowers the Department's ability to extract value from AI investments and to internalize externalities, decreasing effective returns to AI procurement.
Institutional trend evidence of workforce reductions combined with economic analysis treating institutional capacity as an input factor. No empirical quantification of returns or elasticity provided—this is analytical inference.
Ambiguous standards increase uncertainty for contracting officers, raising the risk that they will either over-rely on vendor claims or inconsistently enforce requirements, both of which harm procurement integrity.
Policy-text analysis identifying vague criteria combined with qualitative analysis of procurement decision workflows; argument based on measurement and enforcement friction literature. No empirical study of contracting officer behavior provided.