Evidence (4333 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Governance
Remove filter
Dispersed work alters identity construction, belonging, and social cohesion; digital interactions reshape workplace rituals and norms.
Sociological literature synthesis and qualitative case illustrations emphasizing identity and ritual processes; no longitudinal or quantitative measures provided in the paper.
Demand for defensive AI engineers and incident responders will rise, while demand for traditional offensive hacking labor may decline as automation substitutes some roles.
Labor-market reasoning based on substitution/complementarity between automation and human tasks (qualitative; no labor-market data).
The paper proposes an 'algorithmic workplace' framework emphasising hybrid agency (agents composed of humans plus GenAI), decentralised decision processes, and erosion of rigid managerial boundaries.
Conceptual synthesis derived from thematic mapping, co‑word analysis and interpretive discussion of the mapped literature; framework presented as the article's conceptual contribution.
AI diffusion and China’s delayed retirement policy jointly shape pre-retirement workers’ willingness to stay employed.
Cross-sectional survey (n=889) of pre-retirement respondents in Beijing, Guangzhou, and Lanzhou; multivariate regression analysis examining associations between employment willingness and regional AI exposure plus policy context (delayed retirement).
Proprietary versus open DPP data regimes will shape competition: closed data can lead to vendor lock-in and market power, while open standards can spur broader innovation but may reduce short-term rent extraction.
Conceptual policy/economics argument informed by observed stakeholder perspectives and literature; not empirically tested in this study.
DPP ecosystems resemble multi‑sided platforms (producers, recyclers, consumers, certifiers) with network effects such that more participants increase DPP data value, potentially creating winner-take-most dynamics unless standards and interoperability are enforced.
Theoretical/platform-economics reasoning grounded in empirical description of stakeholders and DPP roles from the study; not directly tested with market-level data in the paper.
Design choices around openness must balance privacy, proprietary information, and commercial sensitivities with public-good benefits; these choices will shape incentives and model validity.
Conceptual policy analysis highlighting trade-offs; no empirical study of design outcomes provided.
Vulnerability is path-dependent and contingent on states’ adaptive capacity—governance quality, industrial policy, and bargaining leverage determine whether a country captures upgrading opportunities or becomes a strategic casualty.
Comparative case analysis using indicators of governance, industrial policy presence, and bargaining outcomes; process tracing of critical junctures showing divergent trajectories. (Data sources: governance indicators, case comparisons; sample sizes not specified.)
Trade diversion caused by tariff escalation and restrictions re-routes production and trade flows, but benefits are asymmetric: countries with stronger institutions, infrastructure, and policy capacity capture more investment and value-added.
Analysis of bilateral trade and FDI flow changes after tariffs; supply-chain mapping of relocation events; firm announcements of relocation; comparative cases emphasizing institutional/infrastructure differences. (Data sources: trade and investment flow data, supply-chain maps, firm-level announcements; sample sizes not specified.)
The benefits of AI come with governance, ethical, and sustainability challenges (standards, control, accountability) that require balancing against innovation incentives.
Synthesis of policy, ethics, and governance literature documenting concerns about standards, accountability, and incentive trade-offs; argument is qualitative and prescriptive rather than empirically tested within this paper.
AI has enhanced delivery in education, health, transportation, and government, improving some service outcomes while persistent issues like bias, privacy, transparency, and accountability remain.
Synthesis of applied-AI case studies and sectoral evaluations drawn from interdisciplinary literature; evidence described qualitatively without new empirical aggregation or meta-analysis in this paper.
AI reshapes demand for skills, redefines occupations, and accelerates the need for reskilling, with distributional effects that can increase inequality.
Narrative review of labor-economics and workforce studies documenting task reallocation and shifting skill requirements; based on observational studies and sectoral analyses summarized in the review (no unified sample size or new empirical test in this paper).
A multi-hazard, multi-risk approach increases societal resilience but is complex and cross-disciplinary.
Project-wide synthesis, in-depth place-based case studies, and stakeholder engagement reported in MYRIAD-EU activities indicating benefits to resilience alongside noted disciplinary and practical complexity.
Shifting disaster risk management toward a genuinely multi-hazard, multi-risk paradigm is feasible and valuable but requires coordinated advances across conceptual mainstreaming, evidence on spatio-temporal hazard–exposure–vulnerability dynamics, scenario methods, usable decision-support tools, explicit equity integration, deep case-study coproduction, support for MHEWS, and strengthened ECR leadership.
Synthesis and reflection across MYRIAD-EU (2021–2025) project outputs, comparative synthesis of activities, lessons learned, and stakeholder feedback reported by the project.
Technical milestones (scalable, error-corrected qubits; hybrid algorithms) create fat-tailed outcome distributions where a small probability of breakthrough could yield outsized long-run effects.
Monte Carlo experiments and scenario ensembles that include low-probability, high-impact technical breakthrough parameters; expert elicitation of milestone probabilities.
R&D funding, standards, regulatory clarity, export controls, and public–private partnerships shape quantum adoption trajectories; policy missteps can slow adoption and concentrate benefits.
Policy counterfactual scenarios and qualitative analysis of ecosystem roles; calibration informed by historical effects of policy on diffusion of strategic technologies.
Aggregate gains hinge on how quickly and broadly quantum technologies diffuse; early gains concentrated in frontier firms/sectors can take decades to propagate economy-wide.
Diffusion modeling using logistic/S-curve and Bass models calibrated to historical analog technologies; scenarios show long lag between frontier adoption and economy-wide diffusion.
As successive pilot batches of urban green data center policies are rolled out, the aggregate policy impact follows a nonlinear rise-then-fall (increase followed by decline) diffusion trajectory.
Analysis across pilot-batch rollout timing showing a nonlinear (rise-then-fall) pattern in aggregate estimated effects as the number of pilot batches expands; modeled/visualized within the staggered-adoption DID framework.
Realizing NLP value in banks requires organizational investments (data pipelines, model deployment, CRM integration) and complementarity between AI tools and managerial/IT capabilities; returns will depend on these complementarities.
Conceptual implication derived from review of applied/engineering papers and literature on technology complementarities; not directly estimated empirically in the review.
Automated tax-preparation and filing could increase compliance rates but also make tax bases more sensitive to automated tax-optimization strategies, requiring updated regulatory oversight and audit tools.
Paper's policy and economic implications section combining case-based observations and literature; presented as plausible outcomes rather than measured effects.
Ethics is distinct from and prior to law: legal codification cannot fully capture the primordial ethical demand.
Philosophical engagement with Derrida and Levinas; normative argumentation and conceptual examples. No empirical validation of precedence.
Legal norms and technical reforms are necessary but incomplete: they must remain responsive to a primordial, non-codifiable ethical obligation that structures how responsibility is perceived and allocated in practice.
Conceptual analysis drawing on Derrida and Levinas; argument supported by illustrative cases across three domains (care robotics, AVs, algorithmic governance). No empirical measurement of legal efficacy.
AI-driven productivity and data externalities can reconfigure which countries/regions specialize in which activities, with implications for labor demand, offshoring, and services trade patterns.
Mechanism and theory-based analysis drawing on literature about comparative advantage, automation, and data externalities; empirical testing recommended but not performed in the paper.
Standard international trade models should be updated to incorporate data as an input, platform-mediated matching, algorithmic complementarities, and costs of regulatory fragmentation.
Theoretical critique and modeling recommendations based on mechanism analysis; no new formal model calibration or empirical testing presented in the paper.
AI-enabled markets tend toward winner-take-most platforms amplified by network effects.
Theoretical reasoning supported by platform literature and case illustrations of platform concentration dynamics; empirical magnitudes not estimated in the paper.
Competitive advantage is shifting away from asset- and labor-intensive models toward data-, model-, and platform-driven advantages, altering comparative advantage and market structure.
Mechanism/theoretical analysis drawing on platform and AI economics literature and qualitative examples; no empirical estimation provided in the paper.
Regulatory design acts as an economic instrument that can balance social value from AI with protection of rights, affecting social welfare, public trust, and long-term adoption rates.
Normative synthesis combining legal and economic reasoning; suggested as a theoretical mechanism rather than empirically validated within the paper.
Automation of routine administrative tasks may reduce demand for certain clerical roles while increasing demand for oversight, auditing, and legal-technical expertise, altering public-sector labor composition and retraining needs.
Qualitative labor-market reasoning based on task-based automation literature and the administrative context; no field labor-data or sample provided.
AI feedback may either augment teacher productivity (complementarity) or substitute for routine teacher feedback tasks (substitution), with unclear net labor impacts.
Workshop deliberations among 50 scholars highlighting competing theoretical scenarios; no causal labor-market evidence provided.
Easier conversational access to models can substitute for routine cognitive labor while complementing high-skill work; miscalibrated trust affects labor outcomes and supervision costs.
Labor and task-allocation implications argued conceptually; no labor-market empirical evidence or quantified substitution/complementarity rates presented.
Firms can compete on front-end design (transparency, trustworthiness) as a socially beneficial quality signal, but absent regulation competition may favor more persuasive (less honest) interfaces.
Economic argument about product differentiation and competitive incentives, drawn from market theory and literature; no empirical market study provided.
Misleading cues can create short-term surplus (user satisfaction) but long-term welfare losses if overtrust causes harms or misinformation.
Theoretical economic argument based on information asymmetry and externalities; no empirical quantification in the paper.
LLM-based chatbots’ conversational naturalness increases usability and adoption but also triggers misleading mental models (e.g., anthropomorphism, overtrust).
Paper-level main finding based on conceptual analysis and literature synthesis from HCI, ethics, and conversational analysis; no new large-scale empirical study or sample reported.
Human experts will likely shift roles from sole decision-makers to adjudicators, challengers, and validators of AI-generated arguments, changing required skills toward critical evaluation and dialectical oversight.
Conceptual labor-market projection; no empirical labor studies or surveys presented.
Productivity gains from partial automation may be offset by negative externalities (incorrect legal outcomes, appeals, reputational damage) that impose social and private costs not captured by narrow productivity measures.
Theoretical economic analysis and illustrative case vignettes describing error propagation; no empirical quantification of externalities.
Market demand will likely split between providers offering generative convenience with liability exposure and providers offering certified/verified, explainable tools at a premium, creating a two-tier market.
Market-structure analysis and illustrative projections; no empirical market data or sample size.
Reported monetary supervision cost was low (~$200) for this project, but the paper cautions that general equilibrium effects and scaling may change costs as demand for supervisors rises.
Paper provides reported supervision cost (≈$200) for the single project and includes a caveat about external validity and scaling; cost is self-reported and contextualized by authors.
Because these agents will be embedded in safety‑critical infrastructure, economic and technical outcomes will depend heavily on system architecture choices.
Systems‑engineering and policy reasoning drawing on analogies to Internet/IoT evolution and domain examples (disaster response, healthcare, industrial automation, mobility); conceptual argumentation rather than empirical measurement.
Policymakers must weigh productivity gains from higher autonomy against increased systemic risk and governance costs; optimal allocation will vary by sector (high-consequence systems justify stricter human oversight; lower-consequence tasks may tolerate more autonomy).
Normative policy analysis and cost–benefit reasoning; sector-differentiated triage framework proposed (no quantitative welfare or sectoral optimization performed).
Bounded-autonomy governance internalizes some externalities from automated interactions, reducing the probability of cascading failures and associated economic damages, but misaligned or heterogeneous governance across firms/sectors can still generate systemic vulnerabilities.
Theoretical argument combining externalities literature and governance design principles; illustrative scenarios and policy reasoning (no empirical validation).
Modern critical infrastructure increasingly uses embodied AI for monitoring, predictive maintenance, and decision support, but these systems are typically trained for statistically representable uncertainty rather than systemic, cascading crises.
Review and synthesis of policy texts, industry descriptions, and safety/AI standards cited in the paper (EU AI Act, ISO standards) and literature on embodied-AI applications; conceptual argument (no original empirical sample).
Increasing benign-agent count and agent stubbornness are practical levers for improving robustness, but both carry costs: added compute/operational cost for scaling agents, and degraded consensus/coordination when stubbornness is high.
Argumentation supported by simulation results showing improved robustness with more agents or higher stubbornness, combined with discussion of computational cost (scaling) and observed consensus degradation; computational cost is presented as conceptual/operational reasoning rather than quantified in the summary.
Naïvely lowering trust weights assigned to suspected adversaries can limit adversarial influence but may also hinder cooperation and reduce task performance.
Simulations manipulating fixed trust weights and observing tradeoffs between reduced adversarial sway and decreased cooperative task performance/convergence; conceptual analysis of the tradeoff is provided.
Raising agents' innate stubbornness (peer resistance) reduces susceptibility to adversarial manipulation but impairs the network's ability to reach consensus or coordinate effectively.
Combined theoretical reasoning from FJ model (stubbornness is weight on innate opinion) and simulation experiments varying stubbornness parameters; measured outcomes include adversarial influence and measures of convergence/coordination or task performance.
Investments in interpretability that aim to fully 'rule‑ify' LLM competence may have diminishing returns; economic value may be better captured by research into robust behavioral evaluation, stress testing, and hybrid human‑AI workflows, while partial interpretability remains valuable.
R&D allocation and interpretability economics argument built on the central thesis; suggestion rather than empirical finding.
The paper challenges a purely rule‑based view of scientific explanation: some explanatory power will remain in implicit model structure rather than explicit rules.
Philosophical/epistemological argument based on the main thesis about tacit competence; no empirical validation.
Liability regimes and penalties should account for limits of enforced compliance and false positives/negatives from probabilistic policy evaluations.
Normative/economic discussion in the paper highlighting probabilistic outputs of the Policy function and calibration challenges; no empirical validation.
Firms will trade off compliance strictness against service quality (task completion rates), creating an economic tradeoff that shapes market offerings (e.g., safer-but-slower vs. faster-but-riskier agents).
Economic reasoning and conceptual models in the paper; suggested objective balancing task completion and legal/reputational costs; no empirical market data.
Alignment and instruction tuning approaches intended to encourage up-to-date answers improve some behaviors but do not reliably solve time-sensitivity and cross-modal consistency issues.
Experiments applying alignment/instruction-tuning methods with measurement of correctness and consistency; reported partial or inconsistent improvements rather than full resolution.
Diagnostic analysis links outdated predictions to (i) the static, time-stamped nature of training/evaluation datasets and (ii) mechanistic limits in how multimodal representations encode and retrieve temporal facts.
Error attribution analyses connecting incorrect answers to training snapshot timestamps and dataset provenance; representation-level analyses and qualitative case studies demonstrating multimodal encoding/retrieval limits.