Evidence (5126 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Adoption
Remove filter
Generative AI serves as an effective 'wingman' for employment lawyers, capable of replacing substantial junior associate work while requiring continued human expertise for client counseling, supervision, and final legal advice preparation.
Authors' synthesis of experimental results showing AI-produced substantive analysis plus discussion about remaining limitations (e.g., citation errors) and required human oversight; qualitative assertion about substitutability for junior associate tasks.
We evaluate 14 LLMs under zero-shot prompting and retrieval-augmented settings and witness a clear performance gap.
Experimental evaluation reported in the paper: authors state they ran experiments on 14 different large language models, under zero-shot and retrieval-augmented configurations, and observed differing performance across models.
Policy implication: smarter, better-coordinated green governance is needed to address the negative local impacts and the crowding-out interaction between AI and environmental regulation.
Policy recommendation drawn in the abstract based on the empirical spatial findings (negative local effects and negative interaction).
Substantial regional gaps persist: leading eastern provinces approach a UCEE value of 1.0 while some northeastern provinces remain below 0.1.
Regional UCEE index estimates from the Super-SBM model across the 30 provinces reported in the abstract.
The systemic implications of AI in finance depend less on model intelligence alone than on how agent architectures are distributed, coupled, and governed across institutions.
Central argumentative claim supported by the AFMM conceptual model and an illustrative empirical application described in the paper (modeling + event-study approach); no full-sample details provided in the excerpt.
The Agentic Financial Market Model (AFMM), a stylised agent-based representation, links agent design parameters (autonomy depth, heterogeneity, execution coupling, infrastructure concentration, supervisory observability) to market-level outcomes including efficiency, liquidity resilience, volatility, and systemic risk.
Presentation of a stylised agent-based model (AFMM) in the paper; conceptual modelling linking specified agent parameters to macro/market outcomes. No empirical sample size reported in the excerpt.
Financial AI agents can be described by a four-layer architecture covering data perception, reasoning engines, strategy generation, and execution with control.
Conceptual framework proposed by the authors (theoretical/architectural proposal); no empirical testing or sample size provided.
These productivity gains are most pronounced for lower-skilled workers, producing a pattern the authors call “skill compression.”
Cross-study pattern reported in the literature review: comparative evidence across worker-skill strata in multiple empirical papers showing larger relative gains for lower-skilled/junior workers; specific underlying studies and sample sizes are not enumerated in the brief.
Financial well-being is not an automatic byproduct of automated credit efficiency but an emergent outcome of architectural alignment among technology, borrower capability, and governance structures.
Theoretical conclusion drawn from empirical results showing mixed effects (positive on repayment and resilience, negative on stress) and significant moderation by human capability and institutional design.
The authors identify ten evaluation practices that teams use, ranging from lightweight interpretive checks to formal organizational processes (examples: qualitative user reviews, red-team testing, A/B experiments, telemetry/log analysis, structured annotation, governance/meta-evaluation).
Thematic coding of 19 interview transcripts produced a taxonomy enumerating ten practices (paper reports the taxonomy as an outcome).
Quantum-driven growth depends critically on adoption rates, infrastructure readiness, complementary investments (digital infrastructure, human capital), and enabling policy/regulatory environments.
Scenario framework that varies (a) technical timelines, (b) sectoral adoption rates (diffusion models), (c) infrastructure readiness, and (d) policy environments; policy counterfactual modeling shows sensitivity of adoption and macro outcomes to these parameters.
The magnitude and timing of macroeconomic impact from quantum computing are highly uncertain.
Monte Carlo / scenario ensemble results showing wide (fat-tailed) outcome distributions driven by uncertainty in technical milestones, adoption rates, and complementarity strengths; use of expert elicitation to parameterize tail risks.
Safeguards such as audit trails, explainability, and human oversight impose additional implementation costs that must be weighed against efficiency benefits.
Normative and economic reasoning based on requirements for compliance and system design; no empirical cost estimates provided.
There is a fundamental tension between AI-driven efficiency and core administrative-law principles—discretion, due process, and accountability.
Doctrinal legal analysis of administrative-law principles in Vietnam and comparative institutional analysis of AI adoption in other systems.
The net educational value of AI-generated feedback depends on alignment with pedagogical goals, quality evaluation, integration with human teaching, and governance to manage equity, privacy, and incentives.
Synthesis statement from the meeting report produced by 50 interdisciplinary scholars; conceptual judgment rather than empirical proof.
LLMs excel at extracting and generating arguments from unstructured text but are opaque and hard to evaluate or trust.
Synthesis of recent LLM literature and observed properties (generation capability vs. opacity); no empirical evaluation within this paper.
The paper is primarily theoretical and historical; empirical validation is needed to quantify the irreducible component of LLM value, and practical degrees of rule‑extractability may exist even if some capabilities remain tacit.
Stated limitations section acknowledging the theoretical nature of the work and the need for empirical follow‑up.
If an LLM's full capability were reducible to an explicit rule set, that rule set would be an expert system; because expert systems are empirically and historically weaker than LLMs, this leads to a contradiction (supporting non‑rule‑encodability).
Logical proof‑by‑contradiction presented in the paper, supported by conceptual mapping between rule sets and expert systems and qualitative historical comparisons.
HindSight has limitations: it depends on citation and venue proxies for impact, uses a finite forward window (30 months), and may undercount delayed-impact research and be domain-specific to AI/ML.
Authors' stated limitations in the paper noting reliance on observable downstream signals (citations/venues), the finite forward window, field heterogeneity, and measurement noise.
Practical caveats: benefits depend on accelerators supporting MXFP formats; despite up to 96% recovery, residual quality gaps may remain for some task-specific or safety-critical cases; integration and tuning cost is required to apply BATQuant.
Discussion/limitation section in the paper outlining hardware dependency, remaining quality gaps despite high recovery percentages, and engineering effort for integration and tuning; these are argumentative caveats rather than results of controlled experiments.
The sign of the Largest Lyapunov Exponent (LLE) gives a precise criterion: negative LLE (contracting dynamics) permits fast convergence and real speedups for parallel Newton methods, whereas positive LLE (expanding/chaotic dynamics) prevents generally achieving fast convergence.
Theoretical derivation relating Lyapunov exponents to the stability of parallel-in-time linearizations and convergence of the parallel Newton iterations; supported by empirical observations reported on representative tasks.
Many fixed-point and iterative schemes (e.g., Picard, Jacobi) are unified as special cases within the parallel Newton framework.
Theoretical analysis and derivations in the thesis that show these classical iterative methods arise from particular choices/approximations in the parallel Newton formulation.
The core problem is a trade-off between computational latency/resource cost and decision correctness: invoking more LLM reasoning improves correctness but increases latency; invoking less reduces latency but can increase failures.
Paper frames the research problem explicitly as this trade-off in the Introduction/Problem framing sections and motivates the need for adaptive orchestration.
Demand for labor will shift toward data scientists, ML engineers, and interdisciplinary scientists, while wet-lab expertise and translational teams remain crucial.
Workforce trend analysis and employer hiring patterns summarized in the paper; interviews/case studies indicating changes in team composition.
AI excels at hypothesis generation but cannot replace scientific reasoning and experimental validation; human expertise remains essential.
Argument and case examples in the paper showing AI-generated hypotheses requiring human-led experimental design, interpretation, and validation.
Net gains from AI are not automatic nor evenly distributed; benefits depend on translation rates to clinical success and on addressing non-technical enablers.
Synthesis and conditional argument informed by sector observations; not backed by empirical distributional analysis in the paper.
Alignment with evolving regulatory expectations (evidence standards, auditing, liability) is necessary to translate AI capabilities into products and reduce adoption risk.
Policy-focused argument referencing regulatory uncertainty; no empirical measures of regulatory impact included.
Realized, sustained impact ('democratized discovery') from AI depends on non-technological enablers: high-quality interoperable data, rigorous validation, transparency/auditability, workforce upskilling, ethical oversight, and regulatory alignment.
Synthesis and prescriptive argument in editorial grounded in observed constraints; no empirical testing of causal dependence provided.
The review synthesizes cross-domain evidence on the use of AI across the continuum from target identification to regulatory integration and critically evaluates existing limitations including data bias, interpretability discrepancy, and regulatory ambiguity.
Statement about the scope and content of the review (literature synthesis and critical evaluation). This is a description of the paper's methods/content rather than an empirical finding; the excerpt indicates these topics are discussed.
The study investigates the benefits and drawbacks associated with the incorporation of innovative artificial intelligence technologies into industrial policies.
Author-stated research objective reported in the text; evidence claimed to come from literature review (novel studies and existing literature), but no specific studies, sample sizes, or empirical measures are provided in the excerpt.
Model output can be treated as evidence for studying human behavior, but there are important epistemic limits to interpreting model-generated text as direct evidence of human beliefs or social facts.
Epistemic analysis and methodological critique in the paper (discussion of limits of treating model outputs as evidence); no single empirical test cited in the provided text.
The paper constructs three policy-contingent labor market scenarios for 2025–2035: (1) an Augmented Services Economy with inclusive productivity gains, (2) a Dual-Speed Labor Market characterized by polarization and uneven adjustment, and (3) a Disruptive Automation Shock involving significant displacement and social strain.
Prognostic, scenario-based approach integrating the three evidence bases (task-level capability mapping, occupational exposure/complementarity analysis, and firm- and worker-level adoption evidence). The scenarios are developed and described in the paper for the 2025–2035 horizon.
The validity of human–AI decision-making studies hinges on participants' behaviours; effective incentives can potentially affect these behaviours.
Conclusion from the authors' thematic review and theoretical rationale linking incentive design to participant behaviour and study validity (no quantitative effect sizes provided in excerpt).
The study's counterfactual analytical model links HR indicators (training intensity, absenteeism, labor productivity, turnover rates, workforce allocation) to organizational performance outcomes using regression-based simulations and predictive estimation.
Methodological claim explicitly stated: model construction from an industrial firm dataset using regression-based simulations and predictive techniques. (Specific sample size, variable operationalizations, and time frame not reported in the description.)
Only one study reported a modest improvement in predicting endoscopic intervention needs (AUC: 0.68).
Single-study result cited in the review reporting AUC = 0.68 for prediction of need for endoscopic intervention.
The review synthesizes findings across five thematic areas: AI‑driven task automation and decision support; digital literacy and capacity building; gender‑sensitive employment patterns; infrastructural and policy challenges; and sustainable development outcomes.
Thematic synthesis of the 55 included articles as described in the paper; themes explicitly listed by the authors.
Prevalence and risk factors for poverty differ by gender, as does the nature of vulnerability.
Stated as a general empirical claim in the introduction, drawing on broader literature (no specific study, method, or sample size provided in the excerpt).
Major actors such as the United States, China, and the European Union pursue distinct models of AI development and regulation.
Comparative policy analysis and qualitative document review of national/regional AI strategies and regulatory proposals for the United States, China, and the EU (specific documents and sample size not specified).
The study identifies the emergence of three competing governance paradigms: the innovation-driven liberal model, the ethics-oriented regulatory model, and the state-controlled authoritarian model.
Finding from the paper's comparative policy analysis and qualitative review of policy documents across major actors (United States, European Union, China); underlying document sources referenced qualitatively but not enumerated as a quantitative sample.
Distinct AI features (recommendation engines, chatbots, and comparison tools) influence consumer outcomes when modeled as latent constructs.
Methodological claim: the study modeled three AI features as latent constructs and analyzed their relationships with dependent variables using SEM (quantitative questionnaire data).
We develop a theoretical framework - the productivity funnel - that traces how technological potential narrows through successive stages, from access and digital infrastructure, through organizational absorption and human capital adaptation, to ultimate value capture.
Conceptual/theoretical development presented in the paper; no empirical sample needed (framework-building).
Effects of curated Skills are highly heterogeneous across domains (e.g., +4.5 pp in Software Engineering vs. +51.9 pp in Healthcare).
Per-domain pass-rate deltas reported in the paper (SkillsBench per-domain analysis). The example domain deltas (+4.5 pp and +51.9 pp) are taken from the reported per-domain results.
Institutional factors (education systems, active labor market policies, mobility, industrial policy, social protection) shape net employment outcomes from AI.
Theoretical and policy-focused synthesis; cross-country comparisons in literature highlight institutional mediation though no single new cross-country empirical estimate is provided.
Net employment effects depend on the balance of substitution and complementarity, sectoral exposure, and institutional responses.
Conceptual labor-economics framework (task-based, skill-biased change) and comparative review of cross-country/sectoral evidence emphasizing institutional mediation.
AI will substantially restructure labor markets.
Task-based theoretical approach and cross-sectoral synthesis of empirical studies showing task substitution and complementarity effects across occupations and sectors.
The pandemic produced a 1.5% increase in people identifying as potential entrepreneurs but a 2.3% contraction in emerging entrepreneurs, indicating a breakdown in converting aspiration into formal entrepreneurial activity (pipeline disruption).
Reported percentage changes in pipeline stages (potential entrepreneurs and emerging entrepreneurs) measured in the survey before/after (or during) the pandemic within the >27,000 respondent sample; comparison of identification and transition rates along the entrepreneurial pipeline.
Long-run integration (degree of long-run association) between core AI and AI-enhanced robotics differs systematically across national innovation systems.
Country-level decomposition of patent filing series and time-series econometric tests for long-run relationships / cointegration between core AI and AI-enhanced robotics patent series for each country/region (China, U.S., Europe, Japan, South Korea).
Core AI, traditional robotics, and AI-enhanced robotics follow distinct historical trajectories over 1980–2019 and do not move together uniformly.
Time-series analysis using annual patent filing counts (1980–2019) for each domain; tests for common long-run relationships / co-movement across the three patent series (as reported in the paper). Country-aggregated and domain-specific patent time series were analyzed; exact sample size (total patents) not specified in the summary.
Kondratieff, Schumpeter, and Mandel each highlight different drivers of capitalist long waves: Kondratieff emphasizes regular technological-driven renewal, Schumpeter emphasizes entrepreneurship and innovation-led creative destruction, and Mandel emphasizes class relations and production structures.
Comparative theoretical analysis and literature synthesis across the three schools; conceptual summary of canonical positions (no original dataset; qualitative interpretation).
XChronos reframes transhumanist technology evaluation in experiential terms, creating both market opportunities and measurement/regulatory challenges for AI economics.
Synthesis and concluding argument in the paper summarizing proposed implications; conceptual reasoning without empirical tests.