Evidence (4333 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Governance
Remove filter
The economic value of deploying DeePC-based controllers depends critically on representativeness of training data and the costs of online adaptation and safety verification.
Authors' deployment-risk analysis and discussion of trade-offs (qualitative), grounded in methodological requirements of DeePC (need for representative, persistently exciting data and safeguards).
System-level improvements from the controller do not imply uniform spatial/temporal benefits—distributional effects may favor certain routes or neighborhoods.
Authors' discussion and caution about distributional effects and equity; possibly supported by spatial analyses in simulation (qualitative discussion in paper).
Deploying conformal factuality systems increases development cost (collecting representative calibration data) and inference cost (verifier compute), though efficient verifiers mitigate inference cost.
Discussion and empirical cost measurements: need for representative calibration datasets to maintain guarantees; measured verifier FLOPs; qualitative economic analysis in the paper.
Conformal filtering improves formal reliability (statistical factuality guarantees) but does not, by itself, deliver robustness and task utility without careful system design.
Aggregate empirical results: improved factuality guarantees after calibration/filtering, but concurrent reductions in informativeness and sensitivity to distribution shift/distractors unless calibration/data-processing are adapted.
DeepSeek-R1 exhibits a distributed memorization signature: 76.6% partial reconstruction rate but 0% verbatim recall on the TS‑Guessing probe.
Model-specific results from Experiment 3 (TS‑Guessing) reporting per-model rates of partial reconstruction and verbatim recall across the 513 MMLU items for DeepSeek-R1.
Quantitative comparisons across tested models show systematic Misapplication Rate even in settings where Appropriate Application Rate is high.
Aggregated MR and AAR statistics reported for multiple frontier models across the benchmark showing co‑occurrence of high AAR and nontrivial MR.
Prompt‑based defensive instructions (explicitly instructing models to suppress preferences where inappropriate) reduce misapplication but fail to fully eliminate it.
Ablation experiments adding prompt‑based safety/defenses to model inputs and measuring MR and AAR; defenses produced reductions in MR but residual misapplication remained.
Attempts to mitigate misapplication with stronger reasoning prompts (e.g., chain‑of‑thought) reduce Misapplication Rate but do not eliminate it.
Ablation applying reasoning prompts and chain‑of‑thought style instructions to models, comparing MR before and after; reported reductions in MR but persistence of non‑zero MR across scenarios.
Models that more faithfully enforce stored preferences achieve higher Appropriate Application Rate (AAR) but also systematically have higher Misapplication Rate (MR), indicating a trade‑off between correct personalization and harmful over‑application.
Ablation experiments varying strength of preference encoding and measuring resulting AAR and MR per model; quantitative comparisons across models showing positive correlation between stronger preference adherence and both higher AAR and higher MR.
Reviving model-based central planning tools (ISB+NDMS) risks political-economy problems and requires evaluation of efficiency and flexibility compared to market coordination.
Analytic discussion and normative argument in the paper; no empirical comparative study provided.
Russia's digitalization and adoption of AI/Big Data are reshaping the country's socio-economic infrastructure in multifaceted and systemic ways.
Qualitative analysis of national strategies and policy documents plus the author's expert assessments; no sample size or statistical testing reported.
Finance, Education, and Transportation show mixed dynamics: both displacement of routine tasks and creation of new hybrid roles.
Descriptive sectoral analyses from the simulated dataset (hybrid share, task-displacement indicators, employment changes) covering Finance, Education, Transportation (2020–2024), plus mixed-evidence studies from the literature synthesis (ACM/IEEE/Springer 2020–2024).
Overall, economic benefits from AI in radiology are plausible but conditional on human-AI interaction design, governance, workforce effects, and payment structures; net value is not determined by algorithmic accuracy alone.
Synthesis of the heterogeneous literature (laboratory, reader, observational, qualitative) and conceptual economic analysis highlighting dependencies beyond algorithmic performance.
The net effect of AI on clinician burnout is ambiguous: tools can remove tedious tasks but may introduce new cognitive, administrative, and liability stresses.
Mixed qualitative and small-scale observational studies with variable findings on burnout-related measures after AI introduction.
Changes in workload composition can reduce routine burdens but may shift cognitive load to follow-up decisions and managing AI outputs.
Observational and qualitative studies of deployed systems reporting redistribution of tasks and clinician-reported changes in cognitive demands.
Economic outcomes depend on complementarity versus substitution: AI that augments radiologists can raise output per worker; AI that substitutes tasks may reduce demand for certain diagnostic activities.
Theoretical economic frameworks and case studies of task reallocation in early deployments; empirical workforce-impact studies limited.
Automation bias can increase undue reliance on AI, while algorithmic aversion can drive underuse of helpful tools.
Cognitive and behavioral studies and reader simulations demonstrating both increased acceptance/overtrust in automated outputs in some settings and rejection/discounting of AI advice in others.
Real clinical value depends critically on how AI tools interact with radiologists in practice (integration design and human-AI interaction).
Conceptual models and synthesis of reader studies, simulation/interaction studies, usability and qualitative deployment evaluations that compare standalone algorithm performance versus clinician+AI workflows.
Explicit governance reduces negative externalities (bias, privacy breaches, loss of trust) but entails compliance costs that should be factored into adoption and diffusion models.
Conceptual claim synthesizing trade‑off arguments from governance and risk literatures and practitioner examples; not measured empirically in the paper.
Embedding AI into workflows may change firm boundaries (e.g., outsourcing models vs. in‑house systems) and make investments in internal auditability and explainability strategic assets.
Theoretical implication drawn from synthesis of organizational boundary theory and practitioner trends; suggested rather than empirically demonstrated within the paper.
Realizing economic gains at scale from AI in drug R&D is constrained by data quality and access, high implementation and integration costs, regulatory uncertainty, and ethical/legal concerns; these constraints will shape how gains are distributed across firms, countries, and patients.
Aggregate conclusion of the narrative review synthesizing documented benefits and recurring constraints from published studies, case reports, industry/regulatory analyses; qualitative synthesis without quantitative projection of distributional outcomes.
Adoption of AI in pharma will increase demand for computational biologists, ML engineers, and data scientists and may displace or redefine some traditional bench roles.
Labor-market trend reports and organizational case studies included in the review noting hiring patterns and role changes; qualitative synthesis rather than comprehensive labor-market study.
AI could lower discovery costs and permit more entrants in niche/specialty therapy discovery, but clinical development costs remain a major barrier to entry.
Synthesis of reported reductions in early-stage discovery costs and persistent high clinical trial costs from studies and industry reports; heterogeneous evidence across therapeutic areas.
Upfront capital and proprietary data requirements may advantage large incumbents or well-funded startups and could increase market concentration unless data-sharing or open platforms emerge.
Market-structure analysis and industry examples in the narrative review; inference based on observed data-asset advantages and investment needs across firms.
AI shifts the cost structure of drug R&D toward higher fixed costs (data infrastructure, compute, ML talent) and potentially lower marginal costs for candidate generation and some preclinical activities.
Economic synthesis and industry reports in the review describing capital-intensive investments and reduced per-unit costs in algorithmic candidate generation; largely conceptual and based on case examples.
Two opposing market forces will act: (a) democratization lowering entry barriers for startups, and (b) concentration where firms with premium proprietary data and integrated AI capture outsized returns.
Conceptual economic analysis and illustrative industry observations; no empirical market-structure measurement presented.
AI (including machine learning, generative AI, and NLP) is reshaping biomedical research and pharmaceutical R&D by creating distinct adoption archetypes within large pharmaceutical companies.
Editorial / conceptual synthesis using qualitative analysis and archetype classification based on cross-industry observations and illustrative examples; no systematic measurement or sample size reported.
Cross-DAO cooperation could reduce duplication and accelerate global public-good R&D (e.g., neglected diseases) but raises jurisdictional, regulatory arbitrage, and equity concerns.
Theoretical discussion and scenario analysis; no cross-DAO empirical case with measured outcomes is provided.
There is potential for timely, personalized interventions (nudges/warnings) that could reduce harm, but causal evidence of long‑term effectiveness is limited.
Many studies propose or evaluate intervention prototypes and report feasibility/short‑term outcomes, while the review notes scarce randomized or longitudinal evaluations measuring welfare outcomes.
Model transparency received 90% approval but still requires further refinement.
Stakeholder validation reporting a 90% approval rate for model transparency, while the authors note transparency needs additional work. (Summary does not specify transparency criteria or evaluation method.)
Ethical governance received 85% approval but requires further refinement.
Stakeholder validation results showing 85% approval for ethical governance aspects, with the paper noting the need for further refinement. (No details given on stakeholder composition or ethical framework used.)
Human capital is no longer defined solely by formal education or accumulated experience; it increasingly takes the form of a multidimensional system in which cognitive abilities, digital competencies, social and communicative skills, and ethical awareness interact and reinforce one another.
Result of the paper's synthesis combining systemic analysis and comparative assessment of international practices; conceptual/qualitative evidence rather than quantified measurement across populations.
Ongoing digital transformation and the widespread adoption of artificial intelligence are reshaping the formation, structure, and practical use of human capital in modern economies.
Paper's core analytical conclusion based on systemic analysis, comparative assessment of international practices, and analytical generalization of organizational learning models; no primary quantitative sample size or experimental data reported.
As compute costs decline, pro-price-competitive policies may lose their effectiveness in improving consumer surplus, while compute subsidies may shift from ineffective to effective.
Comparative statics within the theoretical model tracking how policy effects on consumer surplus change as the model parameter for compute cost is decreased.
Pro-quality-competitive policies increase the provider's profits while reducing the downstream firms' profits.
Model equilibrium analysis indicating that enhancing downstream quality competition shifts surplus toward the provider (higher provider profit) while lowering downstream firms' profits in the modeled equilibria.
Compute subsidies are effective at improving consumer surplus only when compute or data preprocessing costs are low.
Model analysis and comparative statics in the paper: introducing compute subsidies raises consumer surplus in parameter regions where compute/preprocessing costs are low.
Policies that promote price competition in downstream markets boost consumer surplus only when compute or data preprocessing costs are high.
Comparative-static results from the game-theoretic model showing that pro-price-competitive policy interventions increase consumer surplus under parameter regimes where compute or data preprocessing costs are high.
Factors identified as relevant to AI emergence/adoption include Technology Adoption Rate (AI1), Government Policies and Regulations (AI2), Labor Market Dynamics (AI3), Technological Advancements (AI4), Corporate Strategies (AI5), and Socio-cultural Factors (AI6).
Author-provided list of factors in the paper; no empirical quantification, weighting, or methodology for selecting these factors is given in the excerpt.
The maturity of an organization's data governance framework influences the success of AI and Big Data in lowering market uncertainty.
Findings from the qualitative case studies and overall analysis highlighting organizational data-governance maturity as a moderating factor (no standardized maturity measure or sample breakdown provided in the summary).
The stringency of the regulatory environment moderates how effectively AI and Big Data reduce market uncertainty.
Moderation identified via the study's analysis and case studies (specific regulatory measures and empirical tests not detailed in the summary).
The effectiveness of AI and Big Data in reducing market uncertainty is contingent upon industry type.
Observed variation across industries in the paper's qualitative case studies and analysis (the summary does not specify which industries or comparative sample sizes).
These findings have important implications for understanding how political ideology may influence party members’ perspectives on AI in relation to labor markets, job losses, and regulation in OECD countries.
Interpretive implication drawn by the authors from their reported results (synthesis rather than a new empirical claim).
Political ideology shapes party members’ positions on AI education and training programs intended to assist workers in environments where AI is more prevalent.
Inferred finding stated by the authors based on content analysis of party member statements; the excerpt indicates the authors examined positions on AI education/training but does not provide specific results or metrics.
Political ideology significantly affects party members’ views on the need for government regulations to protect workers from labor market disruptions caused by AI.
Reported finding from the paper's content analysis of media interviews, speeches, and debates by party members in OECD countries (2016–2025); details on coding categories, inter-rater reliability, and quantitative significance measures are not included in the excerpt.
Political ideology significantly affects party members’ concerns regarding AI-related job losses.
Result reported by the authors based on content-analysis of party member comments and statements across OECD countries (2016–2025); specific analytic procedures, coding scheme, sample size, and statistical tests are not provided in the excerpt.
Artificial intelligence (AI) is poised to transform the distribution and sources of income.
Analytical assertion in the paper (theoretical/policy analysis); no empirical data or specific study citations provided in the excerpt.
AI has emerged as a transformative force that influences economic systems, institutional functions, and daily human behaviors.
Stated as an overarching observation in the paper (theoretical/interpretive claim); no empirical methods or sample sizes are reported in the abstract.
Improvements in caseworker accuracy level off as chatbot accuracy increases (an "AI underreliance plateau").
Observed pattern in experimental results: incremental gains in caseworker accuracy diminish at higher chatbot accuracies, described by authors as an 'AI underreliance plateau' (specific curves or thresholds not in the excerpt).
The rapid global proliferation of Artificial Intelligence (AI) has created a profound paradox: while promising unprecedented productivity gains, its current trajectory exacerbates labor market polarization, deepens inequality, and threatens to fracture the 20th-century social contract.
Asserted in abstract; no empirical methods, datasets, or sample sizes described in the abstract (presumably supported in paper by literature review/argumentation).
AI’s labor market impacts in the Philippines are not technologically predetermined; outcomes will depend on policy choices related to skills development, governance, social protection, and innovation.
Integrated conceptual framework in the paper linking AI capabilities, occupational structure, and institutional mediation, supported by the scenario analysis which shows divergent outcomes conditional on policy settings.