Evidence (2480 claims)
Adoption
5227 claims
Productivity
4503 claims
Governance
4100 claims
Human-AI Collaboration
3062 claims
Labor Markets
2480 claims
Innovation
2320 claims
Org Design
2305 claims
Skills & Training
1920 claims
Inequality
1311 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 373 | 105 | 59 | 439 | 984 |
| Governance & Regulation | 366 | 172 | 115 | 55 | 718 |
| Research Productivity | 237 | 95 | 34 | 294 | 664 |
| Organizational Efficiency | 364 | 82 | 62 | 34 | 545 |
| Technology Adoption Rate | 293 | 118 | 66 | 30 | 511 |
| Firm Productivity | 274 | 33 | 68 | 10 | 390 |
| AI Safety & Ethics | 117 | 178 | 44 | 24 | 365 |
| Output Quality | 231 | 61 | 23 | 25 | 340 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 158 | 68 | 33 | 17 | 279 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 88 | 31 | 38 | 9 | 166 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 105 | 12 | 21 | 11 | 150 |
| Consumer Welfare | 68 | 29 | 35 | 7 | 139 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 71 | 10 | 29 | 6 | 116 |
| Worker Satisfaction | 46 | 38 | 12 | 9 | 105 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 11 | 16 | 94 |
| Task Completion Time | 76 | 5 | 4 | 2 | 87 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 16 | 9 | 5 | 48 |
| Job Displacement | 5 | 29 | 12 | — | 46 |
| Social Protection | 19 | 8 | 6 | 1 | 34 |
| Developer Productivity | 27 | 2 | 3 | 1 | 33 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 8 | 4 | 9 | — | 21 |
Labor Markets
Remove filter
Public‑interest concerns (bias, misuse, systemic risk) may be harder to mitigate via simple transparency rules; policies should emphasize outcome‑based regulations, mandatory behavioral testing, and marketplace disclosure obligations for stressed scenarios.
Policy implication derived from the non‑rule‑encodability thesis; no empirical policy evaluation included.
Standard contracts and regulatory audits that rely on inspection of rule sets or source code will be insufficient to assess model behavior or risk; regulators and buyers must rely more on behavior‑based testing, standards, and outcome measures.
Policy and regulatory argument derived from the main theorem about non‑rule‑encodability; no empirical regulatory studies presented.
Full interpretability via rule extraction may be impossible for the most valuable parts of LLM competence, limiting the utility of some transparency approaches for safety and auditing.
Argumentative consequence of the main theoretical claim and structural mismatch; supported by historical limitations of rule‑based systems; no empirical tests reported.
There is a structural mismatch between explicit human cognitive tools (rules, checklists) and the pattern‑rich, high‑dimensional competence encoded in LLMs.
Theoretical/structural argument about distributed statistical representations in LLMs versus discrete rules; no experimental quantification provided.
Historical expert systems failed to generalize or scale to complex, ambiguous tasks, contrasting with LLMs' broader empirical successes.
Historical case analysis and literature review-style discussion of expert systems versus contemporary LLM performance; no new quantitative historical dataset provided.
LEAFE's benefits depend on informative, actionable feedback; environments with noisy or adversarial feedback may limit improvements.
Limitations stated in the paper noting sensitivity to feedback quality; conceptual reasoning that the method relies on extracting actionable signals from environment feedback.
Outcome-driven post-training (optimizing final rewards) underutilizes rich environment feedback and causes 'distribution sharpening' — policies overfit a narrow set of successful behaviors and fail to broaden problem-solving/recovery capacity in long-horizon settings.
Problem diagnosis in the paper supported by comparison of outcome-driven RL (GRPO) performance versus LEAFE and by conceptual argument about how optimizing final success signals can narrow behavioral support; supported by empirical observations of poorer recovery/generalization in baselines.
If left unchecked, managerial short-termism combined with AI adoption can create a feedback loop where firms cut labor to boost short-term profits, undermining aggregate demand and eroding the market that sustains those profits.
Conceptual macroeconomic and organizational synthesis drawing on theory and historical patterns; no new empirical time-series demonstrating this loop in current AI-driven layoffs.
Work-time reduction policies carry distributional and implementation risks (heterogeneous effects by occupation, firm size, capital intensity; risk of hidden wage cuts) that require careful compensation rules and monitoring.
Theoretical reasoning and references to heterogeneous outcomes in prior work-hour studies; no new empirical quantification of heterogeneity in AI-era implementations.
Lower household demand resulting from payroll cuts can precipitate further cost-cutting and automation, creating a self-reinforcing feedback loop that risks persistent demand shortfalls and higher structural unemployment.
Theoretical models of demand-driven adjustment and cited historical patterns; conceptual argument rather than empirical causal identification in contemporary AI contexts.
AI-justified layoffs are driven more by managerial short-termism and misaligned executive incentives than by immediate technological necessity.
Interdisciplinary conceptual synthesis drawing on labor-economics theory, organizational behavior literature linking executive compensation/short-termism to layoffs, and selected prior empirical studies; no new firm-level causal identification or large-scale dataset provided.
Distributional impacts of AI are uneven: younger workers and individuals with lower formal education face greater disruption.
Descriptive breakdowns of occupational vulnerability and employment changes by demographic groups (age and education) derived from labor statistics and vulnerability mapping; supported by qualitative case observations. Exact subgroup sample sizes not given.
Routine service and administrative occupations show the highest vulnerability to automation and displacement from AI.
Occupational vulnerability mapping using task/routine exposure methods and descriptive employment trend analysis across occupations; supported by employer survey responses and case-study observations. Sample sizes for surveys/mapping not provided in summary.
Passive monitoring and predictive models are insufficient for governing the complex dynamics of a tech-driven economy.
Conceptual critique based on economic cybernetics literature and the author's expert assessment; no empirical test comparing governance regimes is provided.
Digitalization is deepening digital inequality (unequal access to digital tools, skills, and benefits) across social groups and regions.
Qualitative analysis and expert assessment; the paper calls for new metrics but does not present systematic empirical measures of inequality.
Digital transformation can generate technological unemployment if not managed with appropriate retraining and social protection measures.
Expert assessment and literature-informed argumentation in the paper; no empirical longitudinal analysis isolating technology-driven job losses presented.
Forced or poorly regulated digitalization risks exacerbating social stratification.
Conceptual argument supported by qualitative analysis of policy documents and expert assessment; no empirical causal estimates provided.
Manufacturing and Retail experienced net employment contractions attributable mainly to task automation and substitution.
Simulated employment-level series and net change calculations by sector (Manufacturing, Retail) across 2020–2024 in the paper's dataset, together with literature-derived mechanisms emphasizing automation/substitution in these sectors (systematic review of selected publishers 2020–2024).
Explainability, trust, and demonstrated real-world effectiveness are key demand-side frictions; small-scale laboratory gains rarely translate into broad clinical uptake without workflow fit.
Adoption studies, qualitative interviews with clinicians and purchasers, and observations that many high-performing lab models see limited clinical use due to workflow and trust issues.
Hidden costs can arise from increased liability exposure, workflow redesign burden, and potential productivity loss during transition periods.
Qualitative deployment studies and procurement narratives reporting unanticipated legal, operational, and productivity impacts during early rollouts.
Human-AI collaboration can also generate harms, including automation bias, deskilling, and workflow disruption.
Behavioral laboratory experiments, simulation/reader studies demonstrating automation bias, qualitative reports and observational deployment accounts documenting workflow frictions and concerns about reduced trainee exposure.
Trust, verification costs, and legal/governance requirements remain consequential even with AI mediation and may limit or shape adoption.
Theoretical discussion of governance and verification costs; no empirical measurement of these costs in adopter firms provided.
AI-mediated interpretation and action carry risks related to quality, bias, and misalignment, which can produce miscommunication or incorrect automated actions.
Paper's discussion section raising caveats; conceptual risk analysis without empirical incident data; references to general concerns in AI safety literature (no new empirical evidence provided).
Despite positive outcomes, challenges such as workforce displacement, ethical concerns, and limited access to AI technologies were identified as barriers to full adoption.
Study respondents reported barriers in the survey; descriptive statistics summarized the prevalence of workforce displacement concerns, ethical issues, and limited access to AI technologies as impediments to broader adoption.
There is a growing tension between relatively rigid education and training systems and the rapidly changing skill requirements of digitally driven labor markets.
Argument motivated and supported by comparative assessment of international practices and systemic analysis; descriptive/comparative evidence rather than quantified empirical testing.
Analyses of online job postings indicate significant declines in demand for highly automatable and entry-level roles.
Empirical studies using online job-posting data described in the paper (methods: job-posting frequency/trend analysis; sample size/timeframe not specified in the excerpt).
Since the public release of ChatGPT in November 2022, concerns regarding job displacement, wage reduction, and labor market restructuring have intensified.
Temporal observation in the paper referencing heightened public and policy concerns after ChatGPT's release; based on cited literature and discourse (no sample size given).
Low‑skill installation and maintenance jobs have increased, but wage levels and upward mobility for these jobs remain lower than those in high‑skill industries.
Finding reported from the literature review and cited reports/studies indicating growth in low‑skill installation/maintenance employment alongside comparative analyses of wages and career mobility; no specific datasets or sample sizes provided in the summary.
Job polarization is occurring in solar power plants as a result of automation or digital transformation and changes in required skill sets.
Synthesis from the systematic literature review and referenced reports/studies indicating links between automation/digitalization and occupational shifts in solar plants; specific studies and sample sizes not provided in the summary.
The paper highlights that urgent policy intervention is required to reestablish a balance between the benefits of AI and the ethical ramifications that arise from these technologies, with a particular emphasis on job displacement.
Author conclusion drawn from the stated literature-based analysis; the excerpt does not list the specific studies, empirical findings, or criteria used to reach this policy recommendation.
There has been an increase in the level of concern regarding the ethical implications arising from the automation of tasks and the subsequent job displacement due to AI.
Author statement based on a review of (unspecified) novel studies and existing literature; no empirical sample size, instrumentation, or quantitative measure of 'concern' reported in the provided text.
The limitations of systems that prioritize academic pathways constrain workforce adaptability and inclusive labor market development.
Argument based on synthesis of empirical studies and secondary data connecting education pathway composition to workforce adaptability and inclusiveness (presented as a policy-relevant conclusion rather than a quantified causal estimate).
Skills mismatch in the labor market is structural and linked to education systems that prioritize academic pathways without adequate support for vocational and continuing training.
Integrated interpretation of comparative evidence and secondary data showing imbalances between academic and vocational provision and associated labor-market frictions (paper frames this as a structural conclusion; specific causal tests not described in the summary).
Expansion of intermediate vocational skills has been limited relative to the expansion of higher education.
Comparative evidence and secondary data showing smaller increases in intermediate vocational qualifications compared with higher education attainment (specific metrics/country coverage not provided in the summary).
The risk to the tax system is heightened by the federal government’s dependence on individual labor income even as economic value shifts toward mobile capital and AI ownership by large firms.
Analytical claim in the paper linking tax base dependence to shifts in economic value; no empirical measurement of 'mobile capital' or quantified shift included in the excerpt.
AI threatens to disrupt the tax system’s ability to fulfill its fundamental goals of raising revenue, redistributing income, and regulating taxpayer behavior.
Normative/policy argument made in the paper (no empirical testing or quantified projections provided in the excerpt).
These AI-driven outcomes will have far-reaching impacts on the federal tax system, which heavily relies on taxing individual labor income and payroll rather than capital or consumption.
Paper's policy analysis asserting the composition of federal tax reliance (no revenue breakdowns or statistical evidence included in the excerpt).
Even under optimistic projections, AI is expected to exacerbate wealth inequality because ownership and immense value are concentrated within a subset of Big Tech companies and AI startups.
Argumentative claim in the paper asserting concentration of ownership and value in certain firms; no empirical measures or firm-level data presented in the excerpt.
Some experts predict widespread job displacement due to AI.
Statement in the paper referencing expert predictions (no specific experts, studies, or sample sizes cited in the excerpt).
Perceived autonomy amplifies the negative effects of perceived algorithmic behavioral constraint on riders' outcomes (i.e., strengthens the adverse impact on mental health and risky riding via work pressure).
Moderation results from SEM and bootstrapping on a sample of 466 Chinese food delivery riders showing interaction between behavioral constraint and perceived autonomy increases negative indirect effects through work pressure.
Perceived autonomy enhances the positive effect of perceived algorithmic standardized guidance in reducing risky riding behavior.
SEM moderation analysis with bootstrapping on data from 466 Chinese food delivery riders showing perceived autonomy strengthens the standardized guidance -> work pressure -> risky riding indirect pathway.
Perceived algorithmic standardized guidance reduces risky riding behavior among food delivery riders by reducing work pressure.
Survey of 466 Chinese food delivery riders analyzed with SEM and bootstrapping showing standardized guidance -> work pressure -> risky riding behavior (indirect effect).
Perceived algorithmic behavioral constraint impairs food delivery riders' mental health through increased work pressure.
Survey of 466 Chinese food delivery riders analyzed via SEM and bootstrapping with work pressure as mediator (behavioral constraint -> work pressure -> mental health).
Perceived algorithmic tracking evaluation impairs food delivery riders' mental health through increased work pressure.
Survey data from 466 Chinese food delivery riders analyzed with structural equation modeling (SEM) and bootstrapping; work pressure modeled as mediator based on the Job Demands-Resources (JD-R) framework; indirect effect from tracking evaluation -> work pressure -> mental health reported.
Global AI governance, regulatory fragmentation, and the effects of privacy laws on market competition are under-studied areas.
Low topic prevalence for topics corresponding to global governance, regulatory fragmentation, and privacy-law effects on competition in the >4,600-paper corpus as identified by topic modeling and policy-alignment analysis.
The economic impacts of risk-based AI regulations are under-studied in the current literature.
Topic-modeling indicates few papers focusing on economic impacts of risk-based regulation; authors' crosswalk with policy documents shows this as a gap.
Research on effective industrial policy for AI is relatively underexplored.
Low prevalence of industrial-policy-related topics in the topic-modeling output and comparison to stated policy priorities in national AI strategies and legislation across regions.
There are notable gaps in the literature in measuring AI-driven economic growth.
Comparison of topic prevalence from the topic-modeling exercise with policy priorities derived from national AI strategies and legislation across regions, showing low coverage of research explicitly measuring AI-driven economic growth.
Short-run labor market disruptions raise concerns regarding wage inequality and workforce adaptation.
Claims based on observed short-run labor market adjustments in publicly available data and theoretical implications for inequality and adaptation; specific empirical measures, time horizons, and sample sizes are not reported in the excerpt.
AI simultaneously increases adjustment pressures for routine tasks.
Argument and cited observations from publicly available labor market data indicating displacement or adjustment in routine-task-intensive occupations (no specific empirical estimates or samples provided).