Evidence (1286 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Inequality
Remove filter
Current regulations fall short in effectively protecting children in an evolving digital landscape; there are persistent gaps and a growing need for internationally coordinated approaches.
Conclusion presented in the book's comparative legal analysis; implies review of EU (and US) legal frameworks and identification of gaps, but the excerpt does not list the analytical method, jurisdictions reviewed in detail, or specific legal provisions examined.
Europe has emerged as a major hub for hosting child sexual abuse material (CSAM), including newer forms such as deepfake abuse content and AI-generated 'DeepNudes.'
Asserted in the summary; would be supported by law-enforcement takedown data, hosting statistics, or forensic analyses of seized material, but the excerpt provides no specific datasets, agencies, or sample sizes.
Violations of privacy, exposure to disturbing content, unwanted sexual approaches, and cyberbullying are becoming more common.
Trend claim made in the book summary; would be supported by longitudinal or comparative prevalence data on online harms, but no specific studies, methods, or sample sizes are cited in the provided text.
Nearly one in three reports feeling unsafe.
Specific prevalence statement included in the summary; implies self-report survey data on perceived safety among youth, but the excerpt does not identify the survey instrument, population, timeframe, or sample size.
The scalability of the Photo Big 5 enables new academic insights into the role of personality in labor markets, but its growing use in industry screening raises important ethical concerns regarding statistical discrimination and individual autonomy.
Argument in the paper based on the methodological scalability (AI + large LinkedIn microdata) and observed predictive links to labor-market outcomes; authors raise normative concerns about industry adoption and implications for discrimination and autonomy.
What remains needed is rigorous advice to policymakers concerned about rapid increases in labor churn, scientific development, labor–capital shifts, or existential risk.
Normative conclusion drawn by the author from gaps identified in the seven-book review (qualitative assessment of unmet policy-relevant analysis); sample = 7 books.
The reviewed works offer little guidance regarding the transformative scenarios considered plausible by many AI researchers.
Author's evaluative judgment based on the content and emphases of the seven books (qualitative gap analysis); sample = 7 books.
Gendered perceptions of AI's social and ethical consequences, rather than access or capability, are the primary drivers of unequal GenAI adoption.
Comparative model results from the 2023–2024 nationally representative UK survey showing perceptions (societal-risk index) have greater explanatory/predictive power than measures of access (e.g., device/internet access) or capability (digital literacy, education).
Intersectional analyses show the largest gender disparities in GenAI use arise among younger, digitally fluent individuals with high societal risk concerns, where gender gaps in personal use exceed 45 percentage points.
Subgroup (intersectional) analysis of the nationally representative 2023–2024 UK survey data stratified by age, digital fluency, and societal-risk concern levels; reported gender gap >45 percentage points in specified subgroup.
The societal-risk concerns index ranks among the strongest predictors of GenAI adoption for women across all age groups, surpassing digital literacy and education for young women.
Multivariable models and predictor ranking using the 2023–2024 UK survey data showing relative predictive strength of the concerns index versus measures of digital literacy and education, with subgroup (age × gender) comparisons.
The societal-risk concerns index explains between 9 and 18 percent of the variation in GenAI adoption.
Regression/statistical models using the composite concerns index as a predictor of GenAI adoption in the nationally representative 2023–2024 UK survey; reported explained variation (9–18%).
Women adopt GenAI less often than men because they perceive its societal risks differently.
Statistical analysis linking a constructed composite societal-risk concerns index (mental health, privacy, climate impact, labor market disruption) to GenAI adoption, using the UK 2023–2024 survey; models compare explanatory power of perceptions versus access/capability variables.
Women adopt GenAI substantially less often than men.
Analysis of the 2023–2024 nationally representative UK survey data comparing personal use/adoption rates by gender.
There are ethical concerns surrounding AI and automation including algorithmic decision-making, workforce exclusion, and inequality in access to reskilling opportunities.
Raised as an ethical analysis within the paper's conceptual framework; no empirical study, surveys, or quantified measures of these ethical issues are reported in this paper.
AI is eliminating repeated (routine) jobs.
Stated as part of the paper's argument about AI's dual impact; supported by conceptual analysis rather than new empirical evidence in this manuscript (no sample size or empirical method reported).
Artificial intelligence and automation are reshaping jobs, transforming them from a steady source of income to a dynamic process highly influenced by technology, flexibility, and uncertainty.
Central analytical claim made in the paper based on conceptual reasoning; the paper does not report empirical measures, datasets, or sample sizes to support the transformation quantitatively.
AI and automation pose significant challenges to employment stability, skill relevance, and human dignity.
Claim presented within the paper's conceptual and analytical discussion of AI's dual impacts; no empirical study, sample size, or quantitative measures provided in this paper.
There are challenges to adopting AI in HRM within IT firms.
Identified through the literature review and the empirical study involving HR professionals; the summary notes challenges but does not enumerate or quantify them.
AI use also poses risks, including systemic discrimination, privacy invasion, and commodification of talent.
Qualitative synthesis and documented instances in the reviewed literature (n=85) reporting discriminatory outcomes, privacy concerns, and labor commodification effects associated with algorithmic HR tools.
Qualitative synthesis reveals a 'gray zone' in labor relations and a 'black box' in algorithmic data processing, both exposing businesses to procedural injustice risks.
Thematic/qualitative synthesis of findings from the reviewed literature (n=85) highlighting issues of labor relations and algorithmic opacity leading to procedural fairness concerns.
Digital transformation raises challenges related to privacy, inequality, and regulatory scrutiny.
Identified as a key challenge in the paper; the abstract provides no details on how privacy concerns, inequality measures, or regulatory incidents were documented or quantified.
Traditional methods for assessing and developing employees' skills often fail to provide real-time feedback.
Statement supported by literature review cited by the authors; the abstract does not provide empirical comparisons, metrics, or sample sizes.
Skills mismatch and SME adoption constraints constitute a binding bottleneck for inclusive digital–green upgrading.
Synthesis of studies on skills, firm capabilities, and SME adoption of digital and green technologies (review-level evidence; no single dataset or sample size provided).
Absent complementary institutions and infrastructure, digitalization may increase electricity demand, widen inequality, and incentivize strategic disclosure (greenwashing).
Literature review drawing on empirical studies of energy consumption from digital systems, labor-market studies, and analyses of ESG disclosure practices (review-level synthesis; no single sample size reported).
Occupational sorting explains a somewhat larger share of the gender gap in Ireland than in other European countries, but a substantial portion remains unexplained, pointing to possible unobserved structural, cultural or organisational factors specific to the Irish labour market.
Decomposition analysis for Ireland using ESJS data showing occupation contributes more to the explained component in Ireland than on average, while the unexplained residual remains large.
Gender gaps are larger and less well explained by observable characteristics among younger cohorts (aged under 35), implying under-representation of women in advanced digital roles is emerging early in careers.
Age-cohort subgroup regressions and decomposition analyses on ESJS data comparing explained/unexplained gaps for workers aged under 35 versus older cohorts.
Gender disparities widen significantly at the very upper end of the distribution of digital job intensity — a 'digital glass ceiling' — while lower and middle levels show more modest differences.
Distributional analysis of the Job Digital Intensity Index (JDII), constructed from ESJS digital task items, showing larger gender gaps at the upper tail of the JDII distribution.
AI causes job loss due to the automation of repetitive tasks.
Narrative literature review and synthesis of recent economic studies presented in the paper; no original empirical sample or primary data collection reported.
The findings raise ethical concerns about using such models in sensitive selection processes and highlight the need for transparency and fairness in digital labour markets.
Interpretive/concluding claim based on the observed adjective-based gendering and the broader literature on algorithmic fairness; recommendation rather than direct empirical result.
Gendered linguistic patterns emerged in the adjectives attributed to female and male candidates: GPT-5 tended to associate women with emotional and empathetic traits and men with strategic and analytical traits.
Empirical/qualitative analysis of the adjectives and descriptive language in GPT-5's outputs for the 24 simulated profiles; categories reported (emotional/empathetic vs strategic/analytical).
Large language models (LLMs) risk reproducing, and in some cases amplifying, gender stereotypes and bias already present in the labour market.
Framed as an assertion supported by prior literature and used as motivation for the study; partially evaluated empirically in this paper via the GPT-5 experiment.
Developing economies face heightened risks from AI due to large informal sectors, limited reskilling infrastructure, weaker labor mobility, and constrained social protection.
Comparative institutional analysis and application of structural-transformation theory; argument is qualitative and no explicit cross-country regression or representative sample of developing countries is provided in the paper.
Displacement often occurs faster than job creation and worker reallocation, producing transitional unemployment and skills gaps.
Temporal-mismatch argument based on historical patterns of technological adoption and task-based substitution theory; paper synthesizes prior theoretical work rather than presenting new time-series microdata or measured reallocation speeds.
Developing economies are more vulnerable where employment is concentrated in routine or informal tasks and where reskilling, mobility, and institutional buffers are limited.
Comparative consideration of advanced vs developing economies drawing on macro/sectoral indicators, labor market structure discussions, and existing empirical studies cited conceptually.
Creation of new jobs often lags displacement, producing transitional unemployment and reallocation frictions in the short- to medium-term.
Dynamic/task-based theoretical framing and synthesis of empirical evidence on technology adoption episodes showing delayed job creation relative to displacement.
AI disproportionately automates routine and many middle-skill tasks (both manual and cognitive), displacing corresponding occupations.
Synthesis of occupation- and task-level exposure studies and task-based automation literature referenced in the paper (no new empirical sample provided).
Compensation-based frameworks for personal data may advantage those better able to monetize data, potentially worsening inequality.
Theoretical argument and literature synthesis on distributional effects of markets and bargaining power; paper does not present empirical distributional simulations or data.
Data markets tend to concentrate benefits and rents in large platforms while externalizing harms onto individuals and society.
Argument based on descriptive facts about platform business models and literature on market concentration in digital markets; no original econometric concentration analysis provided in the paper.
Standard market-failure fixes (better information, pricing, contracting) are insufficient to address the moral and social-structural harms of commodifying privacy.
Philosophical argument drawing on noxious-markets literature and limitations of informational/contractual remedies; supported by conceptual examples rather than empirical testing.
Harms from data commodification are often externalized, diffuse, and long-term (e.g., profiling, algorithmic discrimination, chilling effects on behavior).
Normative and descriptive synthesis of existing literature on algorithmic harms and privacy externalities; no original longitudinal or causal empirical evidence presented.
Consent in data markets is frequently weak, uninformed, or coerced (due to information asymmetries, complexity, and behavioral biases), undermining the ethical legitimacy of transactions.
Argumentative claim grounded in literature on privacy notice problems, behavioral economics, and descriptive reports on digital consent practices; no new empirical study included in the paper.
Commodifying personal information poses distinctive harms to individuals and social practices, including exploitation, corruption of personal autonomy, distributional injustice, and information asymmetries.
Conceptual analysis supported by literature review across ethics, political philosophy, and descriptive facts about digital-era data practices; uses illustrative examples and secondary sources rather than original empirical data.
Creating a market for personal data is equivalent to making the right to privacy a tradeable right, and such a market should be treated as a 'noxious market' in the sense articulated by Debra Satz.
Normative, conceptual argument applying Satz's noxious-markets framework to personal data; literature review and philosophical argumentation; no original empirical sample or econometric analysis.
Family- and purpose-driven entrepreneurs (motivated by social stability) experienced larger declines in innovation following income shocks than wealth-driven entrepreneurs.
Subgroup quantitative analysis comparing self-reported post-shock innovation activity across identity-defined groups (family/purpose-driven vs. wealth-driven) within the survey sample; outcome measured conditional on reported income shocks.
Stronger internal corporate governance weakens the AI → executive pay relationship, consistent with governance limiting managerial rent capture during technological change.
Moderation analysis in the paper interacting the firm AI indicator with corporate governance measures; results show a smaller AI effect on pay in firms with stronger governance (same sample and regression framework).
Policy levers matter: increasing openness/shared ownership of AI, strengthening rent-sharing (higher ξ), and reducing concentration of complementary assets (antitrust, data portability) can reduce the probability that AI widens aggregate inequality.
Model counterfactuals and policy experiments in the calibrated framework that vary ownership/access parameters, ξ, and asset concentration to show distributional outcomes shift accordingly.
Inflation and geopolitical fragmentation can raise the cost of AI deployment (hardware shortages, supply constraints) and complicate cross-border data flows, slowing diffusion or creating regionalized AI ecosystems.
Conceptual argument linking macroeconomic and geopolitical constraints to AI deployment costs; no empirical cost-accounting or cross-country diffusion analysis provided in the paper.
Mandel's account—that capitalist production relations, class struggle, and global imbalances shape the course and consequences of waves—implies that crises expose and amplify supply-chain fragilities and bargaining conflicts that affect profitability.
Theoretical interpretation of Mandel's political-economy literature and historical examples (qualitative).
One-size-fits-all AI competency approaches fail to account for local labor markets, pedagogical traditions, and resource realities; respondents favor context-aware frameworks allowing discipline-specific adaptation.
Thematic analysis of open-ended responses expressing preferences for context-aware, flexible frameworks; survey items mapped to UNESCO competency frameworks asking about adaptability and local relevance.
Infrastructural limitations (bandwidth, computing resources, licensing costs) disproportionately affect respondents in the Global South and smaller institutions.
Comparative descriptive analysis by region (Global South vs Global North) and institution size/type within the >600 respondent sample; survey items on infrastructure and costs; thematic coding supporting differential impact.