Evidence (1835 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5877 claims
Human-AI Collaboration
5157 claims
Innovation
3492 claims
Org Design
3470 claims
Labor Markets
3224 claims
Skills & Training
2608 claims
Inequality
1835 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 736 | 1615 |
| Governance & Regulation | 664 | 329 | 160 | 99 | 1273 |
| Organizational Efficiency | 624 | 143 | 105 | 70 | 949 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 348 | 109 | 48 | 322 | 836 |
| Output Quality | 391 | 120 | 44 | 40 | 595 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 275 | 143 | 62 | 34 | 521 |
| AI Safety & Ethics | 183 | 241 | 59 | 30 | 517 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 105 | 40 | 6 | 187 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 78 | 8 | 1 | 151 |
| Regulatory Compliance | 69 | 64 | 14 | 3 | 150 |
| Training Effectiveness | 81 | 15 | 13 | 18 | 129 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
Inequality
Remove filter
Unless governments develop industrial policy strategies centered on strengthening democratic economic governance, they risk consolidating corporate control of critical technologies.
Main argumentative claim of the paper as stated in the abstract/introduction; presented as a normative risk argument supported in the paper by conceptual analysis and review of policy trends and historical examples (no empirical sample size reported in the excerpt).
Under-represented groups tend to be systematically under-observed because of historical exclusion and selective feedback, which exacerbates uncertainty for those groups.
Conceptual claim supported by illustrative examples (e.g., lending context) and simulations demonstrating selective feedback effects; literature citation likely included in paper.
Policies that ignore the unobserved (counterfactual) space can harm decision makers (via unrealized gains or losses) and subjects (via compounding exclusion and reduced access).
Theoretical argumentation and illustrative examples (e.g., loan denial counterfactuals) and modelled simulations showing downstream harms when ignoring unobserved outcomes.
Experiments on simulated data with varying bias show that unequal uncertainty and selective feedback produce disparities across groups.
Simulation experiments described in the paper manipulate bias and feedback patterns and report resulting group disparities (synthetic datasets; experiment details in methods/results sections).
A threat model taxonomy mapping misuse vectors to hardware, software, institutional, and liability layers illustrates why no single governance mechanism suffices.
Threat model taxonomy developed in the paper (conceptual taxonomy; illustrative mapping rather than empirical testing).
Restricting access to open-weight models deepens asymmetries while driving proliferation into unsupervised settings.
Argumentation and threat-model reasoning in the paper describing likely consequences of restrictions (theoretical analysis; no empirical sample cited).
Access restrictions, without governed alternatives, may displace risks rather than reduce them.
Theoretical argument and threat-model analysis in the paper showing possible risk displacement (conceptual reasoning; no empirical sample reported).
Disparities emerge and compound across stages of the ML pipeline (training data, model predictions, and post-processing).
Pipeline-level analysis reported in paper showing sources of disparity at multiple stages and how effects accumulate from training data through prediction to post-processing.
Post-processing amplifies these disparities by collapsing heterogeneous probabilities into percentile-based risk tiers.
Analysis of the pipeline showing that converting model probabilities into percentile-based risk tiers (post-processing step) increases observed disparities across demographic groups.
Older and female students with comparable dropout risk are under-identified by the EWS.
Audit comparison showing lower identification/flagging rates for older and female students who have comparable modeled or observed dropout risk to other groups; reported as part of the pipeline disparities analysis.
Younger, male, and international students are disproportionately flagged for support by the EWS, even when many ultimately succeed.
Empirical results from the replica-based audit comparing model predictions and post-processing flags against eventual student outcomes; disparities reported by demographic groups (age, gender, residency). Exact sample size and numerical metrics not provided in the abstract.
Recent policy and academic discourse has increasingly acknowledged the infeasibility of fullstack AI sovereignty, but has not yet provided an integrating theoretical architecture for governing dependence under these conditions.
Literature/policy-discourse claim made in the paper (review/interpretation). No empirical sampling or quantitative evidence reported in the provided text.
The concentration of AI-related infrastructures is coalescing into distinct geocognitive power poles whose competing infrastructural ecosystems generate structural asymmetries that position small and medium-sized states within regimes of cognitive-informational dependence.
Theoretical/geopolitical argument introduced in the paper (conceptual framing). No empirical sample size or quantitative measurement provided in the excerpt.
There is a growing concentration of computational capacity, data ecosystems, and advanced model architectures within a limited number of technological actors, signaling the emergence of a cognitive-informational order in which influence is exercised through the architectures that shape how knowledge is generated, interpreted, and operationalized.
Theoretical/observational assertion in the paper (conceptual synthesis). No empirical details, sample sizes, or quantitative analyses provided in the supplied text.
The policy and research challenge posed by platform-mediated automation is not merely job quantity (technological unemployment) but institutional continuity — how societies reproduce practical competence when platforms optimize for efficiency rather than formation.
Normative and conceptual claim developed through literature synthesis (institutional economics, platform governance, workforce development); presented as an analytical reframing rather than an empirically tested hypothesis.
Entry-level roles have historically functioned as apprenticeships in which workers acquire tacit knowledge and critical judgment; if platforms curtail these formative occupational layers, organizations may lack future workers capable of exercising contextual reasoning required to manage complex systems.
Institutional economics and workforce development literature cited in the paper; conceptual synthesis without original empirical measurement reported.
Platform-mediated automation risks hollowing out labor structures from both directions: eroding repetitive, junior roles from below and automating supervisory coordination functions from above.
Theoretical argument synthesizing institutional economics and platform literature; articulated as a conceptual risk rather than demonstrated with original empirical data.
Algorithmic systems are displacing routine tasks across both low-wage entry-level work and middle-management functions.
Stated in paper's argumentation; supported by a literature-based review drawing on platform governance literature and recent research on AI-enhanced automation (no original empirical sample or quantitative study reported).
There exist inequalities in the emergence of algorithmic bias and in transparency of these systems.
Paper states that inequalities and lack of transparency were observed/identified (citing Memarian, 2023; Bello, 2023; Gambacorta et al., 2024) and discusses these as findings.
Algorithmic bias in automated credit scoring systems may block marginalized groups from accessing financial services.
Explicit statement in the introduction citing prior literature (Agboola, 2025; Nwafor et al., 2024; Oguntibeju, 2024) and motivating the study.
Platforms can exploit workers' uncertainty about the cost of labor to effectively suppress wages.
Interpretation / implication drawn from the theoretical model and the result that a platform can achieve coverage while paying only O(log(M)/M) fraction of total labor cost under assumptions about workers' cost estimates.
There exists a simple pricing strategy for the platform that covers all M tasks with wait time O(M) while paying only an O(log(M)/M) fraction of the total cost of labor.
Theoretical result from the paper's posted-price procurement model under stated assumptions on workers' estimated costs; formal analysis/proof showing existence of such a pricing strategy for general M (no empirical sample).
The authors identify five 'decoys' that seemingly critique—but in actuality co-constitute—AI's emergent power relations and material political economy.
Analytical contribution of the paper: identification and conceptual description of five decoys based on literature synthesis; this is a descriptive/theoretical taxonomy rather than an empirical enumeration with sample size.
Decoys contribute to the network-making power that is at the heart of the Project's extraction and exploitation.
Theoretical synthesis and interpretive argument grounded in literature across relevant fields; the paper posits a mechanism (decoys → strengthened networks → increased extraction/exploitation) but provides no empirical quantification.
Decoys often create the illusion of accountability while masking the emerging political economies that the Project of AI has set into motion.
Conceptual critique supported by literature from communication, STS, and economic sociology; argument that particular practices/instruments function rhetorically to appear accountable while obscuring material political economy. No empirical sample or quantified measures reported.
As AI funders and developers expand their access to resources and configure sociotechnical conditions, they benefit from decoys that animate scholars, critics, policymakers, journalists, and the public into co-constructing industry-empowering AI futures.
Theoretical analysis and literature review; paper identifies and interprets how discursive and institutional phenomena (termed 'decoys') function to produce consent and co-construction of industry-aligned futures. No empirical sample size provided.
Those who fund and develop AI systems operate through and seek to sustain networks of power and wealth.
Conceptual argument and literature synthesis drawing on communication studies, science & technology studies (STS), and economic sociology; no empirical sample reported.
The study identified significant implementation challenges including algorithmic bias, digital divide concerns, data privacy risks, and low technology readiness among HR teams in Tier 2 cities.
Synthesis of qualitative case study findings from 4 organizations plus survey responses (N=150) reporting barriers and risks encountered during adoption.
AI can exacerbate occupational polarization, digital exclusion, and discriminatory outcomes when models are trained on biased data or deployed without transparency and accountability.
Thematic synthesis across included studies identifying mechanisms (biased training data, lack of transparency/accountability) linked to negative distributional outcomes (occupational polarization, digital exclusion, discrimination).
Inherent algorithmic opacity and historical data biases tend to give rise to obvious group prejudices based on gender, educational background, age, and regional origin, thereby further exacerbating the structural inequalities that exist in the current employment market.
Claim made in abstract referencing known sources of algorithmic bias (opacity, historical data bias) and listing affected group attributes; presented as a problem motivating the study, without specific empirical statistics in the abstract.
AI adoption is reinforcing existing structural disparities within the BRICS bloc, creating a two‑tier productivity hierarchy (China & India vs. Brazil, Russia & South Africa).
Observed divergence in TFP trajectories and differing links between AI indicators and TC/EC across the five BRICS economies; comparative analysis shows stronger frontier-shifting effects in China and India and weaker or negative effects in the other three economies.
Brazil, Russia, and South Africa experience stagnation or decline in both efficiency and technological advancement over 2005–2023.
Malmquist TFP decomposition (EC and TC) for each BRICS economy showing flat or negative trends in EC and TC for Brazil, Russia, and South Africa during 2005–2023.
AI infrastructure owners may command more wealth and capability than most governments, threatening the future viability or authority of the nation-state.
Futuristic projection based on the paper's modeling and synthesis of wealth/capability concentration under AI; no empirical measures or comparative data versus governments provided in the excerpt.
Universal Basic Income (UBI), evaluated through incentive-structure lens, will default to a pacification mechanism rather than a genuine solution in the absence of a revolutionary threat that historically forced redistribution.
Normative and theoretical analysis of incentive structures and historical mechanisms of redistribution; the excerpt presents this as an argument rather than reporting empirical trials or quantified outcomes.
Unlike previous feudal orders, this one may prove uniquely resistant to revolution because the mechanisms of enforcement (autonomous weapons, AI surveillance, algorithmic propaganda) do not require human cooperation and therefore cannot be undermined by human dissent.
Logical and theoretical claim based on characteristics of AI-enabled enforcement technologies; presented as an argument rather than an empirically tested finding in the excerpt.
Under this emerging order, the vast majority of humanity will lose their political leverage.
Theoretical and historical argument linking concentration of infrastructure control to political disempowerment; no empirical metrics or sample size provided in the excerpt.
Under this emerging order, the vast majority of humanity will lose their labor value.
Claim made via theoretical argument about automation and AI replacing labor value; no quantitative empirical evidence or sample detailed in the excerpt.
This structural transformation could stabilize into a neo-feudal equilibrium in which a vanishingly small class of infrastructure owners wields power comparable to pre-Enlightenment monarchs.
Futuristic projection and normative/historical analogy based on conceptual modeling of class structure under AGI; the excerpt gives no empirical data or formal model outputs.
The convergence of geopolitical fragmentation (democratic decline) and AI-driven economic concentration is producing a structural transformation unprecedented in human history.
Theoretical synthesis and historical comparison; the paper presents this as an argument based on conceptual modeling and historical analogy; no specific empirical test or sample noted in the excerpt.
The post-World War II international order is undergoing an accelerating concentration of economic power driven by advances in artificial intelligence.
Asserted in the paper as an observed trend linking AI advances to concentration of economic power; presented as a conceptual/historical claim without empirical specification in the excerpt.
The post-World War II international order is undergoing geopolitical fragmentation driven by twenty consecutive years of democratic decline.
Stated as a historical/political claim in the paper; implies reliance on democracy-trend data and historical analysis but no specific dataset, method, or sample size provided in the excerpt.
Income inequality, measured by the Gini index, rises moderately in every scenario we examine due to the polarising effect of job losses and wage and capital income increases on the income distribution.
Calculation of Gini index across multiple simulated scenarios using the SWITCH-linked distributional analysis; reported in the report.
The largest average losses are experienced by middle and higher income households, for whom job displacement outweighs any wage or capital income gains. Lower income households also lose, but by much less.
Distributional results from microsimulation (SWITCH) applying scenarioled job displacement, wage and capital effects across income groups; reported in the report.
When these effects are combined, we find an average decline in household disposable income as a result of AI adoption.
Combined scenario simulations incorporating job displacement, wage effects and capital income effects linked to the Irish tax-benefit system using SWITCH; result reported in the report's main findings.
These wage gains are not large enough to counterbalance the average fall in income due to job displacement.
Combined simulation results (displacement + wage effects) using scenario assumptions and microsimulation (SWITCH), reported in the report's distributional analysis.
Those most likely to experience this disruption are found in higher income households, where the share of workers transitioning into unemployment is substantially larger than in lower income families.
Microsimulation (SWITCH) linking simulated job displacement scenarios to household income groups; results reported in the report.
In our central scenario — drawn from credible international estimates — around 7 per cent of current jobs could be displaced in the short–medium run.
Scenario simulation based on international estimates of AI exposure/adoption; central scenario reported in the report (linked to SWITCH microsimulation for distributional analysis).
AI tends to place higher earning and highly educated workers at greater risk of disruption, because the occupations most exposed to AI are predominantly in these groups.
Synthesis of international research on occupational exposure to AI and the report's analysis linking exposure to worker characteristics (education and earnings); presented as descriptive finding in the report.
These dynamics risk trapping workers in a 'low-skill trap'.
Synthesis of observed labour-market polarisation, persistent low-skill segment, and limited reskilling coverage from secondary sources (2020–2024); presented as a likely risk/consequence.
Limited reskilling coverage constrains workers' ability to adapt to AI-driven changes.
Paper reviews official reports and secondary data (2020–2024) indicating low coverage/uptake of reskilling programs in India and links this to limited adaptation capacity.