Evidence (1902 claims)
Adoption
5126 claims
Productivity
4409 claims
Governance
4049 claims
Human-AI Collaboration
2954 claims
Labor Markets
2432 claims
Org Design
2273 claims
Innovation
2215 claims
Skills & Training
1902 claims
Inequality
1286 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 369 | 105 | 58 | 432 | 972 |
| Governance & Regulation | 365 | 171 | 113 | 54 | 713 |
| Research Productivity | 229 | 95 | 33 | 294 | 655 |
| Organizational Efficiency | 354 | 82 | 58 | 34 | 531 |
| Technology Adoption Rate | 277 | 115 | 63 | 27 | 486 |
| Firm Productivity | 273 | 33 | 68 | 10 | 389 |
| AI Safety & Ethics | 112 | 177 | 43 | 24 | 358 |
| Output Quality | 228 | 61 | 23 | 25 | 337 |
| Market Structure | 105 | 118 | 81 | 14 | 323 |
| Decision Quality | 154 | 68 | 33 | 17 | 275 |
| Employment Level | 68 | 32 | 74 | 8 | 184 |
| Fiscal & Macroeconomic | 74 | 52 | 32 | 21 | 183 |
| Skill Acquisition | 85 | 31 | 38 | 9 | 163 |
| Firm Revenue | 96 | 30 | 22 | — | 148 |
| Innovation Output | 100 | 11 | 20 | 11 | 143 |
| Consumer Welfare | 66 | 29 | 35 | 7 | 137 |
| Regulatory Compliance | 51 | 61 | 13 | 3 | 128 |
| Inequality Measures | 24 | 66 | 31 | 4 | 125 |
| Task Allocation | 64 | 6 | 28 | 6 | 104 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 10 | 16 | 93 |
| Worker Satisfaction | 42 | 32 | 11 | 6 | 91 |
| Task Completion Time | 71 | 5 | 3 | 1 | 80 |
| Wages & Compensation | 38 | 13 | 19 | 4 | 74 |
| Team Performance | 41 | 8 | 15 | 7 | 72 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 17 | 15 | 9 | 5 | 46 |
| Job Displacement | 5 | 28 | 12 | — | 45 |
| Social Protection | 18 | 8 | 6 | 1 | 33 |
| Developer Productivity | 25 | 1 | 2 | 1 | 29 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 7 | 4 | 9 | — | 20 |
Skills Training
Remove filter
Generative AI introduces risks such as model hallucinations and potential erosion of human skills over time.
Practitioner interview reports and authors' interpretive synthesis; qualitative evidence from consulting firms describing hallucination incidents and concerns about reduced skill practice. No longitudinal or quantitative measurement reported.
Current AI tooling often mismatches existing team workflows and CI/CD pipelines, reducing seamless adoption.
Qualitative observations and practitioner reports from the Netlight study describing tooling and workflow frictions; specific integrations or lack thereof discussed but not quantitatively evaluated.
Generated code can introduce security vulnerabilities and licensing/IP ambiguity, raising quality, security, and IP concerns.
Practitioner concerns and examples documented in interviews and observations at Netlight; paper cites security and IP uncertainty as recurring themes; no systematic security scans or legal analyses reported.
There is substitution risk: routine ideation and drafting tasks may be automated, altering task-level labor demand and wage structure.
Task-automation literature and empirical studies of LLMs performing routine drafting/ideation tasks summarized in the review; no long-run labor-market causality established in the paper.
Generative AI lacks reliable situational judgment on ambiguous problems and on ethical trade-offs, making it insufficient for autonomous decision-making in such contexts.
Case examples and experimental studies cited in the synthesis showing inconsistent or inappropriate responses to ambiguous/ethical scenarios; no large-scale causal evidence provided.
LLMs are prone to bias, mediocrity, and factual or logical errors when domain-specific context or experiential knowledge is absent.
Review of empirical evaluations documenting biased outputs, superficial or mediocre suggestions, and factual errors in open-ended tasks and domain-specific prompts; evidence comes from multiple short-term studies and applied examples.
LLMs are predominantly recombinative — they tend to rework and recombine existing material rather than produce deeply novel insights.
Analytical synthesis of output analyses and creativity assessments from multiple empirical studies demonstrating frequent recombination of existing concepts and lower rates of highly original novelty; studies and measures vary.
Proliferation of low-quality or biased AI-generated ideas creates externalities: increased filtering and reputational costs for firms and risks of poor product designs, ethical lapses, or regulatory violations if evaluation is insufficient.
Case studies and qualitative reports documenting filtering burdens and instances of biased/misleading outputs; theoretical reasoning about reputational and regulatory risks; direct quantification of these externalities is limited.
Standard productivity metrics (e.g., TFP) may undercount the value of ideation and creative augmentation provided by generative AI, making attribution between human and AI contributions difficult.
Methodological discussion in the review supported by heterogeneity in outcome measures across studies and challenges in measuring implemented idea quality and long-run impacts.
Generative models exhibit recombination bias: they tend to remix existing patterns rather than produce deeply original, paradigm-shifting insights.
Synthesis of output analyses across studies showing frequent recombination of known patterns and limited evidence of wholly novel, paradigm-changing ideas; claim based on qualitative and comparative analyses in reviewed literature.
AI illiteracy (lack of understanding of AI capabilities/limits) impedes adoption and appropriate use of AI tools in finance.
Survey and interview data reporting lower adoption/intended use among respondents with limited self-reported AI understanding; supplemented by qualitative explanations; sample described as finance professionals across multinational institutions (size unspecified).
Excessive reliance on algorithmic suggestions can erode human judgment and create systemic risks.
Interview reports and, where available, operational/risk metrics indicating overreliance patterns; authors note systemic-risk implications based on combined qualitative and quantitative observations (no causal identification reported).
Cognitive biases and inappropriate trust (both overtrust and distrust) distort decision outcomes and limit the benefits of AI-assisted decision-making.
Qualitative interview evidence describing instances of cognitive bias and misplaced trust; some quantitative indicators of decision distortion and risk where operational performance/risk metrics were available; sample: finance professionals across multinational institutions (detailed metrics not specified).
Market dominance by global platforms can stifle local entrants and distort competition; policies should address market power and data monopolies.
Review of platform economics and competition policy literature; policy argumentation rather than new empirical competition analysis in this paper.
If local data ownership, capacity and governance are weak, economic gains from AI risk accruing to foreign firms and exacerbating income and wealth concentration.
Conceptual synthesis referencing empirical studies on platform rents and data monetization; no original economic distribution analysis presented.
AI and automation can displace labour—particularly routine tasks—heightening the need for retraining, active labour policies and social protection.
Review of literature on automation and labour markets combined with normative inference for African contexts; no primary labour market data presented.
AI adoption raises a risk of digital colonialism: foreign control of data, platforms, and value capture may divert economic gains away from local actors.
Conceptual analysis drawing on policy documents and empirical literature about data flows, platform economics, and international investment; no original quantitative measurement in this paper.
Increased monitoring and algorithmic management raise concerns about worker autonomy and privacy and will prompt regulatory responses (data protection, algorithmic transparency) that shape adoption costs and trajectories.
Recurring concerns reported across included studies and the review's policy implication section; grounded in qualitative and normative discussions within the literature.
Over-standardisation of curricula can create mismatches between certified competencies and firm-specific needs.
Stated in Risks: the paper warns that overly standardized curricula may not fit firm-specific requirements. This is a conceptual caution, not supported by within-paper empirical comparisons.
High fixed costs may concentrate training capacity among a few providers, risking reduced competition.
Listed under Risks to Watch: the paper warns that high fixed costs could concentrate capacity. This is a theoretical market-concentration risk; no empirical market analysis is provided.
Upfront and maintenance costs are substantial; economic evaluation should compare these costs to downstream benefits such as placement rates and productivity gains.
Paper recommends economic evaluation, lists cost-per-curriculum and other cost metrics; presented as advice rather than results. No empirical cost–benefit data provided.
Complexity and lock-in to specific standards may raise barriers to innovation and increase switching costs.
Discussed in Regulation and compliance economics and Risks: claims that standardisation and embedded processes could produce vendor/standard lock-in. This is a theoretical risk flagged by the authors, not supported by empirical data in the paper.
Biased training data or objective functions in AI models could perpetuate gender disparities by offering different products or risk scores to men and women.
Review of AI fairness literature and examples of algorithmic disparate impacts summarized in the paper (conceptual and case evidence; not an empirical test tied specifically to fintech products in the review).
Optional LLM access without training was associated with shorter written answers compared with no LLM access.
Measured answer length in the randomized trial (n = 164); comparison between untrained optional-access arm and no-access arm showed shorter answers in the untrained-access group.
Short-run displacement risks from AI adoption create distributional concerns that warrant active labor market policies (retraining, wage insurance) and portable social protections.
Worker-level evidence of short-run employment losses in routine occupations, particularly in emerging economies, and literature synthesis on displacement effects motivating policy recommendations.
AI-enabled platforms can increase market concentration and platform power, creating competition and data-governance risks and uneven distributional effects across regions and worker skill levels.
Observational platform-concentration indicators and distributional analyses in the case material; scenario and sensitivity checks on distributional outcomes under alternative adoption/policy regimes.
AI substitutes for and displaces many routine and low-skill occupations, increasing automation risk for those roles.
Multiple empirical studies in the reviewed sample document higher automation/substitution risk and observed employment declines in routine/low-skill tasks and occupations.
Young workers experience pronounced negative effects in occupations exposed to AI.
Demographic breakdowns in occupation-level analyses showing larger employment declines (or weaker employment growth) for younger cohorts in AI-exposed occupations.
Diffusion of AI skills is associated with lower employment in occupations that are both highly exposed to AI and have low complementarity with it.
Panel/cross-sectional analyses linking occupation-level AI exposure and measured worker–AI complementarity to employment changes, using occupation classifications of exposure and complementarity.
Middle-skilled occupations are most at risk, contributing to a shrinking middle class (declines in middle-skilled employment).
Occupation-level analyses showing employment declines concentrated in middle-skilled occupations as new skills (IT/AI) diffuse.
AI adoption can reinforce winner‑take‑most market dynamics and increase market concentration due to data‑ and AI‑driven advantages.
Theoretical arguments and industry analyses on platform markets and data economies; empirical market‑structure studies and descriptive evidence cited in the review; the claim is derived from synthesis rather than a single causal identification design.
Impacts of AI on labor are uneven globally: developing regions face larger risks due to digital infrastructure gaps, limited reskilling capacity, and weaker social protections.
Cross‑country comparative analyses, policy and industry reports highlighting infrastructure and institutional differences, and sectoral case studies; review notes geographic bias toward advanced economies in the empirical literature, making some cross‑region inference provisional.
There is widespread displacement of routine and lower‑skilled tasks associated with AI and automation.
Task‑based analyses decomposing occupations into automatable vs augmentable tasks, econometric studies correlating measures of automation/AI exposure with declines in employment and/or hours in routine occupations, and industry reports documenting automation of routine tasks; evidence is largely from cross‑country and country‑specific empirical work summarized in the review.
Prevailing reskilling strategies assume access to stable employment, time and funds for training, certification systems, and institutional support — conditions that are weak or absent for informal platform workers; therefore standard reskilling policies are poorly suited to this context.
Qualitative synthesis of policy analyses and literature on reskilling programs and labour-market institutions; conceptual critique rather than new empirical testing.
Algorithmic management (opaque algorithms for assignment, pricing, and performance metrics) restructures platform work in ways that both change task composition and intensify precarity, reducing workers' ability to adapt to automation.
Draws on prior empirical studies and policy analyses of algorithmic management cited in the literature review; no new empirical data collected in this paper.
Task versus job displacement operate differently across institutional contexts: in formal labour markets, task automation can be accommodated through reallocation or protections, while in informal platform work task loss typically becomes outright job loss.
Argument built from secondary literature comparing formal and informal labour-market institutions and existing empirical studies on reallocation mechanisms; conceptual analysis in the paper (qualitative synthesis only).
AI-driven automation in platform-based informal work in India primarily displaces tasks, but because workers lack job security, institutional protections, and access to alternative labour tracks, task-level automation often manifests as full job displacement.
Synthesis of prior empirical studies, policy analyses, and theoretical work on platform-based labour and automation focused on India and comparable developing-country settings; conceptual framing distinguishing task-level vs job-level effects; no primary data or new empirical analysis in this paper.
Reduced labor shares disproportionately harm lower- and middle-skill workers relative to higher-skill workers, increasing distributional inequality.
Micro and firm-case analyses linking K_T exposure to occupation- and skill-level wage/employment outcomes; regressions showing heterogeneous effects across skill groups; supporting evidence from sectoral studies.
The loss of labor share and payrolls materially undermines PAYG pension sustainability and payroll-tax revenue bases under realistic adoption trajectories.
Dynamic general equilibrium overlapping-generations model calibrated and simulated to incorporate substitution between labor and K_T and a PAYG pension sector; fiscal simulations show declining contributor bases and pressure on pension balances; sensitivity analyses across adoption speeds.
Wages for workers in K_T‑intensive firms/industries fall or grow more slowly relative to less-exposed counterparts, compressing wage contributions to income.
Panel regressions estimating wage outcomes conditional on K_T intensity measures, with controls and robustness specifications; supported by matched employer‑employee microdata in case studies and industry-level decompositions.
AI changes the nature of capital (digital/algorithmic assets) and complicates productivity accounting; researchers should decompose firm-level productivity gains into AI technology, complementary organizational capital, and human capital effects.
Theoretical proposal grounded in productivity accounting literature and conceptual discussion; no single decomposition empirical result presented.
Policy and governance issues become salient: liability, IP, security, and certification of AI-generated code require new standards for provenance, testing, and accountability.
Argument based on practitioner-raised concerns about security, IP, and provenance in the Netlight study; authors recommend policy attention; no legal/regulatory analysis or empirical policy evaluation provided.
Time-series metrics (e.g., derivatives like d/dt(student enrollment)) are useful monitoring signals for validation and system oversight.
Methodological suggestion in the paper proposing time-series analysis of enrollment and other administrative data; no empirical demonstration or threshold criteria provided.
A Sankey diagram of thematic evolution shows lexical convergence over time and indicates that a small set of authors has disproportionate influence in structuring the discourse.
Thematic evolution analysis visualized with a Sankey diagram; author influence inferred from performance trends (citations/publication counts) in the bibliometric data.
Overall employment in Albania has not fallen sharply; instead, changes are concentrated within occupational groups (i.e., occupational restructuring).
Official labor market statistics analyzed descriptively over the recent period, complemented by business survey and case-study evidence of within-occupation shifts. No causal identification; sample details not provided.
AI adoption in Albania is driving occupational restructuring rather than producing large net job losses.
Descriptive analysis of official labor market statistics, business surveys, and selected firm case studies comparing employment levels and occupational composition over the recent period; study notes limited causal identification. Sample size not specified in summary.
The study is the first empirical investigation of human–AI assistance in a live CTF setting with a direct comparison to autonomous AI agents on the same fresh challenges.
Authors' positioning of their work as novel; methodology involved a live onsite CTF, instrumentation of human–AI interactions (41 participants), and direct benchmarking of four autonomous agents on the same fresh challenge set.
Many early-stage AI advances have not translated into higher Phase II/III success rates.
Synthesis of reported outcomes and failures from industry experience; no new systematic statistical analysis provided.
After roughly a decade of adoption in large biopharma, AI has not yet changed late-stage (Phase II/III) clinical success rates.
Qualitative assessment of industrywide experience and reported outcomes; statement based on narrative review rather than systematic, long-run quantitative analysis or causal estimates.
Three primary adoption archetypes in large pharma are (1) partnership-driven acceleration, (2) culture-centric transformation, and (3) production-first democratization.
Conceptual classification in the editorial derived from trends and illustrative examples rather than empirical survey or sampling; no quantitative validation provided.