Evidence (4333 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Governance
Remove filter
Strict upfront compliance may slow deployment but also reduce long‑run liabilities and reputational externalities, affecting venture timelines and expected returns.
Policy trade‑off analysis in the compendium; theoretical and normative argumentation without empirical longitudinal study.
Enforced explainability and non‑discrimination tests may change the design and variable use in credit models, affecting risk assessment, interest spreads and access for historically excluded groups.
Technical and policy analysis synthesizing literature on model design and fairness trade‑offs; normative projections rather than empirical demonstration.
Attribution and measurement innovations affect how value is credited across channels, altering budget allocation across publishers and influencing platform revenues.
Conceptual and policy analysis, supported by literature on attribution effects on budgets; no new empirical allocation dataset presented.
AI-driven bid optimization can increase short-term allocative efficiency (better matching) but may generate welfare-reducing externalities like privacy loss and attention capture.
Auction-market theory and empirical studies cited in literature on bid optimization; the paper synthesizes these findings rather than presenting a new randomized experiment.
Model performance, fairness, robustness, and sustainability are co-produced by technical choices plus contracts, platform policies, and regulation (co-production claim).
Conceptual synthesis combining technical evaluation literature with institutional analysis; no controlled empirical partitioning of effects provided.
Automated market and model optimization create economic efficiencies but reduce transparency for buyers, sellers, and regulators (Efficiency vs opacity trade-off).
Auction and market analysis literature and theoretical arguments; examples from RTB market structure and opaque bid optimization policies discussed; no new controlled experiment provided.
More targeted messaging can improve relevance and conversion but increases risks of nudging and informational harms (Relevance vs manipulation trade-off).
Conceptual trade-off illustrated via causal inference and targeting literature; supported by empirical studies in cited literature (not reproduced here) showing higher conversion with targeting and separate literature on persuasion risks.
The economic performance, social impacts, and durability of AI-driven advertising are determined as much by institutional arrangements (platform design, governance, regulation, market structure) as by model accuracy.
Theoretical and institutional analysis, case-study style arguments and literature references; paper does not present new randomized or large-sample empirical results quantifying the relative contribution.
Federated systems can lower barriers for advertisers and publishers who previously lacked aggregated data, but they also create coordination and infrastructure costs that may favor organizations able to invest in shared infrastructures or consortium governance.
Economic analysis and policy discussion outlining effects on entry, competition, and coordination costs. Evidence is conceptual; no empirical market-entry case studies provided.
Automation reshapes job tasks — reducing demand for some routine manual roles while increasing demand for technical, supervisory, logistics-planning, and service roles — implying substantial reskilling needs rather than outright net job collapse.
Labor-market analysis using occupational employment and job-posting data (task content), supplemented by qualitative interviews and surveys tracing task changes and reskilling needs; scenario sensitivity checks on net employment under alternative adoption paths.
Labor market institutions (unions, collective bargaining), education and training systems, social safety nets, and regulations substantially mediate distributional and aggregate outcomes of AI adoption.
Comparative institutional analysis and equilibrium models linking institutional settings to wage-setting and reallocation dynamics, supported by empirical cross-jurisdiction comparisons where available.
Developing economies face different trade-offs from AI adoption than advanced economies, due to different occupational structures and complementarities.
Comparative analyses and sectoral studies drawing on cross-country microdata and institutional comparisons; theoretical models highlighting differences in task composition and absorptive capacity.
Occupational reallocation occurs: declines in some routine occupations alongside growth in AI-complementary roles (e.g., AI maintenance, oversight, and creative tasks).
Administrative and household employment data analyzed with occupational breakdowns, supplemented by task-mapping methods and panel/event-study approaches documenting shifting occupational shares over time.
Lower-skill roles experience mixed outcomes: some see adverse effects from automation while others benefit where AI is complementary to their tasks.
Microdata analyses and case studies showing heterogeneous effects by task complementarity; task-based exposure measures that differentiate which low-skill tasks are automatable versus augmentable.
AI contributes to wage polarization: earnings grow at the top of the distribution and stagnate or fall for middle occupations.
Wage distribution decompositions and panel regression studies that examine percentile-level wage changes, combined with task-based exposure measures linking AI adoption to differential impacts across the wage distribution.
The employment impact of automation depends crucially on labour-market structure (formal vs informal), availability of alternative employment, and social protections.
Theoretical framing supported by secondary literature comparing institutional contexts and their mediating effects on automation outcomes; no primary causal estimates in this paper.
Standard policy responses focused on retraining and active labor-market programs are necessary but insufficient to fully offset structural job losses where K_T substitutes broadly for tasks.
Model simulations and policy experiments in the calibrated dynamic model comparing scenarios with aggressive retraining versus structural fiscal/interventionist reforms; discussion of empirical limits from case studies and historical reskilling outcomes.
Routine automation of routine drafting tasks by GLAI may reduce demand for junior drafting labor while increasing demand for skilled reviewers, auditors, and legal technologists.
Labor-market reasoning based on task automation literature and illustrative vignettes; no labor-force survey or longitudinal employment data provided.
That measured machine-equivalent work appeared on no financial statement, workforce report, or government statistical return.
Claim about absence of reporting for the deployment's measured work (asserted in the paper for the deployment case).
The emergence and diffusion of these technologies create an era of labor displacement.
Framed in the paper as a premise motivating policy proposals; presented as a conceptual claim rather than supported by original empirical estimates in the text provided.
The economic inevitability of technological transformation (in agentic finance) and the critical urgency of proactive intervention.
Author claim synthesizing the paper's argument and modeling results (normative conclusion based on earlier analysis and assertions, not a validated empirical finding).
Beyond an environment-specific optimum, scaling further degrades institutional fitness because trust erosion and cost penalties outweigh marginal capability gains.
Analytical argument from the Institutional Scaling Law together with illustrative examples and discussion of mechanisms (trust erosion, cost penalties) in the paper.
Bias effects vary by vulnerability type, with injection flaws being more susceptible to framing bias than memory corruption bugs.
Subgroup analysis in Study 1 comparing framing sensitivity across vulnerability classes (injection vs memory corruption) within the experiment dataset.
Model convergence in DRL can lead to crowded trades, which has implications for market stability and motivates a robust regulatory framework balancing innovation with market stability.
Analytical argument in the paper linking convergence/crowding to systemic effects; the excerpt does not include empirical market-impact studies, simulations, or measured incidence rates of crowding.
Deploying DRL at scale requires socio-technical infrastructure considerations including algorithmic governance, systemic risk management, and accounting for the environmental cost of large-scale computational finance.
Conceptual and system-level analysis presented in the paper; no empirical auditing data, carbon-footprint measurements, or governance case studies are provided in the excerpt.
Two sources of spurious performance addressed are memorization bias from ticker-specific pre-training and survivorship bias from flawed backtesting.
Problem identification and methodological focus: the paper names memorization bias and survivorship bias as primary confounders it aims to mitigate. The excerpt does not detail experiments that quantify the magnitude of those biases or the degree to which they were reduced.
Traditional ex ante regulatory approaches struggle to keep pace with AI development, exacerbating the 'pacing problem' and the Collingridge dilemma.
Theoretical/legal literature review and conceptual argument presented in the paper (no empirical sample or quantitative data reported in the abstract).
Low internal conflict or unanimity can be diagnostic of variance depletion (i.e., exclusion) rather than healthy integration, so governance systems should treat low conflict as a potential red flag until heterogeneity integration is verified.
Interpretive policy implication derived from the model's demonstration that exclusionary processes can produce deceptively low observed disagreement while increasing fragility; this recommendation is based on theoretical reasoning without empirical validation in the paper.
Underprovision of verification is likely if left to market forces because information quality has positive externalities and misinformation imposes negative externalities, justifying public funding, subsidies, or regulation.
Economic reasoning and policy implications drawn from the study's findings and the literature on public goods/externalities.
Censorship, restricted data flows, and government interference fragment markets, limit economies of scale, and favor well-resourced, internationally connected actors—widening capacity gaps.
Interpretive economic analysis grounded in observed access constraints and comparative case material across the three platforms.
Limited data access and censorship reduce the efficacy of AI tools by creating training and validation gaps; legal risks complicate use of proprietary platforms and cloud services.
Interviews describing constraints on data availability and legal/operational barriers to using some platforms and cloud services; interpretive analysis of implications for AI training/validation.
Generative AI increases the volume and sophistication of misinformation (deepfakes, fabricated documents), raises false-positive risks, and can be weaponized by state or nonstate actors.
Interview accounts and qualitative analysis noting observed or anticipated misuse of generative models and associated verification challenges.
Resource constraints—limited staff time, funding, and technical capacity—are recurring operational challenges for these platforms.
Staff and stakeholder interviews plus analysis of organizational reports indicating staffing, funding, and technical limitations.
Platforms experience difficulty building and retaining audience trust and engagement, especially in contexts of high public skepticism or polarization.
Interview data from platform staff describing audience engagement challenges, supported by analysis of audience-focused platform formats and community-reporting strategies.
Platforms face limited or asymmetric access to primary data sources such as platform APIs, state data, and archives.
Interview accounts and document analysis noting restricted API access and barriers to state-held data and archives across the three cases.
Censorship and legal risks constrain reporting and distribution for these fact-checking platforms.
Consistent reports from interview subjects and corroborating document analysis indicating legal/censorship-related limitations on publishing and distribution.
Political instability, legal pressure, and censorship strongly shape what platforms can investigate, publish, and access in the region.
Thematic findings from semi-structured interviews with platform staff and document analysis of public reports and policy statements across the three country cases.
Investments in alignment interventions (pluralistic evaluation, transparency) produce public‑good benefits that private firms may underinvest in absent regulation, standards, or procurement incentives.
Economic reasoning about public goods and incentives, supported by conceptual synthesis of firm behavior literature, not by original empirical investment data.
Misalignment generates negative externalities (misinformation, biased decisions, harms to vulnerable groups) that markets may underprovide solutions for, motivating public‑interest interventions.
Economic argumentation and literature synthesis on externalities and public goods; supported by referenced examples in prior work though not quantified here.
AI can augment measurement (e.g., collaboration patterns, output tracking) but if poorly designed may reinforce visibility biases that disadvantage remote workers.
Theoretical reasoning and literature citations about algorithmic bias and monitoring; illustrated with secondary examples rather than primary empirical tests.
Hybrid arrangements can exacerbate inequities in access to informal networks and career advancement, often privileging co-located or better-networked employees.
Theoretical integration of sociological and management studies with comparative case illustrations; secondary data examples referenced but no new causal empirical tests reported.
Hybrid and remote work create risks of professional invisibility, fragmented social networks, and unequal access to workplace social capital.
Literature synthesis and illustrative case studies drawn from secondary sources; qualitative/comparative case evidence rather than primary quantitative data.
Micro and small firms exhibited weak or limited responses to CAFTA spillovers because of financing constraints, lower innovation capacity, and limited international market information.
Firm‑level heterogeneity and subgroup analyses indicating attenuated effects for micro/small firms; authors attribute weaker responses to observed constraints (financing, innovation, information) in the industrial enterprise database.
CAFTA reduced procurement costs for firms importing agricultural goods, lowering marginal procurement costs.
Mediator tests in the paper linking CAFTA to reduced procurement costs using firm‑level cost/price/procurement indicators from the industrial enterprise database and customs data within DID design.
HACCA proliferation increases negative externalities and public-good failure risks, meaning private markets will underinvest in mitigation absent public intervention.
Public-goods and externality economic theory applied to cybersecurity; policy analysis (qualitative).
Widespread HACCA availability compresses the capability gap between resource-rich and resource-poor actors, empowering criminal groups and smaller states and concentrating harms in less-protected sectors and geographies.
Diffusion and strategic externalities analysis; scenario reasoning about capability democratization (qualitative).
Firms will shift investment toward cybersecurity and away from other productive uses; small and medium enterprises (SMEs) will be disproportionately affected due to limited defenses.
Investment-allocation reasoning and distributional analysis of firm capabilities (qualitative; no firm-level panel data).
Cyber insurance markets will face increased premium pressure and uncertainty; insurers may raise prices, restrict coverage, or withdraw from some lines.
Economic analysis of risk pricing under higher uncertainty and tail risks; analogy to prior insurance market reactions to emerging risks (qualitative).
Automation lowers fixed and marginal costs of conducting high-skill cyber operations, changing the supply-side economics and enabling a rapid expansion in the number of attackers.
Cost-structure reasoning about automation effects on labor and tool costs; conceptual economic analysis (no empirical cost data provided).
Widespread diffusion of HACCAs will raise the baseline cyber threat and reduce the monopoly of advanced states and groups on high-end offensive capabilities.
Capability diffusion assessment and historical analogies to proliferation of technologies (qualitative; no large-scale empirical diffusion model).