Evidence (11677 claims)
Adoption
7395 claims
Productivity
6507 claims
Governance
5921 claims
Human-AI Collaboration
5192 claims
Org Design
3497 claims
Innovation
3492 claims
Labor Markets
3231 claims
Skills & Training
2608 claims
Inequality
1842 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 609 | 159 | 77 | 738 | 1617 |
| Governance & Regulation | 671 | 334 | 160 | 99 | 1285 |
| Organizational Efficiency | 626 | 147 | 105 | 70 | 955 |
| Technology Adoption Rate | 502 | 176 | 98 | 78 | 861 |
| Research Productivity | 349 | 109 | 48 | 322 | 838 |
| Output Quality | 391 | 121 | 45 | 40 | 597 |
| Firm Productivity | 385 | 46 | 85 | 17 | 539 |
| Decision Quality | 277 | 145 | 63 | 34 | 526 |
| AI Safety & Ethics | 189 | 244 | 59 | 30 | 526 |
| Market Structure | 152 | 154 | 109 | 20 | 440 |
| Task Allocation | 158 | 50 | 56 | 26 | 295 |
| Innovation Output | 178 | 23 | 38 | 17 | 257 |
| Skill Acquisition | 137 | 52 | 50 | 13 | 252 |
| Fiscal & Macroeconomic | 120 | 64 | 38 | 23 | 252 |
| Employment Level | 93 | 46 | 96 | 12 | 249 |
| Firm Revenue | 130 | 43 | 26 | 3 | 202 |
| Consumer Welfare | 99 | 51 | 40 | 11 | 201 |
| Inequality Measures | 36 | 106 | 40 | 6 | 188 |
| Task Completion Time | 134 | 18 | 6 | 5 | 163 |
| Worker Satisfaction | 79 | 54 | 16 | 11 | 160 |
| Error Rate | 64 | 79 | 8 | 1 | 152 |
| Regulatory Compliance | 69 | 66 | 14 | 3 | 152 |
| Training Effectiveness | 82 | 16 | 13 | 18 | 131 |
| Wages & Compensation | 70 | 25 | 22 | 6 | 123 |
| Team Performance | 74 | 16 | 21 | 9 | 121 |
| Automation Exposure | 41 | 48 | 19 | 9 | 120 |
| Job Displacement | 11 | 71 | 16 | 1 | 99 |
| Developer Productivity | 71 | 14 | 9 | 3 | 98 |
| Hiring & Recruitment | 49 | 7 | 8 | 3 | 67 |
| Social Protection | 26 | 14 | 8 | 2 | 50 |
| Creative Output | 26 | 14 | 6 | 2 | 49 |
| Skill Obsolescence | 5 | 37 | 5 | 1 | 48 |
| Labor Share of Income | 12 | 13 | 12 | — | 37 |
| Worker Turnover | 11 | 12 | — | 3 | 26 |
| Industry | — | — | — | 1 | 1 |
The regulatory architecture is in place; the verification instrument is not.
Paper's high-level diagnosis asserting that regulations establish obligations but lack a technical instrument for quantitative verification of acceptable risk.
The systems most in need of oversight are opaque statistical inference engines that resist white-box scrutiny.
Paper's characterization/analysis of contemporary high-risk AI systems as opaque statistical models that are difficult to inspect via white-box methods.
This gap is not theoretical: as the EU AI Act moves into full enforcement, developers face mandatory conformity assessments without established methodologies for producing quantitative safety evidence.
Argument in paper linking imminent enforcement of EU AI Act to practical conformity-assessment requirements for developers and asserting lack of established methodologies for quantitative safety evidence.
None [of these regulatory frameworks] specifies what 'acceptable risk' means in quantitative terms, and none provides a technical method for verifying that a deployed system actually meets such a threshold.
Paper's critical analysis of existing regulatory instruments, arguing absence of quantitative definitions and verification methods.
Ethical concerns—such as transparency, explainability, psychological effects, and responsible AI governance—are critical factors influencing employability outcomes.
Review synthesis highlighting ethical issues from empirical and industry literature as influential on employability outcomes.
There are significant AI adoption challenges in education and industry that affect employability and role transformation.
Synthesized evidence from industry reports and empirical studies discussed in the review highlighting barriers to adoption in education and industry.
Technological interdependence is not dissolving but being selectively restructured, producing a durable condition of partial, segmented decoupling in which interdependence persists under increasingly politicized rules of access.
Interpretation based on case-study observations of export controls, allied coordination, Chinese countermeasures, and emergent supply-chain and regulatory changes described in the paper.
When the United States employs export controls and allied coordination to manage perceived technological risks, China responds through defensive reconfiguration aimed at reducing asymmetric vulnerability, in addition to retaliation in rare-earth export controls in certain instances.
Case-study evidence centered on advanced-technology sectors (particularly semiconductors) and observed policy responses following U.S. export restraints after the first Trump administration (qualitative policy and reaction examples described in the paper).
From the perspectives of 'personal subordination' and 'economic subordination', AIGC deeply and implicitly controls the labor process through mechanisms such as dynamic path planning, blurring the boundaries of determination.
Analytical/legal argument in the paper linking conceptual standards of subordination to specific algorithmic mechanisms (e.g., dynamic path planning); supported by mechanistic discussion but no reported empirical measurement or sample.
AIGC constantly challenges traditional standards for determining labor relations.
Paper's analytic claim based on conceptual/legal argument that algorithmic features of AIGC complicate application of existing labor-relation tests; no quantitative validation or sample size provided.
The transformation toward algorithmic enterprises raises critical concerns regarding agency, accountability, data monopolization, and algorithmic bias.
Presented as a principal concern in the paper's conceptual discussion and interdisciplinary critique; based on analysis of governance and ethical literature rather than new empirical evidence in the abstract.
Algorithmic management and monitoring have reduced employees’ autonomy and perceived work meaningfulness, contributing to 'AI anxiety' characterised by concerns about job loss, skill obsolescence, and diminished control.
Qualitative studies, survey evidence, and theoretical literature reviewed that document impacts of algorithmic management on autonomy, meaningfulness, and worker anxiety (mixed-methods literature).
Automation has intensified income inequality between high-skilled and low-skilled workers.
Synthesis of empirical literature linking automation adoption to widening wage and income gaps across skill groups (literature review).
Displacement effects have extended from manufacturing into cognitive roles such as clerical work and customer service.
Review of empirical studies documenting automation/substitution effects in cognitive, clerical, and customer-service roles (literature synthesis).
Automation has put downward pressure on wages.
Cited empirical studies and wage analyses in the reviewed literature indicating wage suppression associated with automation adoption (literature review).
AI and robotics have led to contractions in low-skilled occupations.
Synthesis of empirical literature reporting occupational contractions in low-skilled jobs following automation adoption (literature review).
Extensive empirical evidence shows that AI and robotics can substitute for rule-based, codifiable routine tasks.
Review cites extensive empirical studies demonstrating substitution of rule-based, codifiable routine tasks by AI/robotics (literature synthesis).
Artificial intelligence and robotic technologies are fundamentally reshaping labour markets and pose multifaceted challenges to workers engaged in routine and low-skilled tasks.
Narrative review of domestic and international scholarly literature over the past decade (literature review / synthesis).
Structural barriers, workforce biases, and digital skill gaps affect women’s participation in AI-enabled sectors.
Claim derived from the paper's synthesis of literature (peer-reviewed studies, policy analyses, preprints) identifying common barriers; the abstract does not report quantitative meta-analysis or specific sample sizes.
Routine-intensive sectors exhibit higher susceptibility to automation.
Synthesis result reported in the paper based on the systematic review of sector-specific literature (no numeric aggregation or sample size provided in the abstract).
Vibe coding (unstructured GenAI-driven coding) promises rapid prototyping but often suffers from architectural drift, limited traceability, and reduced maintainability.
Paper asserts this as a motivating observation and characterizes vibe coding's weaknesses; the abstract frames these as commonly observed problems motivating the Shift-Up approach (no sample size given in abstract).
In post-AGI economies the presupposition of agent autonomy becomes nontrivial because artificial systems may exhibit varying degrees of autonomy, functioning as tools, delegates, strategic market actors, manipulators of choice environments, or possible welfare subjects.
Theoretical argumentation and conceptual classification in the paper; no empirical data reported (modeling/motivating discussion).
Market incompleteness distorts the efficient development of AI (i.e., distorts innovation/output).
Claim made in the abstract as a theoretical implication of the asset-pricing model; no empirical data provided.
Market incompleteness distorts valuations.
Stated in the abstract as an implication of the model (theoretical analysis); no empirical quantification provided.
Every additional mechanism we test (planner evolution, per-tool selection, cold-start initialization, skill extraction, and three credit assignment methods) degrades performance.
Findings from the nine-variant ablation study reported in the paper; comparison of variants that add each listed mechanism versus the memory+reflection combination.
There is a stark geopolitical divide between 'AI Core' nations and the Global South; the Global South faces acute risks of 'Digital Dependency' and eroded digital sovereignty.
Cross-study synthesis in the systematic review (2018-2026) identifying geopolitical patterns and risks; abstract does not quantify the number of studies or present empirical effect sizes.
The 'black box' nature of automated systems undermines the democratic social contract and principles of procedural justice, epitomised by the Australian 'Robo-debt' scandal.
Case study material and literature synthesized in the systematic review referencing the Australian Robo-debt case as an exemplar; abstract does not provide primary data or sample sizes.
Agentic AI introduces novel challenges related to market stability, regulatory compliance, interpretability, and systemic risk.
Survey discussion synthesizing literature on systemic and governance risks of autonomous systems in markets; draws on conceptual and empirical prior work but does not present new quantitative results.
Consolidation of corporate control of critical technologies (driven by AI industrial strategies that do not center democratic economic governance) threatens key democratic and societal objectives.
Stated implication in the paper's opening argument; supported by the paper's conceptual framing and (as indicated) review of how past and emerging tech/AI industrial strategies interact with democratic objectives. No quantitative sample size provided in the excerpt.
Unless governments develop industrial policy strategies centered on strengthening democratic economic governance, they risk consolidating corporate control of critical technologies.
Main argumentative claim of the paper as stated in the abstract/introduction; presented as a normative risk argument supported in the paper by conceptual analysis and review of policy trends and historical examples (no empirical sample size reported in the excerpt).
Scalable AI tutoring for procedural skill learning requires structured knowledge representations, yet constructing these representations remains a labor-intensive bottleneck.
Background/claim made in the paper's introduction framing the problem; no specific quantitative evidence reported in the abstract.
Under-represented groups tend to be systematically under-observed because of historical exclusion and selective feedback, which exacerbates uncertainty for those groups.
Conceptual claim supported by illustrative examples (e.g., lending context) and simulations demonstrating selective feedback effects; literature citation likely included in paper.
Policies that ignore the unobserved (counterfactual) space can harm decision makers (via unrealized gains or losses) and subjects (via compounding exclusion and reduced access).
Theoretical argumentation and illustrative examples (e.g., loan denial counterfactuals) and modelled simulations showing downstream harms when ignoring unobserved outcomes.
Experiments on simulated data with varying bias show that unequal uncertainty and selective feedback produce disparities across groups.
Simulation experiments described in the paper manipulate bias and feedback patterns and report resulting group disparities (synthetic datasets; experiment details in methods/results sections).
Industrial firms face a dual challenge: (1) the development and deployment of digital technologies and (2) the proliferation and integration of the corresponding skills portfolios.
Conceptual framing and literature synthesis presented in the paper (identification by authors); not tied to a specific quantitative sample in the provided text.
Renewable energy adoption further reinforces the beneficial effect of digital trade on emissions under stronger regulatory stringency (mediation via renewable energy and regulation).
Structural equation modelling (SEM) on the monthly panel (38 OECD economies, 2000–2024) assessing mediation paths through renewable energy adoption and regulatory stringency; reported as reinforcing the digital trade effect.
There is a carbon-pricing threshold at USD 40 per tonne, above which emissions decline significantly (Δ = −15%, p < 0.01).
Carbon-pricing threshold analysis applied to the monthly panel of 38 OECD economies (2000–2024); threshold identified and associated pre/post comparison reports a 15% decline with p < 0.01.
The environmental effect of digital trade becomes stronger (more negative on emissions) when combined with AI-enhanced logistics (interaction effect).
Econometric models including interaction terms for AI-enhanced logistics and digital trade on the monthly panel (38 OECD economies, 2000–2024); interaction effects identified via regression and machine-learning threshold techniques.
GVC participation is significantly associated with lower CO2 emissions (β = −0.064, p < 0.01).
Econometric analysis on a monthly panel of 38 OECD economies from 2000–2024 using fixed-effects models; coefficient and p-value reported in paper.
Traditional forecasting and optimization approaches often operate in isolation, limiting their real-world effectiveness in volatile-demand, uncertain-supply industries.
Positioning/background statement in the paper motivating the integrated framework (literature-based claim).
The study is framed based on Job Demands-Resources (JD-R) theory, positing that HAI-C task complexity is a job demand and AI self-efficacy/humble leadership act as resources that can mitigate negative effects on engagement.
Introduction states JD-R theory as the theoretical basis and describes job demands (HAI-C task complexity) and job/personal resources (humble leadership, AI self-efficacy) in the hypothesized model.
HAI-C tech-learning anxiety reduces employees' work engagement (serves as the mediator between HAI-C task complexity and work engagement).
Mediation analysis via hierarchical regression and bootstrapping on the three-wave survey sample of 497 employees; reported in Results as the mediating mechanism.
Human-AI collaboration task complexity (HAI-C task complexity) negatively affects employees' work engagement by amplifying their HAI-C tech-learning anxiety.
Three-wave longitudinal survey of matched data from 497 employees; mediation analysis using hierarchical regression and bootstrapping reported in the Results section.
LLMs are not only less accurate on ideologically contested economic questions, but systematically less reliable in one ideological direction than the other, underscoring the need for direction-aware evaluation in high-stakes economic and policy settings.
Synthesis of empirical findings: lower accuracy on contested items, higher accuracy for intervention-aligned cases in 18/20 models, and error skew toward intervention-oriented predictions; policy recommendation follows from these empirical patterns.
This directional skew is not eliminated by one-shot in-context prompting.
Intervention of one-shot in-context prompting applied to models; evaluation shows the intervention-oriented error skew persists despite one-shot prompting.
Ideology-contested items are consistently harder than non-contested ones.
Comparison of model performance (accuracy) on contested subset (1,056 items) versus non-contested items in the 10,490-triplet benchmark; reported consistent lower accuracy on contested items.
Important boundary conditions include data maturity, process integration, governance discipline, and the degree of functional trust between finance and operating units.
List of boundary conditions reported in the paper based on documentary case analysis and synthesis with literature.
GenAI does not improve management accounting decision quality primarily by replacing managerial judgment.
Interpretive finding based on documentary analysis of disclosures from the three case firms and relevant literature; presented as a summary conclusion in the paper.
The stakes are particularly high in spreadsheet environments, where process and artifact are inseparable: each decision the agent makes is recorded directly in cells that belong to and reflect on the user.
Conceptual / domain-specific argument made by the authors (no empirical sample attached to the claim).
AI agents can perform sophisticated, multi-step knowledge work autonomously from start to finish, yet this process remains effectively inaccessible during execution: by the time users receive the output, all underlying decisions have already been made without their involvement.
Author assertion / conceptual description in the paper (no empirical quantification provided for this general statement).