Evidence (4333 claims)
Adoption
5539 claims
Productivity
4793 claims
Governance
4333 claims
Human-AI Collaboration
3326 claims
Labor Markets
2657 claims
Innovation
2510 claims
Org Design
2469 claims
Skills & Training
2017 claims
Inequality
1378 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 402 | 112 | 67 | 480 | 1076 |
| Governance & Regulation | 402 | 192 | 122 | 62 | 790 |
| Research Productivity | 249 | 98 | 34 | 311 | 697 |
| Organizational Efficiency | 395 | 95 | 70 | 40 | 603 |
| Technology Adoption Rate | 321 | 126 | 73 | 39 | 564 |
| Firm Productivity | 306 | 39 | 70 | 12 | 432 |
| Output Quality | 256 | 66 | 25 | 28 | 375 |
| AI Safety & Ethics | 116 | 177 | 44 | 24 | 363 |
| Market Structure | 107 | 128 | 85 | 14 | 339 |
| Decision Quality | 177 | 76 | 38 | 20 | 315 |
| Fiscal & Macroeconomic | 89 | 58 | 33 | 22 | 209 |
| Employment Level | 77 | 34 | 80 | 9 | 202 |
| Skill Acquisition | 92 | 33 | 40 | 9 | 174 |
| Innovation Output | 120 | 12 | 23 | 12 | 168 |
| Firm Revenue | 98 | 34 | 22 | — | 154 |
| Consumer Welfare | 73 | 31 | 37 | 7 | 148 |
| Task Allocation | 84 | 16 | 33 | 7 | 140 |
| Inequality Measures | 25 | 77 | 32 | 5 | 139 |
| Regulatory Compliance | 54 | 63 | 13 | 3 | 133 |
| Error Rate | 44 | 51 | 6 | — | 101 |
| Task Completion Time | 88 | 5 | 4 | 3 | 100 |
| Training Effectiveness | 58 | 12 | 12 | 16 | 99 |
| Worker Satisfaction | 47 | 32 | 11 | 7 | 97 |
| Wages & Compensation | 53 | 15 | 20 | 5 | 93 |
| Team Performance | 47 | 12 | 15 | 7 | 82 |
| Automation Exposure | 24 | 22 | 9 | 6 | 62 |
| Job Displacement | 6 | 38 | 13 | — | 57 |
| Hiring & Recruitment | 41 | 4 | 6 | 3 | 54 |
| Developer Productivity | 34 | 4 | 3 | 1 | 42 |
| Social Protection | 22 | 10 | 6 | 2 | 40 |
| Creative Output | 16 | 7 | 5 | 1 | 29 |
| Labor Share of Income | 12 | 5 | 9 | — | 26 |
| Skill Obsolescence | 3 | 20 | 2 | — | 25 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
Governance
Remove filter
Temporal mapping and citation networks reveal distinct technology maturity patterns, which are visualised using S-curve and hype cycle models.
Paper describes use of temporal mapping and citation network analysis and visualization via S-curve and hype cycle models; methodological description without quantitative sample-size details.
Technologies such as AI-driven healthcare, quantum communication, hydrogen energy, and smart educational AI are identified as key domains of convergence.
Paper reports these domains were identified via the applied analytic framework and multi-source data triangulation; no numeric counts/sample sizes provided.
The study applies advanced techniques such as LDA topic modelling, BERT-based clustering, and co-citation analysis to detect innovation trajectories.
Paper states these specific analytic techniques were applied (method description).
The research leverages large AI models and multi-source data—including global patent databases (WIPO, USPTO, Lens.org), scientific literature corpora, and industry intelligence platforms (CB Insights, Qichacha).
Paper statement of data sources and use of large AI models; methodological description (no sample sizes reported).
A stylized-facts analysis using OECD and World Bank indicators shows that economies with higher digital capacity, greater R&D intensity, and stronger institutions exhibit superior productivity and growth performance.
Stylized-facts (cross-country) analysis based on OECD and World Bank indicators; descriptive correlations reported in the paper (sample of countries not enumerated in the provided summary).
AI adoption stimulates institutional innovation, which in turn increases total factor productivity (TFP) and supports sustainable economic growth.
Theoretical mediation claim developed in the paper (integration of Schumpeterian growth theory with institutional economics); supported conceptually and argued with stylized-facts analysis but not presented as causally identified empirical estimates.
AI improves governance quality.
Argument within the conceptual framework linking AI capabilities (information processing, monitoring) to improved governance; stated qualitatively in the paper rather than supported by causal empirical tests.
AI lowers transaction costs.
Paper's conceptual/theoretical framework that characterizes AI as lowering transaction costs through improved information and coordination; no quantitative causal estimate reported.
AI reduces information asymmetries.
Theoretical/conceptual argument in the paper framing AI as a general-purpose technology that improves information flows; supported by the paper's conceptual framework (no experimental or causal identification reported).
These systems are now being widely used to produce software, conduct business activities, and automate everyday personal tasks.
Authors' statement describing observed applications and uses (policy/legal analysis; specific empirical data or sample size not provided in excerpt).
AI agents have entered the mainstream.
Authors' declarative statement based on their review of recent developments and observed uptake (policy/legal analysis in the paper). No empirical sample size reported in excerpt.
Economies and organizations that prioritize adaptability, workforce transformation, and real-time decision-making capabilities are better positioned to sustain growth under volatile conditions.
Claim based on the paper's cross-cutting analysis of global indicators and the conceptual AEPM framework; the excerpt does not provide a quantified causal estimate, experimental evidence, or sample size supporting this assertion.
AEPM is structured around five core pillars—energy resilience, supply chain flexibility, human capital adaptability, financial sustainability, and AI-enabled decision systems—which together provide a comprehensive approach to managing uncertainty and enabling dynamic responses to structural disruptions.
Conceptual design of the AEPM presented in the paper; described as a multidimensional framework combining these five pillars. No empirical validation or quantified impact measures reported in the excerpt.
The paper proposes shifting from forecasting-centric economic management to an adaptive preparedness paradigm and introduces the Adaptive Economic Preparedness Model (AEPM), a multi-dimensional framework designed to enhance resilience at both organizational and national levels.
Presentation of a conceptual model (AEPM) in the paper structured around five pillars; this is a proposed framework rather than an empirically validated intervention (no evaluation sample or randomized test reported in the excerpt).
The contribution is a falsifiable architectural thesis, a clear threat model, and a set of experimentally testable hypotheses for future work on distillation resistance, alignment, and model governance.
Theoretical contribution claim: the paper proposes hypotheses and a threat model intended to be testable in future empirical work; no experiments in the paper itself are reported.
The authors call for shifting evaluation and assurance from tool qualification toward workflow qualification to achieve trustworthy Physical AI.
Normative recommendation based on the paper's theoretical analysis (policy/recommendation; no empirical sample reported).
The paper derives non-degradation conditions that characterize shadow-resistant workflows for AI-assisted safety analysis.
Analytic derivations and formal criteria presented in the paper (theoretical result; no empirical validation/sample size reported).
The paper formalizes four canonical human–AI collaboration structures and derives closed-form performance bounds for them.
Theoretical/mathematical derivations and models in the paper (no empirical verification/sample size reported).
A five-dimensional competence framework captures safety competence via domain knowledge, standards expertise, operational experience, contextual understanding, and judgment.
Theoretical contribution: paper defines and formalizes a five-dimension framework (no empirical validation/sample size reported).
To facilitate adoption of our evaluation framework, we detail our testing protocols and make relevant materials publicly available.
Statement in paper that testing protocols and materials are documented and released publicly (paper claims to provide materials).
We assess an AI model with 10,101 participants spanning interactions in three AI use domains (public policy, finance, and health) and three locales (US, UK, and India).
Reported sample size and study design details stated in abstract: N = 10,101; three domains and three locales specified.
This paper introduces a framework for evaluating harmful AI manipulation via context-specific human-AI interaction studies.
Paper describes a proposed evaluation framework (methodological contribution); claimed in abstract/introduction as new contribution. No numeric sample required for the claim itself.
The result is evidence-based triggers that replace calendar schedules and make governance auditable.
Claimed outcome of applying the decision-theoretic framework in the paper (argumentative; no empirical deployment or case-study evidence reported in the summary).
The paper provides a decision-theoretic framework for retraining policies.
Explicit claim about the paper's contribution; the article presents a framework (conceptual/methodological exposition).
The retraining decision is a cost minimization problem with a threshold that falls out of your loss function.
Decision-theoretic derivation presented in the paper (analytical/theoretical reasoning; no empirical validation reported).
Retraining can be better understood as approximate Bayesian inference under computational constraints.
Theoretical argument and decision-theoretic framing presented in the paper (conceptual/mathematical derivation rather than empirical testing).
The framework is designed for direct application to engineering processes for which operational event logs are available.
Statement of intended applicability in the paper and demonstration on a large enterprise procurement workflow (BPI 2019 log).
The same quantities that delimit statistically credible autonomy (blind masses, escalation gate, m(s), etc.) also determine expected oversight burden (the framework includes an expected oversight-cost identity over the workflow visitation measure).
Theoretical identity and discussion in the paper plus demonstration on the empirical workflow showing how the introduced quantities relate to expected oversight costs.
On the held-out split, m(s) = max_a \hat{\pi}(a|s) tracks realized autonomous step accuracy within 3.4 percentage points on average.
Empirical evaluation on the paper's held-out test split (chronological 20%); reported average discrepancy between the maximum predicted action probability and realized autonomous-step accuracy.
Refining the operational state to include case context, economic magnitude, and actor class expands the state space from 42 to 668.
Empirical report in the paper showing state-space expansion when additional contextual variables are included in state definition (numbers 42 and 668 stated).
We instantiate the framework on the Business Process Intelligence Challenge 2019 purchase-to-pay log (251,734 cases, 1,595,923 events, 42 distinct workflow actions) and construct a log-driven simulated agent from a chronological 80/20 split of the same process.
Empirical instantiation described in the paper using the BPI 2019 purchase-to-pay event log; dataset statistics (cases, events, distinct actions) and an 80/20 chronological train/test split are reported.
We develop a measure-theoretic Markov framework for agentic AI in organizations, whose core quantities are state blind-spot mass B_n(\tau), state-action blind mass B^{SA}_{\pi,n}(\tau), an entropy-based human-in-the-loop escalation gate, and an expected oversight-cost identity over the workflow visitation measure.
Theoretical development presented in the paper (definition and derivation of the measure-theoretic Markov framework and associated quantities).
The framework aims to support more comparable benchmarks and cumulative research on human-AI readiness, advancing safer and more accountable human-AI collaboration.
Stated aims and intended impact in paper; aspirational/conceptual rather than empirically demonstrated in excerpt.
Operationalizing evaluation through interaction traces rather than model properties or self-reported trust enables deployment-relevant assessment of calibration, error recovery, and governance.
Methodological claim/proposed approach in paper; presented as enabling assessment but no empirical evaluation reported in excerpt.
The taxonomy and metrics are connected to the Understand-Control-Improve (U-C-I) lifecycle of human-AI onboarding and collaboration.
Conceptual mapping described in paper; no empirical tests or sample reported in excerpt.
We introduce a four part taxonomy of evaluation metrics spanning outcomes, reliance behavior, safety signals, and learning over time.
Explicit methodological claim in paper announcing a taxonomy; described as a contribution rather than empirically tested in excerpt.
This paper proposes a measurement framework for evaluating human-AI decision-making centered on team readiness.
Methodological contribution presented in paper; conceptual framework proposed (no empirical validation reported in excerpt).
Artificial intelligence (AI) systems are deployed as collaborators in human decision-making.
Statement in paper (conceptual/observational claim); no empirical sample or method provided in excerpt.
Late disclosure of AI involvement improved affective engagement for AI-enhanced content.
Reported experimental result in the abstract from the two online studies (study 1: n = 325; study 2: n = 371) manipulating disclosure timing (early vs. late).
The results of this regional research outline a multi-dimensional policy roadmap that dives deep into the region’s current capabilities and the hurdles it faces in catching up with the AI revolution from a governance and policy perspective, presenting them in a practical framework for public sector leaders.
Report summary claiming that the study's results produce a comprehensive roadmap and practical framework (content description).
This executive report provides a roadmap for establishing an AI governance infrastructure through a set of strategic policy recommendations across seven key pillars.
Document assertion describing the content and structure of the report (authors' deliverable).
The reality of limited AI governance capacity calls for a series of policy interventions at both local and regional levels to empower the AI ecosystem in the Arab region.
Authors' policy recommendation derived from the regional study and synthesis of findings.
A governance model linking 'trustworthy AI' practices to competitive advantage yields reduced uncertainty, faster deployment cycles, and higher stakeholder trust.
Central claim of the paper tying the proposed AIGSF to business benefits; supported by conceptual linkage and illustrative examples rather than quantified empirical evidence or controlled evaluation.
Case illustrations across hiring, credit, consumer services, and generative AI draw lessons on controls such as model documentation, algorithmic audits, impact assessments, and human-in-the-loop oversight.
Paper includes qualitative case illustrations in the listed domains to demonstrate governance controls; these are presented as examples and lessons rather than as systematic empirical studies (no sample sizes reported).
The paper develops an AI Governance Strategic Framework (AIGSF) and an implementation roadmap that connect ethical accountability, regulatory readiness, cybersecurity resilience, and performance outcomes.
Paper contribution described as an integrative conceptual framework and roadmap; supported by theoretical grounding and illustrative cases rather than empirical validation; no sample size provided.
AI governance should be treated as a strategic governance function—anchored in board oversight and enterprise risk management—rather than a narrow technical or compliance task.
Central normative recommendation and thesis of the paper; derived from an integrative conceptual framework grounded in corporate governance theory, ERM, and emerging regulation. No empirical testing or sample reported.
AI has moved from a peripheral digital capability to a central driver of corporate strategy, reshaping decision-making, customer engagement, operations, and risk exposure.
Statement presented in the paper's introduction and motivation; supported by integrative conceptual design and literature grounding (theory and descriptive citations). No empirical sample or quantitative analysis reported.
A policy of 20% mandatory practice preserves 92% more capability than the simulation baseline (baseline includes a 5% background AI-failure rate).
Simulation comparing baseline (5% background AI-failure rate) to a counterfactual with 20% mandatory practice; reported 92% relative preservation of capability.
The model predicts that periodic AI failures improve human capability 2.7-fold (relative improvement reported in simulations).
Simulation experiments comparing scenarios with/without periodic AI failures; reported fold-change in capability of 2.7×.
Validated against 15 countries' PISA data (102 points), the model achieves R^2 = 0.946 with 3 parameters and attains the lowest BIC among compared specifications.
Empirical validation using PISA dataset covering 15 countries and 102 data points; reported fit statistics (R^2, number of parameters, BIC).