Evidence (4137 claims)
Adoption
5267 claims
Productivity
4560 claims
Governance
4137 claims
Human-AI Collaboration
3103 claims
Labor Markets
2506 claims
Innovation
2354 claims
Org Design
2340 claims
Skills & Training
1945 claims
Inequality
1322 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 378 | 106 | 59 | 455 | 1007 |
| Governance & Regulation | 379 | 176 | 116 | 58 | 739 |
| Research Productivity | 240 | 96 | 34 | 294 | 668 |
| Organizational Efficiency | 370 | 82 | 63 | 35 | 553 |
| Technology Adoption Rate | 296 | 118 | 66 | 29 | 513 |
| Firm Productivity | 277 | 34 | 68 | 10 | 394 |
| AI Safety & Ethics | 117 | 177 | 44 | 24 | 364 |
| Output Quality | 244 | 61 | 23 | 26 | 354 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 168 | 74 | 37 | 19 | 301 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 89 | 32 | 39 | 9 | 169 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 106 | 12 | 21 | 11 | 151 |
| Consumer Welfare | 70 | 30 | 37 | 7 | 144 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 75 | 11 | 29 | 6 | 121 |
| Training Effectiveness | 55 | 12 | 12 | 16 | 96 |
| Error Rate | 42 | 48 | 6 | — | 96 |
| Worker Satisfaction | 45 | 32 | 11 | 6 | 94 |
| Task Completion Time | 78 | 5 | 4 | 2 | 89 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 17 | 9 | 5 | 50 |
| Job Displacement | 5 | 31 | 12 | — | 48 |
| Social Protection | 21 | 10 | 6 | 2 | 39 |
| Developer Productivity | 29 | 3 | 3 | 1 | 36 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Skill Obsolescence | 3 | 19 | 2 | — | 24 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Labor Share of Income | 10 | 4 | 9 | — | 23 |
Governance
Remove filter
Adopting equity-by-design (including diverse, non‑European datasets and subgroup evaluation) reduces model bias and improves global generalizability of AI models.
Recommendations and examples in the review; draws on literature documenting subgroup performance differences and bias remediation strategies (narrative evidence).
AI-enabled trial innovations—such as integration with new approach methodologies (NAMs), adaptive and covariate-adjusted designs, and digital biomarkers—can reduce trial inefficiency while preserving scientific and ethical standards.
Narrative review of trial design optimization methods, examples of adaptive and covariate-adjusted analyses, and digital endpoint qualification discussions; case examples and methodological papers referenced without meta-analysis.
Synthesis-aware and physics-informed molecular design increases the downstream feasibility (synthetic accessibility and developability) of AI-designed compounds.
Methodological literature and case examples of synthesis-aware generative models and physics-informed approaches summarized in the narrative review (heterogeneous studies, no pooled estimate).
External validation, explicit applicability-domain reporting, and subgroup performance reporting improve model reliability and support regulatory alignment.
Technical best-practice recommendations and analysis of evolving regulatory frameworks discussed in the review; examples of regulatory guidance and credibility-plan concepts (narrative).
Structural prediction tools and structural-biology advances speed target validation and can accelerate target identification/validation workflows.
Discussion of structural biology datasets (cryo-EM/X-ray and predicted structures) and use cases in the narrative review; examples include use of predicted structures to inform target characterization (heterogeneous examples).
AI-assisted molecular design can improve lead/compound quality (e.g., potency, selectivity, developability) when using synthesis-aware and physics-informed approaches.
Review of method papers and case examples of synthesis-aware generative models and physics-informed neural networks in de novo design; examples drawn from cheminformatics and molecular design studies (heterogeneous, narrative).
AI can raise early-phase (e.g., Phase I/II) success rates when effectively applied with the technical and governance controls described.
Case studies and literature examples summarized in the narrative review reporting improved early-phase outcomes under AI-supported discovery programs; heterogeneous sample sizes and contexts, no aggregated effect estimate.
Artificial intelligence (AI) can materially shorten drug development timelines when models are predictive, interpretable, and integrated with causal/mechanistic priors, synthesis- and physics-aware molecular design, rigorous external validation (with defined applicability domains), and governance aligned to regulatory requirements.
Narrative synthesis and case examples from recent literature reviewed in the paper; heterogeneous studies and case reports across discovery and early development domains (no pooled/meta-analytic effect size provided).
Labor complementarities with agentic AI will shift resources toward oversight, interpretation, and coordination roles rather than routine task execution.
Economic and organizational reasoning; literature synthesis on skill complementarities; no empirical labor-market data analyzed in the paper.
Principal–agent contracting frameworks must be extended to account for evolving agent objectives and open-ended action spaces; contracts should be dynamic and include continuous renegotiation and monitoring.
Theoretical extension and recommendations based on economic reasoning; proposed formal models for future work.
Projection congruence — alignment of forecasts/plans across heterogeneous agents — becomes a central metric for assessing alignment in agentic human–AI teams.
Conceptual modeling and proposal in the paper; introduced as a new measurable construct (projection congruence indices) for future empirical work.
The DAR framework reframes human oversight as a dynamic, auditable process whose micro-level mechanics and macro-level legitimacy have direct economic consequences for productivity, contracting, regulation, and welfare.
Synthesis claim based on the conceptual framework, formal modeling, derived propositions, and policy/economics implications sections. The claim is theoretical and synthesizing rather than empirically validated.
The Reversal Register will create granular, time-stamped administrative data valuable for structural estimation of trust, error externalities, and productivity comparisons between automation and human judgment.
Design claim linking register contents to potential econometric uses; no empirical data shown—claim about potential data utility.
Reversal Register logs can enable descriptive and causal analyses of handovers and support experimental/quasi-experimental tests (e.g., randomized hysteresis thresholds, A/B override policies).
Implied empirical strategies and instrumentation described; paper outlines how register data would be used for experiments and causal inference. No empirical implementation or sample reported.
Operationalizing reversible AI leadership via DAR can preserve human accountability while enabling AI-led decisions where appropriate.
Conceptual argument supported by the combined use of authority states, Reversal Register logging, and override mechanisms; no field validation provided.
DAR incorporates stabilizing mechanisms—hysteresis bands and safe-exit timers—to reduce rapid oscillation of authority and improve stability of handovers.
Formal model components and design proposals (hysteresis and timers) with conceptual argument that these damp oscillation; no empirical validation reported.
Improved targeting and dynamic personalization increase marketing ROI by raising conversion rates and lowering customer acquisition costs (CAC).
Economic implication based on observed performance improvements in conversions and resource allocation in case studies; no comprehensive ROI/CAC empirical analysis or sample-size-backed estimates are given.
Online A/B or multi-armed tests comparing the BERT–GPT pipeline with RAG+RL against baseline marketing automation produce measurable uplifts in CTR, engagement, conversion rate, retention, and revenue per user.
Paper reports that online experiments were conducted measuring these outcomes and observing uplifts; however, the paper does not provide numeric uplift magnitudes, confidence intervals, or sample sizes.
Privacy-preserving techniques such as federated learning, differential privacy (DP), and homomorphic encryption can mitigate privacy leakage while enabling model updates and secure aggregation.
Methods section describes applying federated learning with DP mechanisms on gradient updates and homomorphic encryption for aggregation; feasibility is argued but no empirical privacy-utility trade-off results are provided.
Comparative evaluations and case studies show consistent improvements over traditional marketing automation across engagement and conversion metrics, driven by better intent recognition, contextually appropriate messaging, and adaptive delivery policies.
Reported comparative evaluations (offline metrics and online A/B tests) and case studies attributing gains to improved intent recognition and adaptive policies; empirical details (sample sizes, statistical significance) are not reported in the paper.
Continuous online adaptation of models and policies—updating from streaming user interactions—enables per-session and lifetime personalization that improves engagement and conversion outcomes.
Modeling pipeline includes streaming updates and online adaptation; evaluations include online experiments and retention/engagement measurements. (No numerical magnitudes or update frequencies provided.)
An RL layer that formulates content selection as a contextual bandit / policy optimisation problem improves content selection and delivery using real-time reward signals (CTR, dwell time, conversions).
Paper describes RL-based policy optimisation using reward signals (CTR, session length, conversion events, LTV proxies) and reports online experiments/A/B tests where adaptive policies outperform static rules; exact algorithms and sample sizes not detailed.
RAG anchors generated content to up-to-date product/catalog/contextual knowledge and reduces hallucinations, increasing factuality of marketing messages.
Architectural description of RAG combining retrieved structured/unstructured knowledge with generative models; factuality/reduction in hallucinations evaluated in offline generation quality assessments using human raters and automatic factuality metrics.
GPT-family decoders generate tailored marketing content (ad copy, email text, chat responses) that matches user context and tone more effectively than template-based generation.
System uses GPT conditioned on user context and product info; generation quality evaluated via human raters and automatic relevance/factuality metrics in offline evaluations. (No quantitative effect sizes reported.)
An integrated BERT–GPT pipeline augmented with retrieval-augmented generation (RAG) and reinforcement learning (RL) substantially outperforms conventional rule-based or template-driven marketing automation.
Comparative evaluations and case studies reported in the paper, including online A/B or multi-armed tests comparing the full pipeline vs baseline automation and measuring CTR, engagement, conversion rate, retention, and revenue per user. (Sample sizes and statistical details are not specified in the paper.)
Continuous human-in-the-loop oversight, monitoring, and retraining are required to maintain quality and prevent model drift.
Practitioner reports and conceptual literature synthesized in the review advocating monitoring and retraining; no longitudinal empirical study provided here.
Transparent disclosure to customers about AI involvement helps preserve trust.
Conceptual analyses and referenced empirical/regulatory discussions in the literature aggregated by the review; this paper presents no new experimental evidence on disclosure effects.
Hybrid designs that automate low-risk, high-volume tasks while routing complex, judgment-sensitive cases to humans produce the best operational outcomes.
Inferred best-practice from aggregated empirical studies, industry examples, and conceptual reasoning; no controlled comparative trials presented in this review.
Agent augmentation via suggested responses, summarization, and information retrieval improves agent productivity.
Aggregated evidence from prior empirical research and practitioner reports cited in the review; no new measurements or sample sizes presented here.
Generative AI enables personalization at scale through automated tailoring of messaging and recommendations.
Qualitative synthesis of empirical studies and industry reports showing automated personalization use-cases; no systematic effect-size estimates or new quantitative data in this review.
Generative AI provides 24/7 availability and cost-effective scaling of routine interactions.
Industry case examples and prior empirical studies aggregated in the review; no original data or quantified sample sizes provided in this paper.
Generative AI can materially transform customer service and strategic communication by enabling continuous automation, scalable hyper-personalization, and effective agent augmentation.
Nano review: qualitative aggregation and synthesis of existing empirical studies, industry case examples, and conceptual analyses. No novel primary data or sample size; conclusion drawn from heterogeneous secondary sources and practitioner reports (not a systematic meta-analysis).
There is a need for standards around evaluation, bias mitigation, provenance, and accountability in AI-assisted ideation and design.
Policy recommendation motivated by documented biases, errors, and provenance issues in the reviewed studies; grounded in the synthesis's critique of existing practice.
There will likely be complementarity-driven increases in demand for evaluative, integrative, and domain-expert roles (curators, synthesizers, implementation experts).
Inference from task-level studies and economic reasoning about complementarities between AI generative capability and human evaluative skills; empirical labor-market evidence is limited in the reviewed literature.
Lower search and idea-generation costs enabled by LLMs may speed early-stage R&D and increase the gross flow of candidate innovations.
Theoretical economic interpretation supported by empirical findings of increased idea volumes in experimental/field studies summarized in the review; no long-run causal firm-level evidence presented.
Generative AI accelerates early-stage hypothesis and prototype development by providing scaffolded prompts and procedural suggestions.
Applied case evidence and experimental studies summarized in the review showing reduced time or increased productivity in early-stage experimental/design tasks when using LLM assistance; no pooled effect size presented.
Empirical studies document that AI-assisted tools can help break cognitive fixation and generate cross-domain analogies.
Cited experimental tasks and lab studies in the literature showing higher incidence of analogical or cross-domain suggestions from LLMs and improvements on fixation-related task metrics; heterogeneity across tasks and measures.
Generative AI provides scaffolded, structured support that aids systematic hypothesis formation, prototyping steps, and decomposition of complex problems.
Review of design/ideation studies and applied case evidence where LLMs produced stepwise plans, decomposition prompts, or hypothesis scaffolds; evidence drawn from multiple short-term experimental and applied studies, sample sizes and exact designs vary by study.
Generative models rapidly produce many candidate ideas, analogies, and associative prompts that help overcome cognitive fixation.
Synthesis of experimental ideation and design studies reporting increases in number of ideas and examples of reduced fixation when participants used LLM outputs; heterogeneous sample sizes across cited studies (not reported in review).
Generative AI can raise per-worker productivity for tasks involving brainstorming, drafting, and prototyping, but realized gains depend on downstream filtering and implementation costs.
User studies showing higher output on specific tasks (brainstorming/drafting), combined with qualitative reports of filtering/implementation effort; many studies measure immediate task output but not net realized productivity after implementation.
Generative AI can increase creative output in both lab and field tasks as judged by external raters.
Controlled experiments and field studies reporting higher judged creativity/novelty scores for AI-assisted outputs versus controls; judged creativity/novelty is typically assessed by human raters using rubric-based scoring.
AI assistance helps people overcome fixation and produces cross-domain analogies that they might not generate alone.
Experimental studies and qualitative analyses documenting reductions in fixation effects and increases in cross-domain analogical suggestions when participants use generative models.
Generative AI supports systematic problem breakdown and early-stage prototyping, accelerating hypothesis generation and prototype development.
Field case studies of AI-supported prototyping and lab/user studies reporting reduced time-to-prototype and generated hypotheses; measures include time-to-prototype and user-reported usefulness.
Generative AI boosts ideational fluency—the quantity and diversity of ideas produced in brainstorming tasks.
Controlled experiments and user studies measuring number and diversity of ideas with and without AI assistance; typical study designs compare participant idea counts/uniqueness across conditions (note: many studies use small or convenience samples).
When used as a 'cognitive co-pilot' that expands the solution space and challenges assumptions while humans curate and evaluate, generative AI generates economic value.
Inferred from experimental and field findings showing increased idea quantity/diversity and faster prototyping combined with qualitative studies showing human curation is needed; economic interpretation drawn from the review rather than direct macroeconomic measurement.
Generative AI serves a dual cognitive role: (1) a high-volume catalyst for divergent idea generation and cross-domain analogy-making, and (2) a structured assistant for deconstructing complex problems and scaffolding hypotheses and prototypes.
Synthesis of controlled experiments, lab studies, field case studies, and qualitative analyses summarized in the review; evidence includes measures of idea fluency/diversity, examples of analogy production, and observations of AI-assisted problem decomposition in prototyping tasks. (Note: underlying studies are heterogeneous and often short-term or convenience samples.)
Agent augmentation (drafting replies, summarizing histories, suggesting actions) raises frontline productivity and can improve response consistency.
Pilot deployments and internal A/B tests cited that measure time saved by agents and improvements in draft quality/consistency; mostly short-run and firm-specific reports.
Hyper-personalization at scale can increase relevance of responses and customer engagement when fed high-quality signals.
Case studies and pilot deployments that applied personalization signals (customer history, behavioral data) and reported improved relevance/engagement metrics; evidence conditional on availability and quality of signals and largely non-randomized.
24/7 automation reduces routine handling time and operational costs for simple, repetitive queries.
Operational deployments and pilot studies reporting reduced handling times and cost-per-interaction for routine queries; some vendor-supplied before/after or A/B comparisons, but heterogeneous measurements and limited randomized evidence.
Reproducibility is a practical and valuable goal for the HCI field even where full independent replication remains contested.
Authors' argumentation based on the observed rate of reproducibility, qualitative feedback from authors, and identified gains in credibility and reuse when artifacts are reproducible.