Evidence (3062 claims)
Adoption
5227 claims
Productivity
4503 claims
Governance
4100 claims
Human-AI Collaboration
3062 claims
Labor Markets
2480 claims
Innovation
2320 claims
Org Design
2305 claims
Skills & Training
1920 claims
Inequality
1311 claims
Evidence Matrix
Claim counts by outcome category and direction of finding.
| Outcome | Positive | Negative | Mixed | Null | Total |
|---|---|---|---|---|---|
| Other | 373 | 105 | 59 | 439 | 984 |
| Governance & Regulation | 366 | 172 | 115 | 55 | 718 |
| Research Productivity | 237 | 95 | 34 | 294 | 664 |
| Organizational Efficiency | 364 | 82 | 62 | 34 | 545 |
| Technology Adoption Rate | 293 | 118 | 66 | 30 | 511 |
| Firm Productivity | 274 | 33 | 68 | 10 | 390 |
| AI Safety & Ethics | 117 | 178 | 44 | 24 | 365 |
| Output Quality | 231 | 61 | 23 | 25 | 340 |
| Market Structure | 107 | 123 | 85 | 14 | 334 |
| Decision Quality | 158 | 68 | 33 | 17 | 279 |
| Fiscal & Macroeconomic | 75 | 52 | 32 | 21 | 187 |
| Employment Level | 70 | 32 | 74 | 8 | 186 |
| Skill Acquisition | 88 | 31 | 38 | 9 | 166 |
| Firm Revenue | 96 | 34 | 22 | — | 152 |
| Innovation Output | 105 | 12 | 21 | 11 | 150 |
| Consumer Welfare | 68 | 29 | 35 | 7 | 139 |
| Regulatory Compliance | 52 | 61 | 13 | 3 | 129 |
| Inequality Measures | 24 | 68 | 31 | 4 | 127 |
| Task Allocation | 71 | 10 | 29 | 6 | 116 |
| Worker Satisfaction | 46 | 38 | 12 | 9 | 105 |
| Error Rate | 42 | 47 | 6 | — | 95 |
| Training Effectiveness | 55 | 12 | 11 | 16 | 94 |
| Task Completion Time | 76 | 5 | 4 | 2 | 87 |
| Wages & Compensation | 46 | 13 | 19 | 5 | 83 |
| Team Performance | 44 | 9 | 15 | 7 | 76 |
| Hiring & Recruitment | 39 | 4 | 6 | 3 | 52 |
| Automation Exposure | 18 | 16 | 9 | 5 | 48 |
| Job Displacement | 5 | 29 | 12 | — | 46 |
| Social Protection | 19 | 8 | 6 | 1 | 34 |
| Developer Productivity | 27 | 2 | 3 | 1 | 33 |
| Worker Turnover | 10 | 12 | — | 3 | 25 |
| Creative Output | 15 | 5 | 3 | 1 | 24 |
| Skill Obsolescence | 3 | 18 | 2 | — | 23 |
| Labor Share of Income | 8 | 4 | 9 | — | 21 |
Human Ai Collab
Remove filter
Returns to AI and digital investments are heterogeneous across firms and industries, implying adoption barriers and varied productivity impacts.
Across the 145 studies, reported effect sizes and qualitative findings vary by firm characteristics, industry sector, and technology readiness, as summarized in the review.
Impacts of digital transformation on productivity vary substantially by moderators such as digital competencies, organizational culture, leadership, and technology readiness.
Multiple included studies identified these factors as moderators/mediators in their empirical analyses; moderator effects were synthesized in the review.
Task-based labor effects: GenAI will substitute routine tasks (documentation, triage) and complement complex decision-making; net employment effects are ambiguous and vary by role.
Task-based model of labor and early observational/pilot studies; the paper highlights heterogeneity by specialty and role, but presents no comprehensive empirical employment-impact studies.
GenAI can reduce clinician time per case (productivity gains) but may increase utilization (more tests/treatments) if it lowers thresholds for intervention or aligns with revenue incentives.
Economic reasoning supported by early empirical and simulation work; the paper notes the possibility based on task substitution and induced demand literature; direct causal empirical evidence from large-scale deployments is limited.
Improved personalization via RS techniques can increase consumer surplus by better matching robot behaviors to user needs, but it also creates the potential for finer-grained price or content discrimination if monetized.
Economic reasoning and implications section; conceptual analysis without empirical measurement.
Both initial trust and inertia have statistically significant effects on GAICS adoption decisions.
Inferential statistical tests reported in the quantitative phase indicating significant pathways from initial trust and from inertia to adoption outcome (exact effect sizes and sample size not provided in the abstract).
Organizations’ adoption of Generative AI–enabled CRM systems (GAICS) is driven by initial trust and inertia.
Quantitative inferential analysis in the study's second phase testing the conceptual model (paper reports statistically significant relationships between initial trust, inertia, and GAICS adoption). Sample size and sector/country scope not reported in the abstract.
Effectiveness of ChatGPT varied by discipline; not all course contexts showed significant gains from allowing its use.
Heterogeneous treatment effects observed across the six courses; GLM and non-parametric tests indicated variation in effect sizes and statistical significance by course/discipline.
AI adoption acts as a site of power reconfiguration: roles, relationships, and accountability structures shift as AI is integrated into workflows.
Qualitative workshop data from 15 UX designers describing anticipated or observed shifts in accountability and role boundaries; cross-scale thematic synthesis.
Discourses of efficiency carry ethical and social dimensions—responsibility, trust, and autonomy become central concerns when tools shift who does what and who is accountable.
Recurring themes from the 15 UX designers' discussions and design choices during workshops; thematic coding emphasized responsibility, trust, autonomy linked to efficiency claims.
At the team scale, adoption triggers negotiations over collaboration patterns, division of responsibility, and maintaining design rigor.
Group workshop activities and discussions among UX designers (n=15) where participants described team negotiation scenarios; team-level themes identified in analysis.
At the individual scale, designers expressed trade-offs among efficiency gains, opportunities for skill development, and feelings of professional value.
Individual- and small-group reflections in the 15-person workshop study; thematic coding highlighted these three recurring themes at the individual level.
Organizations frame AI adoption around competitiveness and efficiency, while workers (UX designers) weigh those efficiency framings against professional worth, learning, and autonomy.
Participants' reports during the qualitative design workshops (n=15) showing differences between organizational rhetoric and worker concerns.
Adoption outcomes depend on interactions among individual, team, and organizational incentives and norms (three analytic scales).
Cross-scale coding and synthesis of workshop data from 15 UX designers; analyses grouped themes into individual, team, and organizational scales.
Designers’ decisions about integrating AI reflect trade-offs between efficiency and social/ethical concerns (skill development, autonomy, accountability).
Workshop prompts and group discussions with 15 UX designers; thematic analysis identified recurring trade-off narratives between efficiency and professional/ethical considerations.
AI adoption reconfigures roles, responsibilities, trust, and power within organizations.
Qualitative data from design workshops with 15 UX designers; participants' reflections and group discussions coded using cross-scale thematic analysis (individual, team, organizational).
Heterogeneous and changing users (skill, mental models, incentives) produce heterogeneous and time-varying treatment effects, complicating inference from average uplift estimates.
Practitioner descriptions from 16 interviews highlighting user heterogeneity and learning/adaptation over time; authors' implication that averages may be insufficient.
Human uplift studies (typically RCTs measuring how AI changes human performance relative to a status quo) are a useful tool for informing deployment and policy decisions but face systematic validity challenges when applied to frontier AI systems.
Qualitative thematic synthesis of semi-structured interviews with 16 experienced practitioners across biosecurity, cybersecurity, education, and labor; authors' analytic mapping of interview themes to research lifecycle stages.
AI agents are useful as breadth tools and for pre-deployment checks but lack the protocol-specific and adversarial reasoning required to replace human auditors; human-in-the-loop workflows are the best use.
Study observations: agents reliably flag well-known patterns and respond to human-provided context, but fail to perform robust end-to-end exploit generation and are sensitive to scaffolding and configuration.
NFD can raise productivity in expert-heavy tasks by capturing tacit process knowledge and reducing repetitive cognitive effort, but the effect on employment is nuanced—routine parts may be automated while humans remain central to oversight and knowledge contribution.
Claims drawn from implications and the case study where analyst effort per task decreased and practitioners reported value; employment impact discussion is conceptual and speculative.
Highly personalized agents developed via NFD create stronger switching costs because crystallized knowledge assets are sticky, and economies of scale depend on the transferability of those assets across users or firms.
Conceptual reasoning in the paper's market structure and returns sections; supported by qualitative observations from the case study about personalization and reuse limits. No large-scale market data.
NFD shifts the economic tradeoff from large up-front engineering investment to ongoing human-in-the-loop investment; marginal cost of improving an agent becomes tied to practitioner time and crystallization efficiency rather than purely engineering labor.
Implications for AI economics section—conceptual analysis drawing on the NFD model and case study observations. No large-scale economic data provided.
The particular statement’s wording/ambiguity is a dominant source of labeling variability (statement dependence outweighs annotator-level effects).
Variance observed across repeated labeling of the same statements and strong statement-level effects in GEE models that account for repeated observations per statement and per participant.
Sentiment perception of short, decontextualized messages in team-based software projects is only moderately stable within individuals and is strongly statement-dependent.
Longitudinal repeated-measures study with 81 student participants across four survey rounds. In each round participants labeled 30 decontextualized statements for sentiment. Descriptive stability analyses showed only moderate within-person consistency and large between-statement variation.
Virtual–physical ecosystems and continuous validation raise new regulatory models (post-market surveillance, continuous certification), changing compliance costs and liability allocation.
Regulatory and safety implications raised in workshop panels and consensus recommendations captured in the workshop documentation (NSF workshop, Sept 26–27, 2024).
Human–AI collaboration frameworks will shift task allocation in clinical settings, affecting labor demand in clinical roles with potential for both complementarity and substitution effects.
Workshop discussion on systems/workflows and labor impacts from interdisciplinary participants (clinicians, researchers, industry) summarized in the report (NSF workshop, Sept 26–27, 2024).
Investment trade-offs exist between capital intensity (hardware co-design) and broader access; policy should balance platform funding with incentives for diversity and competition.
Workshop discussion and recommendation on funding trade-offs and policy implications from panels at the NSF workshop (Sept 26–27, 2024).
AI functions like a capital-augmenting technology that substitutes routine tasks while complementing creative and coordination tasks, altering the capital–labor mix and returns to different human capital types.
Conceptual framing and synthesis of literature and survey impressions; not directly tested empirically in the paper.
AI-driven automation will shift labor demand away from routine coding toward higher-order tasks (architecture, design, systems thinking, tool supervision), consistent with skill-biased technological change.
Theoretical implications drawn from observed substitution of routine tasks in literature and practitioner expectations in the survey; no labor-market causal analysis presented.
Benefits and uptake of AI tools are heterogeneous: they vary by team size, application domain (e.g., safety-critical vs. consumer software), and organizational process maturity.
Subgroup comparisons implied from survey (e.g., by role or domain) and literature examples; explicit subgroup sample sizes and statistical tests not provided in the summary.
AI augments developers rather than fully replacing them for complex, creative tasks; automation mainly substitutes routine work and complements higher-skill activities.
Synthesis of literature and survey responses indicating tool usage patterns and practitioner expectations about role changes; no experimental displacement studies reported.
RATs create both opportunities (public goods like shared trails that reduce duplication) and risks (surveillance, monetization without consent, concentration of network effects on large platforms).
Normative and policy analysis in the paper outlining possible externalities; no empirical assessment of magnitude or likelihood.
RAD remains competitive on helpfulness, incurring only modest or no loss in helpfulness in the reported experiments.
Empirical comparisons between RAD and baseline methods on helpfulness metrics reported in the paper (details on tasks, metrics, and sample sizes not provided in the summary).
Effective ISP depends on high-quality internal data and sometimes external data sharing across partners, raising issues around data ownership, incentives to share, and the design of contracting/market mechanisms to internalize coordination gains.
Case evidence on importance of data quality and authors' policy/contractual discussion; conceptual argument informed by interviews about data-sharing frictions.
ISP automation shifts labor demand toward higher-skill roles (data governance, analytics, cross-functional coordination) and reduces demand for routine forecasting and manual reconciliation tasks.
Interview reports and authors' task-based inference across cases, supplemented by economic reasoning about task reallocation.
ISP is relevant across multiple sectors (FMCG, manufacturing, retail) but outcomes and capabilities are heterogeneous by firm size and legacy IT footprint.
Sample composition includes firms from FMCG, manufacturing, and retail; authors report cross-case heterogeneity linked to firm characteristics and IT legacy.
Technology alone is insufficient; successful ISP requires cross-functional collaboration and continuous process improvement to realize gains from digital integration.
Cross-case interview evidence showing cases where digital tools did not produce expected benefits until processes and collaboration were changed; authors' synthesis of recurring barriers and enablers across the five cases.
Integrated Supply Planning (ISP) improves resilience and competitive performance only when advanced technologies (notably AI-enabled forecasting and ERP integration) are combined with organizational alignment, leadership commitment, and a data-driven culture.
Qualitative multi-case study (n = 5 medium-to-large organizations across FMCG, manufacturing, retail); cross-case comparison of semi-structured interviews with supply chain professionals reporting instances where technology adoption produced gains only alongside organizational enablers.
Standardized explainability requirements (audits, disclosure mandates) will affect market entry, favor incumbents with resources to meet standards, and create demand for third-party auditors and certification services.
Policy- and regulatory-focused literature synthesized in the review; claims are deductive implications from governance proposals and descriptive accounts rather than empirical causal tests.
Implementing explainability increases upfront development costs (tooling, documentation, UIs, training) and ongoing compliance/monitoring costs, but can lower downstream costs from litigation, audits, and reputational harm.
Synthesis of economic and policy literature in the review describing cost components and trade-offs; statements are conceptual and based on reviewed case studies and analyses rather than primary cost accounting.
Firm returns to AI adoption depend crucially on sociotechnical investments (training, redesign, knowledge infrastructure), so AI price/performance alone is an incomplete predictor of adoption returns.
Conceptual claim grounded in organizational literature synthesized in the paper; no firm-level econometric evidence presented within the paper itself.
Economic models of AI impact should move beyond simple task-automation/substitution frameworks to incorporate team-level complementarities and cognitive-process primitives (reasoning, memory, attention).
Theoretical recommendation for economists based on the paper's framework; supported by conceptual arguments rather than empirical re-specification or estimation shown in the paper.
Sociotechnical determinants — team composition, trust calibration, shared mental models, training regimes, and task structure — materially shape Human–AI team effectiveness beyond algorithmic performance alone.
Integrative review of multiple literatures (organizational behavior, human–computer interaction, psychology); presented as conceptual determinants; no empirical quantification provided in the paper.
Task reallocation: demand will fall for routine, automatable tasks and rise for complementary, cognitive, and governance tasks.
Task‑level decomposition and theoretical arguments about comparative advantage between AI and humans; no quantitative labor market estimates.
Overall, AI will be augmentative: many roles will transform rather than disappear; transition costs and task reallocation are the primary labor‑market challenges.
Synthesis of task‑based automation/complementarity analysis and scenario reasoning; paper explicitly notes lack of large‑sample causal evidence.
Within the next five years, AI will become an embedded, augmentative co‑pilot across software development and adjacent tech professions, shifting daily work from manual, task‑level activities to higher‑order, idea‑driven collaboration with intelligent systems.
Conceptual, forward‑looking analysis synthesizing current AI capability trends, illustrative examples of existing AI assistants, and scenario reasoning; no empirical longitudinal data or sample size reported.
AI has the potential to reduce diagnostic variability and improve access to specialist-level interpretation in underserved areas, but realized benefits depend on affordability, validation, and regulatory acceptance.
Potential benefits inferred from automation capabilities reviewed; contingent factors drawn from policy and implementation literature included in the narrative review.
AI-driven efficiency gains (reduced reading times, faster documentation) can lower per-patient labor costs and increase throughput, but net savings depend on reimbursement structures and implementation costs.
Empirical reports of time-savings in workflow studies and economic analysis in the review noting dependency on reimbursement and integration costs; no quantitative pooling.
Short-term physician substitution is limited; demand may increase for clinicians with oversight, escalation, and integrative skills.
Economic reasoning and task-complementarity arguments derived in the narrative review, supported by observed limitations of AI tools in open-ended and embodied tasks.
Clinical integration faces challenges including uncertainty quantification, clear escalation pathways, and user interfaces that support effective human oversight.
Policy, implementation, and technical literature included in the narrative review discussing difficulties in providing calibrated uncertainty estimates, embedding escalation workflows, and UX design for clinician-AI interaction.