The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

Labeling content as AI-generated or AI-enhanced reduces user engagement, particularly for emotional posts; delaying disclosure helps only for AI-enhanced content, not fully for AI-generated pieces.

AI content labeling and user engagement on social media: The role of AI level, content type, and disclosure timing
Freya Seeger, Michael Wessel, Christiane Lehrer · March 24, 2026 · Electronic Markets
openalex rct medium evidence 7/10 relevance DOI Source PDF
Labeling social-media visuals as AI-generated or AI-enhanced reduces viewers' affective and reported behavioral engagement relative to human-created labels—especially for emotional content—while delaying disclosure partially restores affective engagement for AI-enhanced but not AI-generated content.

Abstract The rapid adoption of generative AI by content creators, coupled with the emergence of legal requirements for labeling AI-generated content, raises important questions about the implications of AI on user engagement on social media platforms. We examine how the level of AI involvement (human-created, AI-enhanced, or AI-generated), content type (emotional or rational), and disclosure timing (early or late) impact user engagement through two online experiments (study 1: n = 325; study 2: n = 371) conducted via the crowdsourcing platform Prolific. Participants (mean age = 35 years; 55% female) were asked to view Instagram profiles containing visual content labeled as human-created, AI-enhanced, or AI-generated. The results show that labeling content as AI-generated or AI-enhanced reduced both affective and behavioral engagement compared to human-created content, particularly for emotional content. Late disclosure of AI involvement improved affective engagement for AI-enhanced content, but not for AI-generated content. These findings deepen the understanding of algorithmic aversion in the context of generative AI and offer practical guidance for creators and platforms navigating the tension between transparency and engagement in an increasingly AI-mediated content ecosystem.

Summary

Main Finding

Labeling visual social-media content as AI-generated or AI-enhanced reduces both affective and behavioral engagement relative to labeling it human-created, with the effect strongest for emotional content. Delaying disclosure (late vs. early) increases affective engagement for AI-enhanced content but does not restore engagement for fully AI-generated content.

Key Points

  • Experimental manipulations: level of AI involvement (human-created, AI-enhanced, AI-generated), content type (emotional vs. rational), and disclosure timing (early vs. late).
  • Two online experiments on Prolific: study 1 (n = 325), study 2 (n = 371); participants mean age = 35; 55% female.
  • Outcome measures: affective engagement (attitudinal responses) and behavioral engagement (intended actions toward the profile/content).
  • Main pattern:
    • AI labels (both enhanced and generated) reduced affective and behavioral engagement versus human-created labels.
    • Reduction was larger for emotional content than for rational content.
    • Late disclosure mitigated the negative effect only for AI-enhanced content’s affective engagement; it did not help for AI-generated content.
  • Interpreted as evidence of algorithmic aversion in the context of generative AI, especially when content appeals to emotions.

Data & Methods

  • Design: two preregistered online randomized experiments (between-subjects manipulations).
  • Sample: participants recruited via Prolific; combined N = 696 across studies; demographic snapshot provided (mean age 35, 55% female).
  • Stimuli: Instagram-style profiles presenting visual content; labels indicated whether content was human-created, AI-enhanced, or AI-generated.
  • Factors:
    • AI involvement: three levels (human, AI-enhanced, AI-generated).
    • Content type: emotional vs. rational appeals.
    • Disclosure timing: early (label shown upfront) vs. late (label shown after initial exposure).
  • Outcomes: self-reported affective engagement (e.g., liking, warmth, trust) and behavioral engagement (e.g., intentions to like/follow/share — reported intentions rather than real platform behavior).
  • Limitations: online convenience samples, artificial profile exposure, self-reported engagement (not observed real-world behavior), single platform and visual-content focus.

Implications for AI Economics

  • Creator incentives and monetization:
    • Labeling reduces engagement and thus may lower creators’ returns (ad revenue, sponsorships, follower growth) for AI-labeled content.
    • Creators face a trade-off: use generative tools to reduce production costs but risk lower engagement if labeled.
    • Concealment incentives: negative engagement effects could encourage creators to obscure AI involvement, raising enforcement and verification challenges.
  • Platform design and revenue:
    • Platform-level labeling policies can materially affect user engagement metrics and downstream monetization (ads, subscriptions).
    • Timing of disclosure matters: allowing or engineering later disclosure for AI-enhanced content could partially preserve engagement while maintaining some transparency—though this raises ethical and regulatory questions.
  • Regulation and policy trade-offs:
    • Mandatory labeling increases transparency but may impose negative economic externalities (reduced engagement, lower creator incomes).
    • Policymakers should weigh consumer-protection goals against potential dampening of creator and platform revenues; nuanced rules (e.g., distinguishing levels of AI involvement, or disclosure timing/format standards) may be more efficient than blunt mandates.
  • Market dynamics and content composition:
    • If emotional content suffers more from AI labels, creators may shift toward rational or informational content formats, changing the overall content mix and consumer welfare.
    • Platforms could see changes in supply-side behavior, with differential adoption of generative tools across creator types and genres.
  • Research and modeling priorities:
    • Need for economic models of incentives under disclosure regimes (dynamic entry, signaling, enforcement costs).
    • Empirical follow-up: natural-field experiments measuring actual engagement and revenue impacts, heterogeneity across user segments, and long-run behavioral adjustments.
  • Practical recommendations:
    • Platforms and policymakers should consider graduated labeling (distinguish AI-enhanced vs AI-generated) and experiment with disclosure timing/format to balance transparency with engagement.
    • Invest in verifiable audit mechanisms to prevent concealment while reducing blunt engagement losses (e.g., reputational certification, contextual disclosures).

Assessment

Paper Typerct Evidence Strengthmedium — Randomized design provides strong internal validity and the finding is replicated across two studies, but evidence is limited by modest, non-representative crowdsourced samples, artificial stimulus presentation (simulated Instagram profiles), reliance on self-reported affective/behavioral engagement rather than actual platform behavior, and short-term laboratory-like exposure. Methods Rigormedium — The studies use randomized, factorial experiments and replication across two samples, which is methodologically sound; however, the paper appears to rely on online convenience samples (Prolific), does not report field-validation with real engagement metrics, and no information is provided here about pre-registration, power analysis, or robustness checks against demand effects. SampleTwo online experimental samples recruited via Prolific (study 1: n = 325; study 2: n = 371), mean age ~35, 55% female; participants viewed simulated Instagram profiles with visual content manipulated by label (human-created, AI-enhanced, AI-generated), content emotionality (emotional vs rational), and disclosure timing (early vs late); engagement outcomes are affective and self-reported behavioral measures. Themesadoption governance IdentificationRandomized online experiments: participants were randomly assigned to view Instagram-style posts labeled as human-created, AI-enhanced, or AI-generated, with orthogonal manipulation of content type (emotional vs rational) and disclosure timing (early vs late); causal effects are identified via random assignment across conditions. GeneralizabilityCrowdsourced Prolific sample not nationally representative (limits demographic generalizability), Artificial, short-term exposure to simulated Instagram profiles may not reflect real-world consumption or long-run behavior, Engagement measured via self-report rather than observed platform metrics (likes/shares/follow behavior), Findings may vary across platforms, cultures, content genres, and creator reputations not covered in the studies, Legal disclosure formats and real-world label designs may differ from experimental manipulations

Claims (7)

ClaimDirectionConfidenceOutcomeDetails
Labeling content as AI-generated reduced both affective and behavioral engagement compared to human-created content. Adoption Rate negative high affective and behavioral engagement
n=696
0.6
Labeling content as AI-enhanced reduced both affective and behavioral engagement compared to human-created content. Adoption Rate negative high affective and behavioral engagement
n=696
0.6
The reduction in engagement from AI labeling (AI-generated or AI-enhanced) was particularly pronounced for emotional content compared to rational content. Adoption Rate negative high affective and behavioral engagement for emotional content
n=696
0.6
Late disclosure of AI involvement improved affective engagement for AI-enhanced content. Adoption Rate positive high affective engagement for AI-enhanced content under late disclosure
n=696
0.6
Late disclosure of AI involvement did not improve affective engagement for AI-generated content. Adoption Rate null_result high affective engagement for AI-generated content under late disclosure
n=696
0.6
The paper's findings deepen the understanding of algorithmic aversion in the context of generative AI and offer practical guidance for creators and platforms navigating transparency versus engagement trade-offs. Governance And Regulation mixed high interpretation of experimental results (algorithmic aversion / guidance implications)
n=696
0.1
Study methodology: Two online experiments were conducted via the crowdsourcing platform Prolific with sample sizes study 1: n = 325 and study 2: n = 371; participant mean age = 35 years; 55% female. Other null_result high study design and sample characteristics
n=696
1.0

Notes