The Commonplace
Home Dashboard Papers Evidence Digests 🎲
← Papers

Generative audiovisual AI could sharply cut the cost and time of content production, but unresolved data rights, governance gaps and authenticity concerns mean adoption and economic benefits are far from guaranteed; regulation, provenance systems and legal rulings will critically shape who captures value and whether society accepts synthetic media.

Ethical and societal challenges to the adoption of generative AI in communication.
Bruno Caldas Vianna, Carles Sora, Josep Torrelló · Fetched March 10, 2026 · Knowledge Commons (Lakehead University)
openalex review_meta low evidence 7/10 relevance DOI Source PDF
Generative neural audiovisual tools are likely to drive large productivity gains in content creation but face significant legal, ethical, control, and trust challenges across training data, development, and output stages that make adoption paths and market outcomes uncertain.

This paper builds up on secondary data analysis to articulate a narrative review describing the challenges to the adoption of artificial intelligence tools for audiovisual communication, identifying issues in regard to ethics, control, transparency and the legal framework. We make a critique of the term and delimit the research to generative neural-network based computational processes. The problems are examined in three proposed areas: the input, which includes the training material; the process, which deals with the systems necessary to run such services, their development and control; and the output, where the use of generated artifacts is analyzed. International legal and judiciary approaches are reviewed in each step, as they are fundamental for incoming unfoldings of the field. Finally, some hypothetical future scenarios based on current AI debates are imagined, which help readers foresee possible situations regarding the use of generative systems. We conclude that although the adoption of such technologies is unstoppable, their use might come with some limitations, regulations or conditions. More importantly, the general use of AI tools will convey a shift in the way we produce and consume information. While we can clearly foresee dramatic productivity increases, it is not clear how this production will be accepted by society. On one hand, studies already revealed that subjects rank AI-generated content higher than the one made by humans. On the other hand, the plethora of artificial media might provoke an overall rejection of the digital world.

Summary

Main Finding

Adoption of generative neural-network audiovisual tools is effectively inevitable and promises large productivity gains, but faces significant ethical, control, transparency and legal challenges across three stages—input (training data), process (development and control), and output (generated artifacts). These challenges will likely lead to limitations, regulations, and shifts in how information is produced and consumed; however, societal acceptance of AI-generated audiovisual media remains uncertain and could range from widespread uptake to broad rejection.

Key Points

  • Scope and critique

    • Focus is limited to generative neural-network based audiovisual systems (e.g., deepfakes, synthetic voices, video/animation generation).
    • The paper critiques the umbrella term “AI” and urges specificity around generative neural methods.
  • Three-stage framework for risks and issues

    • Input (training data)
      • Copyright, consent, and data provenance problems.
      • Biases and representational harms encoded in training datasets.
      • Difficulty of auditing or reconstructing training sets across jurisdictions.
    • Process (systems and development)
      • Concentration of capabilities among a few platforms/actors.
      • Control, governance, and transparency deficits (explainability, model access).
      • Operational risks: misuse, security, and opaque deployment.
    • Output (use of generated artifacts)
      • Authenticity, provenance, and trust erosion (deepfakes, misinformation).
      • Legal liability for harms caused by generated content.
      • Social acceptance uncertainty: some studies show higher ratings for AI content, but abundance of synthetic media may provoke distrust or rejection.
  • Legal and judiciary review

    • Examination of international legal approaches at each stage; cross-border enforcement and divergent national rules create fragmentation.
    • Judicial decisions and regulatory moves are central to shaping incentives and constraints going forward.
  • Future scenarios

    • Productivity surge in content creation juxtaposed with unclear market and social acceptance.
    • Possible regulatory regimes range from light-touch standards to strict controls on generation/use.

Data & Methods

  • Methodology

    • Secondary data analysis and narrative (qualitative) literature review.
    • Synthesis of academic, legal, and policy sources; review of international statutes and judicial approaches relevant to generative audiovisual AI.
    • Construction of hypothetical future scenarios grounded in current debates and documented trends.
  • Strengths

    • Broad interdisciplinary synthesis linking technical, ethical, and legal literatures.
    • Structured framework (input/process/output) that clarifies where different risks and rules apply.
  • Limitations

    • No original empirical data or primary quantitative analysis—findings rely on existing studies and cases.
    • Potential selection bias in reviewed sources; conclusions are contingent on evolving legal decisions and technology.
    • Narrowed to neural-network generative methods, so insights may not transfer to other AI paradigms.

Implications for AI Economics

  • Labor and productivity

    • Large potential productivity gains in audiovisual production (lower marginal costs, faster content creation).
    • Displacement risk for creative and production roles; changes in skill premiums (demand for prompt engineering, curation, verification).
    • Complementarity vs substitution: human labor may shift toward oversight, curation, post-production, and creative direction.
  • Market structure and competition

    • Economies of scale and data-driven advantages may concentrate market power in a few platforms or studios that control models and datasets.
    • Entry barriers could rise due to data access, compute costs, and IP uncertainty; regulatory interventions will affect market dynamics.
  • Intellectual property and rents

    • Ambiguities in copyright and dataset licensing affect who captures value from models and outputs (original creators vs model operators).
    • New rent sources may arise from provenance, authentication services, and certified “human-made” labels.
  • Demand, prices, and consumer behavior

    • If consumers prefer AI-generated content (as some studies indicate), demand shifts could lower prices for certain media types and increase consumption volume.
    • Conversely, trust erosion from synthetic media could reduce demand for digital content generally, creating negative externalities for platforms and advertisers.
  • Externalities and social welfare

    • Information-quality externalities (misinformation, reduced trust) can create social costs not internalized by producers—justifies policy interventions.
    • Regulatory choices (e.g., liability rules, mandatory provenance/watermarking) will shape innovation incentives and social welfare outcomes.
  • Policy levers and economic design

    • Effective policy instruments include data-governance rules, provenance and watermarking standards, liability frameworks, copyright clarifications, competition policy, and taxes/subsidies to internalize externalities.
    • Harmonization of cross-border legal standards would reduce fragmentation costs but is politically challenging.
    • Monitoring and measurement: economists and regulators will need new metrics for synthetic content prevalence, provenance adoption, and labor market impacts.
  • Investment and strategic behavior

    • Firms may invest in proprietary datasets, model-locking, or certification/verification services to differentiate products.
    • Insurance, compliance, and legal risk management become material costs influencing adoption timing and scale.

Overall, the paper implies that economic outcomes depend strongly on legal and regulatory design: rules that clarify data rights, provenance, and liability can preserve market incentives while mitigating harms; absence of such rules may lead either to concentrated platform rents and rapid adoption or to broad distrust and reduced valuations of digital audiovisual goods.

Assessment

Paper Typereview_meta Evidence Strengthlow — The paper is a narrative, secondary-data literature and legal review with no original empirical or causal analysis; claims about productivity gains and market outcomes are synthesized from existing studies and hypothetical scenarios rather than identified causal estimates. Methods Rigormedium — Provides a clear, structured three-stage (input/process/output) framework and interdisciplinary synthesis of technical, legal, and policy literature, but it is a narrative review (not a systematic review or meta-analysis), susceptible to selection bias, and contains no primary data or formal robustness checks. SampleQualitative synthesis of academic papers, policy reports, legal statutes and judicial decisions across multiple jurisdictions, technical literature on generative neural audiovisual models, and selected empirical studies on consumer/market responses; no original dataset or primary quantitative analysis. Themesproductivity labor_markets governance adoption org_design GeneralizabilityFindings limited to generative neural-network audiovisual systems and may not extend to other AI paradigms (e.g., symbolic AI, non-audio/visual ML)., Conclusions depend on evolving technology and changing legal/regulatory decisions, reducing temporal generalizability., International legal and institutional heterogeneity limits cross-jurisdictional applicability of specific policy recommendations., Selection bias in reviewed sources (narrative review) may omit contrary evidence or sector-specific nuances., Lack of primary empirical estimates limits applicability to precise market or labor-impact magnitudes.

Claims (18)

ClaimDirectionConfidenceOutcomeDetails
Adoption of generative neural-network audiovisual tools is effectively inevitable. Adoption Rate positive speculative adoption rate of generative neural-network audiovisual tools
0.01
Generative audiovisual models promise large productivity gains in content creation (lower marginal costs and faster content production). Firm Productivity positive medium productivity in audiovisual production (e.g., marginal cost per unit of content, throughput/time-to-produce)
0.07
Use of these models faces significant ethical, control, transparency, and legal challenges across three stages—input (training data), process (development/control), and output (generated artifacts). Ai Safety And Ethics negative high presence and severity of ethical/legal/control challenges across input/process/output stages
0.12
Input-stage risks include copyright infringement, lack of consent, poor data provenance, and biases/representational harms encoded in training datasets. Ai Safety And Ethics negative high legal/compliance risk and bias in generated outputs arising from training data
0.12
Process-stage risks include concentration of capabilities among a few platforms/actors and deficits in control, governance and transparency (e.g., limited explainability and restricted model access). Market Structure negative medium market concentration of model capabilities and levels of governance/transparency
0.07
Output-stage risks include challenges to authenticity and provenance, erosion of trust (deepfakes and misinformation), and potential legal liability for harms caused by generated content. Ai Safety And Ethics negative high authenticity/provenance verification success, consumer trust, incidence of misinformation-related harms
0.12
Cross-border enforcement difficulties and divergent national rules produce legal fragmentation in regulation and judiciary responses to generative audiovisual AI. Governance And Regulation negative medium degree of legal fragmentation across jurisdictions (differences in statutes, enforcement outcomes)
0.07
Societal acceptance of AI-generated audiovisual media is uncertain and could range from widespread uptake to broad rejection. Adoption Rate mixed speculative social acceptance/adoption levels of AI-generated audiovisual media
0.01
The paper's methodology is a secondary-data, narrative (qualitative) literature review; it contains no original empirical data or primary quantitative analysis. Other null_result high presence or absence of original empirical data
0.12
A structured three-stage framework (input/process/output) clarifies where different risks and regulatory rules apply to generative audiovisual systems. Ai Safety And Ethics positive high clarity and mapping of risk types to development/use stages
0.12
Limitations of the study include potential selection bias in reviewed sources and contingency of conclusions on evolving legal decisions and technology developments. Research Productivity negative high reliability and generalizability of the review's conclusions
0.12
Generative audiovisual models pose displacement risk for creative and production roles, but also create demand for new skills (prompt engineering, curation, verification) and complementarities in oversight and post-production. Job Displacement mixed medium employment levels in creative/production roles and demand for new skill categories
0.07
Economies of scale, data-driven advantages, and compute costs may concentrate market power in a few platforms or studios, raising entry barriers. Market Structure negative medium market concentration (e.g., HHI), entry rates, and barriers to entry
0.07
Ambiguities in copyright and dataset licensing will affect value capture (original creators versus model operators) and may create new rent opportunities from provenance/authentication services or certified 'human-made' labels. Firm Revenue mixed medium distribution of economic rents and revenue shares between content creators and model operators
0.07
If consumers prefer AI-generated content, demand shifts could lower prices and increase consumption volume for certain media types; alternatively, trust erosion could reduce overall demand for digital content. Consumer Welfare mixed medium consumer demand, price levels, and consumption volume for digital audiovisual content
0.07
Information-quality externalities from misinformation and reduced trust impose social costs that are not internalized by producers, justifying policy interventions such as liability rules or provenance standards. Consumer Welfare negative medium social-welfare losses from misinformation and trust erosion
0.07
Recommended policy levers include data-governance rules, provenance and watermarking standards, liability frameworks, copyright clarifications, competition policy, and taxes/subsidies to internalize externalities. Governance And Regulation positive high effectiveness of specified regulatory instruments in mitigating harms from generative audiovisual AI
0.12
Firms are likely to invest in proprietary datasets, model-locking, certification/verification services, insurance, and compliance/legal risk management, which will influence adoption timing and scale. Adoption Rate positive medium firm investment in defensive/proprietary assets and timing/scale of technology adoption
0.07

Notes