Generative audiovisual AI could sharply cut the cost and time of content production, but unresolved data rights, governance gaps and authenticity concerns mean adoption and economic benefits are far from guaranteed; regulation, provenance systems and legal rulings will critically shape who captures value and whether society accepts synthetic media.
This paper builds up on secondary data analysis to articulate a narrative review describing the challenges to the adoption of artificial intelligence tools for audiovisual communication, identifying issues in regard to ethics, control, transparency and the legal framework. We make a critique of the term and delimit the research to generative neural-network based computational processes. The problems are examined in three proposed areas: the input, which includes the training material; the process, which deals with the systems necessary to run such services, their development and control; and the output, where the use of generated artifacts is analyzed. International legal and judiciary approaches are reviewed in each step, as they are fundamental for incoming unfoldings of the field. Finally, some hypothetical future scenarios based on current AI debates are imagined, which help readers foresee possible situations regarding the use of generative systems. We conclude that although the adoption of such technologies is unstoppable, their use might come with some limitations, regulations or conditions. More importantly, the general use of AI tools will convey a shift in the way we produce and consume information. While we can clearly foresee dramatic productivity increases, it is not clear how this production will be accepted by society. On one hand, studies already revealed that subjects rank AI-generated content higher than the one made by humans. On the other hand, the plethora of artificial media might provoke an overall rejection of the digital world.
Summary
Main Finding
Adoption of generative neural-network audiovisual tools is effectively inevitable and promises large productivity gains, but faces significant ethical, control, transparency and legal challenges across three stages—input (training data), process (development and control), and output (generated artifacts). These challenges will likely lead to limitations, regulations, and shifts in how information is produced and consumed; however, societal acceptance of AI-generated audiovisual media remains uncertain and could range from widespread uptake to broad rejection.
Key Points
-
Scope and critique
- Focus is limited to generative neural-network based audiovisual systems (e.g., deepfakes, synthetic voices, video/animation generation).
- The paper critiques the umbrella term “AI” and urges specificity around generative neural methods.
-
Three-stage framework for risks and issues
- Input (training data)
- Copyright, consent, and data provenance problems.
- Biases and representational harms encoded in training datasets.
- Difficulty of auditing or reconstructing training sets across jurisdictions.
- Process (systems and development)
- Concentration of capabilities among a few platforms/actors.
- Control, governance, and transparency deficits (explainability, model access).
- Operational risks: misuse, security, and opaque deployment.
- Output (use of generated artifacts)
- Authenticity, provenance, and trust erosion (deepfakes, misinformation).
- Legal liability for harms caused by generated content.
- Social acceptance uncertainty: some studies show higher ratings for AI content, but abundance of synthetic media may provoke distrust or rejection.
- Input (training data)
-
Legal and judiciary review
- Examination of international legal approaches at each stage; cross-border enforcement and divergent national rules create fragmentation.
- Judicial decisions and regulatory moves are central to shaping incentives and constraints going forward.
-
Future scenarios
- Productivity surge in content creation juxtaposed with unclear market and social acceptance.
- Possible regulatory regimes range from light-touch standards to strict controls on generation/use.
Data & Methods
-
Methodology
- Secondary data analysis and narrative (qualitative) literature review.
- Synthesis of academic, legal, and policy sources; review of international statutes and judicial approaches relevant to generative audiovisual AI.
- Construction of hypothetical future scenarios grounded in current debates and documented trends.
-
Strengths
- Broad interdisciplinary synthesis linking technical, ethical, and legal literatures.
- Structured framework (input/process/output) that clarifies where different risks and rules apply.
-
Limitations
- No original empirical data or primary quantitative analysis—findings rely on existing studies and cases.
- Potential selection bias in reviewed sources; conclusions are contingent on evolving legal decisions and technology.
- Narrowed to neural-network generative methods, so insights may not transfer to other AI paradigms.
Implications for AI Economics
-
Labor and productivity
- Large potential productivity gains in audiovisual production (lower marginal costs, faster content creation).
- Displacement risk for creative and production roles; changes in skill premiums (demand for prompt engineering, curation, verification).
- Complementarity vs substitution: human labor may shift toward oversight, curation, post-production, and creative direction.
-
Market structure and competition
- Economies of scale and data-driven advantages may concentrate market power in a few platforms or studios that control models and datasets.
- Entry barriers could rise due to data access, compute costs, and IP uncertainty; regulatory interventions will affect market dynamics.
-
Intellectual property and rents
- Ambiguities in copyright and dataset licensing affect who captures value from models and outputs (original creators vs model operators).
- New rent sources may arise from provenance, authentication services, and certified “human-made” labels.
-
Demand, prices, and consumer behavior
- If consumers prefer AI-generated content (as some studies indicate), demand shifts could lower prices for certain media types and increase consumption volume.
- Conversely, trust erosion from synthetic media could reduce demand for digital content generally, creating negative externalities for platforms and advertisers.
-
Externalities and social welfare
- Information-quality externalities (misinformation, reduced trust) can create social costs not internalized by producers—justifies policy interventions.
- Regulatory choices (e.g., liability rules, mandatory provenance/watermarking) will shape innovation incentives and social welfare outcomes.
-
Policy levers and economic design
- Effective policy instruments include data-governance rules, provenance and watermarking standards, liability frameworks, copyright clarifications, competition policy, and taxes/subsidies to internalize externalities.
- Harmonization of cross-border legal standards would reduce fragmentation costs but is politically challenging.
- Monitoring and measurement: economists and regulators will need new metrics for synthetic content prevalence, provenance adoption, and labor market impacts.
-
Investment and strategic behavior
- Firms may invest in proprietary datasets, model-locking, or certification/verification services to differentiate products.
- Insurance, compliance, and legal risk management become material costs influencing adoption timing and scale.
Overall, the paper implies that economic outcomes depend strongly on legal and regulatory design: rules that clarify data rights, provenance, and liability can preserve market incentives while mitigating harms; absence of such rules may lead either to concentrated platform rents and rapid adoption or to broad distrust and reduced valuations of digital audiovisual goods.
Assessment
Claims (18)
| Claim | Direction | Confidence | Outcome | Details |
|---|---|---|---|---|
| Adoption of generative neural-network audiovisual tools is effectively inevitable. Adoption Rate | positive | speculative | adoption rate of generative neural-network audiovisual tools |
0.01
|
| Generative audiovisual models promise large productivity gains in content creation (lower marginal costs and faster content production). Firm Productivity | positive | medium | productivity in audiovisual production (e.g., marginal cost per unit of content, throughput/time-to-produce) |
0.07
|
| Use of these models faces significant ethical, control, transparency, and legal challenges across three stages—input (training data), process (development/control), and output (generated artifacts). Ai Safety And Ethics | negative | high | presence and severity of ethical/legal/control challenges across input/process/output stages |
0.12
|
| Input-stage risks include copyright infringement, lack of consent, poor data provenance, and biases/representational harms encoded in training datasets. Ai Safety And Ethics | negative | high | legal/compliance risk and bias in generated outputs arising from training data |
0.12
|
| Process-stage risks include concentration of capabilities among a few platforms/actors and deficits in control, governance and transparency (e.g., limited explainability and restricted model access). Market Structure | negative | medium | market concentration of model capabilities and levels of governance/transparency |
0.07
|
| Output-stage risks include challenges to authenticity and provenance, erosion of trust (deepfakes and misinformation), and potential legal liability for harms caused by generated content. Ai Safety And Ethics | negative | high | authenticity/provenance verification success, consumer trust, incidence of misinformation-related harms |
0.12
|
| Cross-border enforcement difficulties and divergent national rules produce legal fragmentation in regulation and judiciary responses to generative audiovisual AI. Governance And Regulation | negative | medium | degree of legal fragmentation across jurisdictions (differences in statutes, enforcement outcomes) |
0.07
|
| Societal acceptance of AI-generated audiovisual media is uncertain and could range from widespread uptake to broad rejection. Adoption Rate | mixed | speculative | social acceptance/adoption levels of AI-generated audiovisual media |
0.01
|
| The paper's methodology is a secondary-data, narrative (qualitative) literature review; it contains no original empirical data or primary quantitative analysis. Other | null_result | high | presence or absence of original empirical data |
0.12
|
| A structured three-stage framework (input/process/output) clarifies where different risks and regulatory rules apply to generative audiovisual systems. Ai Safety And Ethics | positive | high | clarity and mapping of risk types to development/use stages |
0.12
|
| Limitations of the study include potential selection bias in reviewed sources and contingency of conclusions on evolving legal decisions and technology developments. Research Productivity | negative | high | reliability and generalizability of the review's conclusions |
0.12
|
| Generative audiovisual models pose displacement risk for creative and production roles, but also create demand for new skills (prompt engineering, curation, verification) and complementarities in oversight and post-production. Job Displacement | mixed | medium | employment levels in creative/production roles and demand for new skill categories |
0.07
|
| Economies of scale, data-driven advantages, and compute costs may concentrate market power in a few platforms or studios, raising entry barriers. Market Structure | negative | medium | market concentration (e.g., HHI), entry rates, and barriers to entry |
0.07
|
| Ambiguities in copyright and dataset licensing will affect value capture (original creators versus model operators) and may create new rent opportunities from provenance/authentication services or certified 'human-made' labels. Firm Revenue | mixed | medium | distribution of economic rents and revenue shares between content creators and model operators |
0.07
|
| If consumers prefer AI-generated content, demand shifts could lower prices and increase consumption volume for certain media types; alternatively, trust erosion could reduce overall demand for digital content. Consumer Welfare | mixed | medium | consumer demand, price levels, and consumption volume for digital audiovisual content |
0.07
|
| Information-quality externalities from misinformation and reduced trust impose social costs that are not internalized by producers, justifying policy interventions such as liability rules or provenance standards. Consumer Welfare | negative | medium | social-welfare losses from misinformation and trust erosion |
0.07
|
| Recommended policy levers include data-governance rules, provenance and watermarking standards, liability frameworks, copyright clarifications, competition policy, and taxes/subsidies to internalize externalities. Governance And Regulation | positive | high | effectiveness of specified regulatory instruments in mitigating harms from generative audiovisual AI |
0.12
|
| Firms are likely to invest in proprietary datasets, model-locking, certification/verification services, insurance, and compliance/legal risk management, which will influence adoption timing and scale. Adoption Rate | positive | medium | firm investment in defensive/proprietary assets and timing/scale of technology adoption |
0.07
|