The EU AI Act: A Landmark Framework

The European Union's Artificial Intelligence Act, which entered into force in 2024, represents the world's most comprehensive attempt to regulate AI systems across a broad range of use cases. Among its many provisions, the Act includes specific requirements directly relevant to synthetic media, deepfakes, and AI-generated content — with significant implications for developers, platforms, and content creators operating in or serving EU markets.

Key Synthetic Media Provisions

Transparency Obligations for AI-Generated Content

Under the AI Act, systems that generate or manipulate image, audio, or video content — including deepfake systems — are subject to transparency obligations. Specifically, outputs from such systems must be labeled as artificially generated or manipulated in a machine-readable format.

This requirement covers:

  • Deepfake video and audio content depicting real persons
  • AI-generated images presented as authentic photography
  • Synthetic text that could mislead about its AI origin (with some exceptions for clearly creative or assistive contexts)

Prohibited Uses of Deepfakes

The Act places certain deepfake applications in the highest-risk category or outright prohibits them. Specifically, the use of AI systems to deploy subliminal manipulation techniques or to exploit individual vulnerabilities — which can include deceptive deepfake impersonations — is prohibited when they cause or are likely to cause significant harm.

Biometric and Emotion Recognition Systems

The Act also places strict limits on real-time biometric identification systems and prohibits certain uses of AI to infer emotions in sensitive contexts, which intersects with technologies commonly used in synthetic media generation pipelines.

Obligations by Stakeholder Type

StakeholderKey ObligationTimeline
AI Developers (EU-based)Label synthetic outputs in machine-readable formatPhased from 2025
Online PlatformsDetect and label AI-generated content; transparency reportsAligned with DSA requirements
Content CreatorsDisclose AI-generated or manipulated media, especially of real personsFrom general application date
Broadcasters/PublishersLabel deepfake content, especially in news and political contextsFrom general application date

Intersection with the Digital Services Act

The EU AI Act does not operate in isolation. The Digital Services Act (DSA) — already in force for large platforms — requires very large online platforms to assess and mitigate systemic risks, which explicitly includes risks from deepfakes and AI-generated disinformation. Together, these two frameworks create a layered regulatory environment for synthetic media in Europe.

Global Implications

While the AI Act is EU legislation, its reach extends beyond European borders due to the "Brussels Effect" — the tendency of EU regulations to become de facto global standards as multinational companies apply consistent policies across markets. Several key points:

  1. Non-EU companies offering AI services to EU users must comply with relevant provisions.
  2. The Act's labeling standards are influencing parallel legislative efforts in the UK, US, and other jurisdictions.
  3. Industry bodies are working to align technical standards (such as C2PA) with the Act's transparency requirements.

What This Means in Practice

For organizations creating, distributing, or hosting AI-generated media, the core takeaway is straightforward: disclosure and labeling are becoming legally required, not just ethically recommended. Building provenance and labeling into content workflows now — before enforcement deadlines — is both a compliance strategy and a trust-building measure.