Skip to content

Content Labeling and Provenance

Content labeling sits at the intersection of technology, law, and politics: the democratic risks of deepfakes, election integrity, reputation, consumer deception, and the very distinguishability of AI output from human creation all converge here.

This topic covers: how content generated or materially modified by AI systems should disclose its synthetic nature to users or downstream systems.

Two technical paths:

  • Explicit labels / visible watermarks: overlays, corner marks, or textual prompts visible to the human eye (“AI-generated”).
  • Implicit labels / provenance: file metadata, invisible watermarks, cryptographic signatures (C2PA).
DimensionChinaUnited StatesEU
Legal tierDepartmental rule + mandatory national standard (dual-track)State law + industry commitments; no unified federal ruleStatute (AI Act) + harmonised standards (in development)
Core obligorsService providers, distribution platforms, uploading users (three-way)Varies by state law; employers (employment) + platforms (elections)Provider + Deployer
Explicit vs. implicitBoth required (dual-track)Explicit-dominantBoth required; implicit must be “machine-readable and robust”
In-force dateSep 1, 2025 (Labeling Measures + GB 45438)State timelines vary; federal NO FAKES / COPIED in considerationAug 2, 2026 (AI Act art. 50)
Technical standardGB 45438-2025 mandatory national standardC2PA (industry standard)CEN-CENELEC JTC 21 (harmonised standards in development)
EnforcementCAC with multi-ministry coordinationFTC + state AGsMember-state MSA + AI Office

The table is an index; consult each jurisdiction page for specific obligations — do not treat it as a compliance conclusion.

Foundational research on deepfakes and synthetic media

Section titled “Foundational research on deepfakes and synthetic media”
  • Chesney & Citron (2019), “Deep Fakes: A Looming Challenge for Privacy, Democracy, and National Security” (California Law Review) — the foundational legal text on deepfakes.
  • Paris & Donovan (2019), “Deepfakes and Cheap Fakes” (Data & Society) — extends the frame to low-tech “cheap fakes” (mismatched photos, edits, voiceovers).
  • Farid (UC Berkeley), “Creating, Using, and Combating Synthetic Media”, and the Farid Lab’s ongoing deepfake-detection research — the authority on forensic AI.
  • Ajder, Patrini, Cavalli, Cullen (2019), “The State of Deepfakes” (Deeptrace): a quantitative baseline for deepfake detection.
  • C2PA (Coalition for Content Provenance and Authenticity): the open content-provenance standard led by Adobe / Microsoft / Intel / BBC and others.
  • Leibowicz, McGregor, Ovadya (2021), “The Deepfake Detection Dilemma” (Partnership on AI).
  • Partnership on AI Synthetic Media Framework (2023): a voluntary industry framework aligning obligations across creator, distributor, and user.
  • Kirchenbauer, Geiping et al. (2023), “A Watermark for Large Language Models” (Maryland) — text watermarking technology.
  • Fernandez, Couairon et al. (Meta, 2023), “Stable Signature”: watermark embedding in image-generation models.
  • Ferrara (USC, 2024): empirical research on generative-AI use in the 2024 US elections.
  • Rini (2020), “Deepfakes and the Epistemic Backstop”: the epistemic-trust problem.
  • Coeckelbergh, “The Political Philosophy of AI”: democratic implications of AI content.
  • Woolley & Howard, the Computational Propaganda series.

International law and cross-border application

Section titled “International law and cross-border application”
  • Bradford, Digital Empires (chapter 9 on comparative deepfake governance).
  • Livingston, Risse, Valeriani: human-rights implications of AI-generated content.
  • Zhang Linghan 张凌寒, Zhang Jiyu 张吉豫, and others: legal analysis of deep synthesis and generative-AI labeling.
  • The official experts’ commentary collection (CAC, Mar 2025): a systematised reading of the Labeling Measures.
  • Matt Sheehan (Carnegie): English-language analysis of China’s Labeling Measures.

1. Robustness: how far can labels resist adversarial attack?

Section titled “1. Robustness: how far can labels resist adversarial attack?”
  • Technical reality: current watermarks can be defeated by screenshots, compression, or re-editing.
  • Research consensus (Farid Lab / Maryland / CMU): a fully robust watermark does not exist; the question is one of cost / capability trade-offs.
  • Regulatory implication: mandating labels may create false confidence — “labelled” is not the same as “authentic”.
  • Reverse problem: if labels are breakable, bad actors can forge “native” content, potentially creating more danger than in the pre-labeling era.

2. Explicit vs. implicit labels: trade-offs

Section titled “2. Explicit vs. implicit labels: trade-offs”
  • Explicit: user-visible but constraints artistic and creative freedom (see EU AI Act art. 50 art / satire exemption).
  • Implicit: does not affect user experience, but requires downstream tooling to detect — ordinary users cannot see it.
  • China’s dual-track: requires both; the EU permits implicit-only in certain scenarios; US state laws emphasise explicit labels.
  • EU AI Act art. 50(4): artistic / fictional / satirical works can use weakened disclosure “that does not disturb artistic appreciation”.
  • China’s Labeling Measures: no express art / satire exemption — the boundary is unclear in practice.
  • United States: state laws vary widely — California AB-2655 (election deepfakes), Texas SB 751, Minnesota HF 1370, etc.
  • China’s Labeling Measures: enforcement against foreign services serving domestic users is limited.
  • EU AI Act art. 50: extraterritorial effect (art. 2), but actual enforcement depends on member-state MSAs.
  • United States: no unified federal rule; state laws cannot bind extraterritorially.
  • When AI training uses AI-generated synthetic data, should it also be labelled?
  • AI Act art. 53 requires disclosure of the share of synthetic data in a training-data summary, but does not mandate chain-of-provenance labeling.
  • No academic consensus has emerged yet.
  • C2PA is the Adobe-led de facto international standard, with version 1.0 released in 2022.
  • Signatories / integrators: OpenAI (parts of DALL-E / Sora), Google (Imagen), Microsoft, BBC, Nikon (in-camera), Leica, Arm, Sony, Truepic, etc.
  • Technology: based on JUMBF containers + X.509 cryptographic signatures; records provenance, modification history, and toolchain.
  • Critique: requires end-to-end tool support; invisible to ordinary users.

GB 45438-2025’s interoperability with C2PA

Section titled “GB 45438-2025’s interoperability with C2PA”
  • GB 45438-2025: a mandatory Chinese national standard (in force Sep 1, 2025, same day as the Labeling Measures).
  • Field definitions: service provider name, content number, date of generation, model information, content type.
  • C2PA compatibility: the technical fields map, but the legal status and signature mechanism differ.
  • Practical challenge: can domestic Chinese AI-service outputs be recognised in overseas C2PA ecosystems? A cross-standard conversion mechanism has not been established.
  • The AI-specific technical committee under the European harmonised-standards body.
  • prEN 18286 (entering the Enquiry stage in Q1 2026) is the first AI Act-related harmonised standard.
  • The full standards suite will not necessarily be ready by Aug 2, 2026, when AI Act art. 50 applies → the Code of Practice is the transitional arrangement.

Labeling status of major AI services (Apr 2026)

Section titled “Labeling status of major AI services (Apr 2026)”
Company / productExplicit labelImplicit labelIntegrated standard
OpenAI DALL-E / SoraCorner mark (partial)C2PA (comprehensive)C2PA 1.3
Google Imagen / Gemini imagesCorner markSynthID (Google proprietary) + C2PASynthID + C2PA
Microsoft Designer / Bing ImageCorner markC2PAC2PA
Anthropic ClaudeTextual prompt (output includes “AI” disclosure)No systematic implicit labeling
Meta AI / LlamaPartial corner markLimited C2PA (Instagram / Facebook)C2PA partial
ByteDance Doubao and other domestic China servicesCompliant with China’s Labeling MeasuresGB 45438-2025National standard
Baidu Wenxin / Alibaba Qwen / DeepSeekChina Labeling Measures compliantGB 45438-2025National standard
MidjourneyWeak / absentLimitedCommunity dispute
xAI Grok imagesWeak / absentNone

Sep 1, 2025 — the day the Labeling Measures took effect:

  • The six major Chinese platforms — WeChat, Weibo, Douyin, Kuaishou, Bilibili, Xiaohongshu — simultaneously rolled out explicit AI corner marks + implicit metadata labeling.
  • Creator-initiated marking + platform auto-detection supplementation.
  • This is the empirical instance of Section 4.1.3 “three-level vertical coordination” from the thesis (central rule issuance + rapid platform roll-out).

2024–2026 Meta / YouTube / TikTok:

  • Proactively added “created or significantly modified by AI” labels.
  • Based on C2PA auto-detection + user self-declaration.
  • Integrated with content-moderation policy.
  • 2023 White House Voluntary Commitments: OpenAI, Google, Meta, Anthropic, and others pledged investment in watermarking and provenance.
  • 2024 Seoul AI Declaration: extended to synthetic-media governance cooperation.
  • PAI (Partnership on AI) Synthetic Media Framework (2023): a voluntary industry framework.

Implementation status across three jurisdictions (Q1 2026)

Section titled “Implementation status across three jurisdictions (Q1 2026)”
  • China: in force (Sep 1, 2025). Dual-track labeling + national standard. Operational for seven-plus months; platforms have cooperated broadly.
  • United States: none federally; dispersed state laws. The TAKE IT DOWN Act (2025) on non-consensual intimate imagery is one of the few AI-related federal statutes.
  • EU: applies Aug 2, 2026 (AI Act art. 50); the GPAI Code of Practice serves as transitional compliance.