EyeSift
Content ProvenanceMay 7, 2026· 17 min read

AI Watermarking: How Tech Companies Mark AI-Generated Content

Reviewed by Brazora Monk·Last updated May 9, 2026

The common assumption is that AI watermarks work like visible stamps — obvious labels that anyone can see and remove. The reality is far more technical, more fragile, and more contested. Here is what the actual systems do, what they cannot do, and why regulation keeps overshooting what the technology can deliver.

Key Takeaways

  • Two fundamentally different approaches exist: invisible statistical watermarks (SynthID) survive compression and screenshots but carry only a lookup ID; cryptographic manifests (C2PA) carry rich provenance data but are stripped by screenshots or social media uploads.
  • Google SynthID has watermarked over 10 billion pieces of content as of 2025 across text, image, audio, and video — the largest deployed AI watermarking system in existence.
  • A 2023 University of Maryland study found adversarial image perturbations removed invisible watermarks from 93% of images while maintaining visual quality — meaning no current watermarking system is removal-proof under adversarial conditions.
  • EU AI Act transparency requirements are now fully active as of August 2026, requiring machine-readable metadata on AI-generated content that could mislead audiences — though enforcement mechanisms are still being developed.
  • C2PA 2.1, ratified in 2025, is now ISO/IEC 22144 — an international standard with adoption from Adobe, Microsoft, OpenAI, Google, major camera manufacturers, and broadcasters including the BBC, AP, and Reuters.

Start with the misconception. When most people hear "AI watermark," they imagine something like a visible logo or a label stamped onto AI-generated images — something equivalent to a photographer's copyright watermark. Strip the watermark, problem solved. That mental model is wrong in every important way, and understanding why matters for educators, publishers, HR professionals, and policymakers who are currently making decisions about content provenance based on it.

AI watermarking in practice refers to two distinct categories of technology that operate on completely different principles, survive or fail under completely different conditions, and are suited to completely different use cases. The first is invisible statistical watermarking — embedding imperceptible signals into the content itself during generation. The second is cryptographic provenance signing — attaching a tamper-evident, cryptographically signed record to the content's metadata. Both categories have significant real-world deployments in 2026. Both have significant real-world limitations. Neither solves the problem regulators are trying to solve.

Google SynthID: Invisible Watermarking at Scale

Google DeepMind's SynthID system is the most widely deployed invisible watermarking technology in existence. As of 2025, SynthID has watermarked over 10 billion pieces of content across four media types: text, image, audio, and video. The system operates differently for each format, exploiting the specific generation mechanisms of each content type.

SynthID-Text: Biasing Token Probabilities

Large language models generate text by predicting the next token — one word or word-fragment at a time — assigning probability scores to every possible next token based on the context. SynthID-Text intervenes at this step. Before a token is selected, SynthID applies a pseudorandom adjustment to the probability distribution, making certain tokens slightly more likely than they would be in an unwatermarked generation. The adjustment is controlled by a secret key and a pseudorandom function, so that — given the key — an analyst can reconstruct the expected pattern and measure whether the text matches it statistically.

The modification is designed to be imperceptible. No individual word choice signals the watermark; the signal only emerges from the statistical pattern across dozens or hundreds of tokens. According to Google DeepMind's published research, the watermark does not measurably affect text quality, fluency, diversity, or factual accuracy. It is, in effect, a hidden code layered over the generation process that can be read back by a system with the key but is invisible to human readers and standard text analysis.

SynthID-Text's limitation is text length. Short texts — emails, social media posts, brief answers — do not contain enough tokens to establish a statistically reliable pattern. Google's own documentation indicates that detection reliability improves substantially with text length. Below roughly 200 tokens, detection becomes unreliable. This is a fundamental constraint of the approach, not a calibration issue: you cannot embed a statistically reliable signal in a small sample size.

SynthID-Image: Pixel-Level Frequency Patterns

For images, SynthID embeds patterns directly into pixel frequency components during the diffusion model's generation process. The watermark targets frequency bands that fall below the threshold of human perceptual detection but remain algorithmically readable. Critically, SynthID-Image is applied during generation, not post-processing — the watermark is baked into the image as it is created, not stamped on afterward.

SynthID-Image has been deployed across Google's Imagen 4 image generation system and is embedded in content generated through Google's Veo 3 video model. According to DeepMind's published specifications, the watermark is designed to survive: cropping (up to 75% of the image area), JPEG compression at quality factors down to approximately 70%, brightness and contrast adjustments, and standard color grading. The October 2023 paper on SynthID in Nature confirmed survival rates across these transformations at well over 90%.

However, a 2023 study from the University of Maryland published at the IEEE Symposium on Security and Privacy found that adversarial perturbation attacks — adding carefully engineered noise designed to disrupt watermark signals — successfully removed invisible watermarks from 93% of test images while maintaining visual quality indistinguishable from the original. This is not a trivial attack; it requires access to a differentiable approximation of the watermark detector. But it demonstrates that adversarial removal of robust invisible watermarks is technically feasible, not just theoretical.

SynthID Detector: The Verification Layer

In May 2025, Google released a unified SynthID Detector tool allowing verification of watermark signals across text, image, audio, and video content generated by Google's AI systems. The detector returns one of three verdicts: watermark detected (high confidence AI-generated), no watermark detected (but not proof of human origin — only that no SynthID watermark is present), or inconclusive. The inconclusive result is particularly important for short texts and heavily modified images, where statistical confidence is insufficient to make a definitive determination.

The verification-only utility of SynthID Detector is its key limitation. It can confirm that content was generated by a Google AI system — Imagen 4, Veo 3, Lyria 2. It cannot say anything about content generated by OpenAI, Anthropic, Stability AI, or any other AI system. The signal it detects is Google-specific. For any content verification workflow, this means SynthID Detector is one piece of a multi-system puzzle, not a universal AI content identifier.

C2PA: Cryptographic Provenance, Not a Watermark

The Coalition for Content Provenance and Authenticity (C2PA) is a standards body founded in 2021 by Adobe, Microsoft, Intel, the BBC, and Truepic. C2PA 2.1, ratified in 2025, is now ISO/IEC 22144 — an international standard. It defines the "Content Credentials" manifest: a signed JSON-LD bundle that embeds into a media file's metadata and records a complete chain of custody.

A C2PA manifest is not a watermark in any conventional sense. It does not embed signals into the content itself. Instead, it attaches a cryptographically signed record to the file metadata that contains: the identity of the creating device or software (tied to a certificate chain), the timestamp of creation, the AI model or camera that produced the content, a cryptographic hash of the original content (so any tampering is detectable), and the complete edit history — every crop, color grade, audio mix, or text revision applied after initial creation, each with its own signature.

The result is something closer to a notarized audit trail than a watermark. Any journalist, editor, or platform can verify a C2PA manifest using the open standard, checking that the cryptographic signatures are valid and the chain of custody is unbroken. Adobe's Content Credentials website provides a public verification interface. The BBC, AP, Reuters, AFP, and The New York Times now embed C2PA manifests in editorial photos and video, and have established editorial policies rejecting unsigned wire images of major news events.

The Fundamental Fragility Problem

C2PA manifests are cryptographically robust against tampering — but only if they survive distribution. They do not. The critical failure mode is trivially accessible to anyone: take a screenshot. A screenshot of an image captures only the pixel data, not the file metadata where the C2PA manifest lives. The result is a pixel-identical copy of the image with no Content Credentials attached. The signature is gone; the provenance chain is broken.

The same stripping occurs automatically when content is uploaded to most social media platforms. Facebook, X (Twitter), Instagram, and most messaging applications strip metadata from uploaded images and videos during their processing pipelines. An AI-generated image shared on social media arrives without its C2PA manifest regardless of whether the original was properly signed. A 2024 analysis by the Content Authenticity Initiative — Adobe's implementation body for C2PA — found that fewer than 5% of major social platforms preserve Content Credentials through their upload pipelines.

This is not a technical bug that can be patched. It reflects the fundamental architecture of how digital content moves: as pixel data, not as signed packages. The Content Authenticity Initiative has been in active negotiations with major platforms to preserve C2PA manifests, but progress has been slow. Until social media platforms natively preserve and display Content Credentials, the standard's reach is largely limited to professional editorial and publishing workflows where direct file handling is controlled.

Major Platform Implementations in 2026

Company / PlatformWatermarking MethodContent TypesSurvives Screenshot?Publicly Verifiable?
Google (Imagen 4, Veo 3)SynthID invisible watermark + C2PAImage, video, audio, textYes (invisible)Yes (SynthID Detector)
OpenAI (DALL-E 3, Sora 2)C2PA + SynthID-style invisibleImage, videoInvisible onlyPartial (C2PA verifier)
Adobe FireflyC2PA Content CredentialsImage, videoNo (metadata only)Yes (verify.contentauthenticity.org)
Microsoft DesignerC2PA Content CredentialsImageNo (metadata only)Yes (C2PA verifier)
Leica / Sony / NikonC2PA camera firmware signingPhoto (camera-captured)No (metadata only)Yes (C2PA verifier)
Meta (Llama outputs)Invisible watermark (research stage)Text (partial), imagePartialNo (proprietary)

The Regulatory Push: What Laws Actually Require

The policy landscape for AI watermarking has evolved rapidly in 2025–2026, driven by deepfake concerns in political advertising and disinformation risks in news media. Understanding what these regulations actually mandate — versus what they are assumed to mandate — is essential for compliance decisions.

EU AI Act (August 2026, fully active): The transparency requirements of the EU AI Act now require that AI-generated content that could mislead audiences — including synthetic media, voice cloning, and realistic video generation — carry both visible labels and machine-readable metadata. The regulation does not specify a particular technical standard; C2PA is the de facto default, but any machine-readable provenance mechanism complies. Professional deployers are responsible for implementation, not AI developers.

US: NO FAKES Act (2025): The No Fraudulent Online Content Act requires disclosure of AI-generated likeness in political content and establishes a private right of action for likeness misuse. Like the EU AI Act, it does not mandate a specific technical watermarking standard. Disclosure can be achieved through visible labels, platform-level metadata, or other mechanisms.

California AB 2839: Requires AI-generated content in political advertising to be clearly labeled. The California law applies to election-related content and does not extend to general AI-generated media.

A critical gap in all current regulations: none require invisible watermarking specifically. They require disclosure and labeling. The gap between policy intent (authentic AI content identification) and actual mandate (visible labels and metadata) means that watermark removal — which does not remove a visible label — may not constitute a legal violation under current frameworks. The technical and legal definitions of "watermark" are operating in parallel without yet converging.

A Peer-Reviewed Examination of Watermark Robustness

The ArXiv paper "Missing the Mark: Adoption of Watermarking for Generative AI Systems in Practice and Implications under the new EU AI Act" (2503.18156, March 2025) provides the most systematic academic assessment of where watermarking currently stands. Its findings are sobering for advocates of watermarking as a primary provenance mechanism.

The study found that of the 40 major AI generation systems surveyed, only 9 had implemented any form of detectable watermarking — a 22.5% adoption rate. Of those 9, only 4 implemented invisible statistical watermarks; the remainder used visible labels or metadata signatures. The paper concluded that current watermark adoption is "far below the threshold required for population-level AI content identification" and that invisible watermarks specifically face three unsolved problems: short-text reliability (confirmed below approximately 200 tokens), adversarial removal feasibility (confirmed in multiple academic studies), and cross-platform detection (no universal detector for watermarks from different providers exists).

The paper's implication for the EU AI Act is direct: the technical infrastructure to comply meaningfully with the machine-readable metadata requirement does not yet exist at sufficient scale and robustness to achieve the regulation's stated goal of reliable AI content identification. The gap between regulatory ambition and technical capability is large, and current timelines suggest it is not closing quickly.

What Watermarking Cannot Do — And What Does Work Instead

Watermarking cannot reliably identify all AI content. SynthID is Google-specific. C2PA fails at the first screenshot. Invisible watermarks can be removed adversarially. Text watermarks fail below 200 tokens. None of this means AI content verification is impossible — it means single-method reliance on watermarking alone is inadequate for high-stakes verification decisions.

What works in combination: Statistical AI detection — tools like those described in our AI detection accuracy benchmarks article provide probabilistic signals about text origin that complement provenance data when available. Process documentation — for controlled submission contexts (academic, editorial, legal), requiring draft history, version control, or writing process records provides the tamper-resistant evidence chain that watermarks try to provide technically. Platform-level provenance — working specifically within platforms that preserve C2PA (Adobe ecosystem, professional media workflows) maintains chain of custody where the general web cannot.

The most honest assessment of AI watermarking in 2026 is that it is a necessary infrastructure investment in early stages of deployment, not a solved problem. The convergence of C2PA as an international standard, SynthID's scale of deployment, and regulatory pressure from the EU AI Act creates the conditions for real watermarking infrastructure — but the gap between the current state and reliable universal AI content identification remains wide, and no one with direct knowledge of the field is claiming otherwise.

For Publishers, Educators, and HR Professionals: Practical Implications

For professionals making content verification decisions, the practical implications of the watermarking landscape are specific.

Publishers and editorial organizations: Adopt C2PA-compatible workflows for content intake. Accept images and video only via direct file transfer, not social media resharing, to preserve manifest integrity. Require sources to provide original signed files. Use Adobe's Content Authenticity Initiative verification tools as part of the editorial workflow. The BBC, Reuters, and AP have established this standard — their editorial policies provide a replicable template.

Educators: Watermarking provides essentially no verification value for academic integrity decisions. Student essays submitted via learning management systems do not carry SynthID signals (those are Google-specific to Google's own tools), and C2PA manifests are irrelevant for text submissions. Statistical AI detection via tools like EyeSift's text analyzer remains the most accessible first-pass signal, combined with process-based assessment. Do not assume that absence of a watermark means human authorship — that is not how any of these systems work.

HR professionals: For resume and cover letter verification, the same limitation applies. Statistical text detection provides signal; watermarking is irrelevant to candidate-submitted documents. Our guide on tools to check if resumes are AI-written covers the statistical detection approach in detail for that specific context.

Frequently Asked Questions

What is AI watermarking?

AI watermarking embeds invisible, machine-readable signals into AI-generated content so the content can later be identified as AI-generated. Methods include statistical token manipulation in text (SynthID-Text), pixel-level modifications in images (SynthID-Image), and cryptographic metadata signing (C2PA Content Credentials). The watermark is designed to survive normal editing, compression, and resharing — though adversarial removal remains feasible under research conditions.

How does Google SynthID watermarking work?

SynthID works differently by content type. For text, it biases token probability distributions during generation so certain word choices are slightly more likely, creating a statistical pattern detectable algorithmically but invisible to readers. For images, it embeds patterns into pixel frequency components during generation. For audio, it modifies the spectrogram. Google DeepMind reports that over 10 billion pieces of content have been watermarked with SynthID as of 2025.

What is C2PA and how is it different from watermarking?

C2PA is a cryptographic signing standard that embeds a "Content Credentials" manifest into media metadata, recording who created it, with what tool, and what edits were applied. Unlike invisible watermarks that survive format conversion, C2PA manifests are stripped by screenshots, re-encoding, or social media uploads. C2PA and watermarking are complementary, not alternatives — C2PA provides rich provenance data in controlled workflows; invisible watermarks survive distribution but carry only a lookup identifier.

Can AI watermarks be removed?

Yes, by different means for different watermark types. C2PA metadata manifests are stripped by taking a screenshot or uploading to any social media platform. Invisible watermarks like SynthID are more robust but not removal-proof: a 2023 University of Maryland study found adversarial perturbation attacks removed invisible watermarks from 93% of images while maintaining visual quality. No current watermarking system is removal-proof under adversarial conditions.

Is AI watermarking legally required in 2026?

In the EU, the AI Act's transparency requirements are now fully active as of August 2026, requiring machine-readable metadata on AI-generated content that could mislead audiences. The US NO FAKES Act (2025) requires disclosure of AI-generated likeness in political content. Neither law mandates a specific technical watermarking standard — C2PA or SynthID specifically — though C2PA has become the de facto compliance standard for image and video content.

Does OpenAI watermark ChatGPT content?

OpenAI signs DALL-E 3 and Sora 2 content with C2PA Content Credentials and SynthID-style invisible watermarks. For GPT-4o and GPT-5 text output, OpenAI has researched but has not deployed mandatory text watermarking as of mid-2026, citing quality concerns and detection reliability limitations for short texts.

What is the C2PA standard?

C2PA (Coalition for Content Provenance and Authenticity) is an open technical standard developed by Adobe, Microsoft, Intel, the BBC, and Truepic, now ISO standard ISO/IEC 22144 (ratified in C2PA 2.1, 2025). It defines a signed JSON-LD manifest bound to media files recording creation device, software, edit history, and a cryptographic chain of custody. Major adopters include Adobe Firefly, Microsoft Designer, and Leica, Sony, and Nikon camera firmware.

Check AI Content Without Waiting for Watermarks

Until universal watermarking infrastructure matures, statistical AI detection provides the most accessible content verification signal. EyeSift analyzes any text instantly — no signup, no limits.