How SynthID Detects AI-Generated Media
Google DeepMind Open-Sources SynthID Watermarking
Developers gain tools to identify AI-generated content through invisible watermarks.
A digital magnifying glass highlights invisible watermarking patterns over text and digital imagery in a clean, high-tech editorial style.
Photo: Avantgarde News
Google DeepMind released its SynthID technology as open-source code on May 15, 2026. This tool allows developers to embed invisible watermarks into AI-generated text and media [1][3]. The technology aims to help users distinguish between human-made content and machine-generated outputs [1][2].
The SynthID tool works by subtly altering the probability of words or pixels during the generation process [2][3]. These changes are invisible to the human eye or ear but remain detectable by specialized software even after editing [1][2]. By making the tool open-source, Google allows more developers to integrate safety features into their own AI models [2].
This move comes as tech companies face pressure to address the spread of AI-generated misinformation [1]. While SynthID is not a perfect solution, it provides a standard for tracking synthetic media across the industry [3]. Google originally launched the technology for its own Gemini and Vertex AI platforms before this public release [1][2].
Editorial notes
Transparency note
AI assisted drafting. Human edited and reviewed.
- AI assisted
- Yes
- Human review
- Yes
- Last updated
Risk assessment
Reviewed for sourcing quality and editorial consistency.
Sources
Related stories
View allTopics
About the author
Avantgarde News Desk covers how synthid detects ai-generated media and editorial analysis for Avantgarde News.