Google DeepMind Releases A Handy Tool To Detect Text Written By AI
A lot of AI generated content, such as Toys R Us’ AI-generated origin story video trailer, have become very good, making it harder to determine whether humans or AI made the content. Because of this, many companies and people are looking for easy ways of being able to tell the difference between AI-generated content, and that created by humans. The DeepMind team has been working on providing effective and easy methods of being able to detect AI-generated content, announcing watermarks for images last year, and later for AI-generated video. Now, the company has unveiled another tool for helping identify AI-generated content, this time for text.
Here’s how SynthID watermarks AI-generated content across modalities. ↓ pic.twitter.com/CVxgP3bnt2
— Google DeepMind (@GoogleDeepMind) October 23, 2024
Watermarks are a recognizable image or pattern embedded into a digital or physical document to identify the owner, prevent unauthorized duplication, or add a decorative effect. They have been around for a while, with many photographers using them to safeguard their work. SynthID works by adding an invisible watermark directly into AI generated text.
In a press release, Google remarked, “Practically speaking, SynthID Text is a logits processor, applied to a model’s generation pipeline after Top-K and Top-P, that augments the model’s logits using a pseudorandom g-function to encode watermarking information in a way that balances generation quality with watermark detectability.”
Today, we’re open-sourcing our SynthID text watermarking tool through an updated Responsible Generative AI Toolkit.
— Google DeepMind (@GoogleDeepMind) October 23, 2024
Available freely to developers and businesses, it will help them identify their AI-generated content. 🔍
Find out more → https://t.co/n2aYoeJXqn pic.twitter.com/4uRKYaz57Y
While Google says SynthID does not compromise the quality, accuracy, creativity, or speed of generated text, it does have limitations. The watermark proved effective to some forms of tampering, such as cropping text, and light editing or rewriting. However, it was less reliable when it came to AI-generated text that had been written or translated from one language to another. It was also found to be less reliable in responses to prompts asking for factual information, such as asking for the capital of a country.
“Achieving reliable and imperceptible watermarking of AI-generated text is fundamentally challenging, especially in scenarios where LLM outputs are near deterministic, such as factual questions or code generation tasks,” remarked Soheil Feizi, an associate professor at the University of Maryland.
Because of all the possible downsides to technologies such as DeepMind’s SynthID, many are saying watermarks can only be one aspect of the solution. Irene Solaiman, Hugging Face’s head of global policy, remarked, “Watermarking is one aspect of safer models in an ecosystem that needs many complementing safeguards. As a parallel, even for human-generated content, fact-checking has varying effectiveness.”