
Google made a number of AI-related announcements at the Google I/O developer conference this week, including stronger security measures in its AI models to clamp down on the spread of misinformation via deepfakes and problematic outputs.
The company expanded its SynthID line of watermarking technologies to include the capability to insert invisible watermarks on AI-generated video and text. This way, documents can be traced back to the original sources. SynthID already had the ability to apply watermarks to AI-generated images and audio.
“We are … developing new tools to prevent the misuse of our models,” James Manyika, senior vice president at Google, said during Google I/O.
Watermarking AI-generated content is increasingly important with AI being used to create all types of content. Deepfake video and audio have already been used to spread misinformation and in business email compromise.
Google also announced two new AI models at I/O — Veo, which generates realistic videos, and Imagen 3, which generates life-like images. The new watermarking techniques will be implemented in both models to easily identify fakes and prevent the spread of misinformation, Manyika said. For example, all videos generated by Veo on VideoFX will be watermarked by SynthID.
“We’re doing a lot of research in this area, including the potential for harm and misuse,” Manyika said.
With SynthID watermarking, the AI model attaches a watermark to generated output, which could be a block of text or an invisible statistical pattern. It then uses a scoring system to identify the uniqueness of that watermark pattern to see if the text was AI generated or came from another source. Google is open-sourcing SynthID text watermarking to other vendors.
“SynthID for text watermarking works best when a language model generates longer responses, and in diverse ways — like when it’s prompted to generate an essay, a theater script or variations on an email,” Google wrote in a blog post.
Another thing the company noted during I/O was how it is protecting AI models with with AI-assisted red-teaming techniques. AI agents are trained to compete with each other to improve and expand their red team capabilities. The primary goal behind this adversarial technique is to reduce problematic outputs.
“We test our own models and try to break them by identifying weaknesses,” Manyika said. “Building AI responsibility means both addressing the risks and maximizing the benefits of people and society.”