Skip to main content

Google’s AI detection tool is now available for anyone to try

Google announced via a post on X (formerly Twitter) on Wednesday that SynthID is now available to anybody who wants to try it. The authentication system for AI-generated content embeds imperceptible watermarks into generated images, video, and text, enabling users to verify whether a piece of content was made by humans or machines.

“We’re open-sourcing our SynthID Text watermarking tool,” the company wrote. “Available freely to developers and businesses, it will help them identify their AI-generated content.”

Recommended Videos

SynthID debuted in 2023 as a means to watermark AI-generated images, audio, and video. It was initially integrated into Imagen, and the company subsequently announced its incorporation into the Gemini chatbot this past May at I/O 2024.

The system works by encoding tokens — those are the foundational chunks of data (be it a single character, word, or part of a phrase) that a generative AI uses to understand the prompt and predict the next word in its reply — with imperceptible watermarks during the text generation process. It does so, according to a DeepMind blog from May, by “introducing additional information in the token distribution at the point of generation by modulating the likelihood of tokens being generated.”

By comparing the model’s word choices along with its “adjusted probability scores” against the expected pattern of scores for watermarked and unwatermarked text, SynthID can detect whether an AI wrote that sentence.

Here’s how SynthID watermarks AI-generated content across modalities. ↓ pic.twitter.com/CVxgP3bnt2

— Google DeepMind (@GoogleDeepMind) October 23, 2024

This process does not impact the response’s accuracy, quality, or speed, according to a study published in Nature on Wednesday, nor can it be easily bypassed. Unlike standard metadata, which can be easily stripped and erased, SynthID’s watermark reportedly remains even if the content has been cropped, edited, or otherwise modified.

“Achieving reliable and imperceptible watermarking of AI-generated text is fundamentally challenging, especially in scenarios where [large language model] outputs are near deterministic, such as factual questions or code generation tasks,” Soheil Feizi, an associate professor at the University of Maryland, told MIT Technology Review, noting that its open-source nature “allows the community to test these detectors and evaluate their robustness in different settings, helping to better understand the limitations of these techniques.”

The system is not foolproof, however. While it is resistant to tampering, SynthID’s watermarks can be removed if the text is run through a language translation app or if it’s been heavily rewritten. It is also less effective with short passages of text and in determining whether a reply based on a factual statement was generated by AI. For example, there’s only one right answer to the prompt, “what is the capital of France?” and both humans and AI will tell you that it’s Paris.

If you’d like to try SynthID yourself, it can be downloaded from Hugging Face as part of Google’s updated Responsible GenAI Toolkit.

Andrew Tarantola
Former Digital Trends Contributor
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
Microsoft 365 Copilot gets an AI Researcher that everyone will love
Researcher agent in action inside Microsoft 365 Copilot app.

Microsoft is late to the party, but it is finally bringing a deep research tool of its own to the Microsoft 365 Copilot platform across the web, mobile, and desktop. Unlike competitors such as Google Gemini, Perplexity, or OpenAI’s ChatGPT, all of which use the Deep Research name, Microsoft is going with the Researcher agent branding.
The overarching idea, however, isn’t too different. You tell the Copilot AI to come up with thoroughly researched material on a certain topic or create an action plan, and it will oblige by producing a detailed document that would otherwise take hours of human research and compilation. It’s all about performing complex, multi-step research on your behalf as an autonomous AI agent.
Just to avoid any confusion early on, Microsoft 365 Copilot is essentially the rebranded version of the erstwhile Microsoft 365 (Office) app. It is different from the standalone Copilot app, which is more like a general purpose AI chatbot application.
Researcher: A reasoning agent in Microsoft 365 Copilot
How Researcher agent works?
Underneath the Researcher agent, however, is OpenAI’s Deep Research model. But this is not a simple rip-off. Instead, the feature’s implementation in Microsoft 365 Copilot runs far deeper than the competition. That’s primarily because it can look at your own material, or a business’ internal data, as well.
Instead of pulling information solely from the internet, the Researcher agent can also take a look at internal documents such as emails, chats, internal meeting logs, calendars, transcripts, and shared documents. It can also reference data from external sources such as Salesforce, as well as other custom agents that are in use at a company.
“Researcher’s intelligence to reason and connect the dots leads to magical moments,” claims Microsoft. Researcher agent can be configured by users to reference data from the web, local files, meeting recordings, emails, chats, and sales agent, on an individual basis — all of them, or just a select few.

Why it stands out?

Read more
Samsung might put AI smart glasses on the shelves this year
Google's AR smartglasses translation feature demonstrated.

Samsung’s Project Moohan XR headset has grabbed all the spotlights in the past few months, and rightfully so. It serves as the flagship launch vehicle for a reinvigorated Android XR platform, with plenty of hype from Google’s own quarters.
But it seems Samsung has even more ambitious plans in place and is reportedly experimenting with different form factors that go beyond the headset format. According to Korea-based ET News, the company is working on a pair of smart glasses and aims to launch them by the end of the ongoing year.
Currently in development under the codename “HAEAN” (machine-translated name), the smart glasses are reportedly in the final stages of locking the internal hardware and functional capabilities. The wearable device will reportedly come equipped with camera sensors, as well.

What to expect from Samsung’s smart glasses?
The Even G1 smart glasses have optional clip-on gradient shades. Photo by Tracey Truly / Digital Trends
The latest leak doesn’t dig into specifics about the internal hardware, but another report from Samsung’s home market sheds some light on the possibilities. As per Maeil Business Newspaper, the Samsung smart glasses will feature a 12-megapixel camera built atop a Sony IMX681 CMOS image sensor.
It is said to offer a dual-silicon architecture, similar to Apple’s Vision Pro headset. The main processor on Samsung’s smart glasses is touted to be Qualcomm’s Snapdragon AR1 platform, while the secondary processing hub is a chip supplied by NXP.
The onboard camera will open the doors for vision-based capabilities, such as scanning QR codes, gesture recognition, and facial identification. The smart glasses will reportedly tip the scales at 150 grams, while the battery size is claimed to be 155 mAh.

Read more
This AI app boosts my productivity in a way that Apple Intelligence can’t
A microphone in front of a MacBook on a desk.

Apple Intelligence offers a bunch of interesting features, but if you’ve tried most of them for more than a few minutes, you realise they’re not quite up to scratch compared to the best artificial intelligence (AI) tools.

Image Playground is fine, for instance, but not particularly useful. Writing Tools work well enough but aren’t as ground-breaking as Apple might make out. And the less that’s said about Siri, the better.

Read more