Can watermarks stop AI deception?

Courtesy of Midjourney

Is it a real or AI-generated photo? Is it Drake’s voice or a computerized track? Was the essay written by a student or by ChatGPT? In the age of AI, provenance is paramount – a fancy way of saying we need to know where the media we consume comes from.

While generative AI promises to transform industries – from health care to entertainment to finance, just to name a few – it might also cast doubt on the origins of everything we see online. Experts have spent years warning that AI-generated media could disrupt elections and cause social unrest, so the stakes couldn’t be higher.

To counter this threat, lawmakers have proposed mandatory disclosures for political advertising using AI, and companies like Google and Meta, the parent company of Facebook and Instagram, are already requiring this. But bad actors won’t be deterred by demands for disclosures. So wouldn’t it be helpful if we had a way to instantly debunk and decipher what’s made by AI and what’s not?

Some experts say “watermarks” are the answer. A traditional watermark is a visible imprint — like what you see on a Getty image when you haven’t paid for it – or the inclusion of a corner logo. Today, these are used to deter theft rather than deception.

But most watermark proposals for AI-generated media center on invisible ones. These are functionally bits of code that tell third-party software that an image, picture, video, audio clip, or even lines of text were generated with AI. Using invisible watermarks would allow the audience to see art without it being visually altered or ruined — but, if there’s any confusion, in theory, the consumer of that media can run it through a computer program to see whether it was human-made or not.

Joe Biden’s administration is curious about watermarks. In his October executive order, the US president told the Commerce Department to “develop guidance for content authentication and watermarking to clearly label AI-generated content.” The goal: To protect Americans from “fraud and deception.”

It’s an effort many private companies are already working on — but solving the watermark issue has involved a lot of trial and error.

In August, Google released SynthID, a new method for embedding a watermark in the pixels of an image that’s perceptible to machine detectors but not the human eye. Still, it warns that SynthID isn’t “foolproof to extreme” methods of image manipulation. And last week, Meta announced it’s adding invisible watermarks to its text-to-image generator, promising that it’s “resilient to common image manipulations like cropping, color change (brightness, contrast, etc.), screen shots and more.”

There are more creative, cross-industry solutions too. In October, Adobe developed a special icon that can be added to an image’s metadata that both indicates who made it and how. Adobe told The Verge that it wants the icon to serve as a “nutrition label” for AI-generated images. But just like nutrition labels on food, the reality is no one can punish you for ignoring them.

And there are daunting challenges to actually making watermarks work.

Adam Conner, the tech policy lead at the Center for American Progress, said that watermarks need to transcend file format changes. “Even the best plans for watermarking will need to solve for the issue … where content is distributed as a normal file type, like a JPEG or MP3,” he said. In other words, the watermarks need to carry over from where they’re generated — say, an image downloaded on DALL-E — to wherever they are copied or converted into various file formats.

Meanwhile, researchers have poked holes in the latest and greatest watermarking tech. Researchers at Carnegie Mellon, for example, published a method for destroying watermarks by adding “noise” (basically, useless data) to an image and then reconstructing it. “All invisible watermarks are vulnerable to the proposed attack,” they wrote in July.

Others think that watermarking efforts might just be a fool’s errand. “I don’t believe watermarking the output of the generative models will be a practical solution,” University of Maryland computer science professor Soheil Feizi told The Verge. “This problem is theoretically impossible to be solved reliably.”

But there is clear political will to get watermarks working. Apart from Biden’s call, the G-7 nations are reportedly planning to ask private companies to develop watermarking technology so AI media is detectable. China banned AI-generated media without watermarks a year ago. Europe has pushed for AI watermarking, too, but it’s unclear if it’ll make it into the final text of its AI Act, the scope of which lawmakers agreed to last week.

The main limitation to achieving these goals is the elephant in the room: If Feizi is right, then watermarking AI will simply … miss the mark.

Please write in and tell us what you think – are watermarks on AI-generated images a good idea? Should they be legally required? Write to us here.

More from GZERO Media

Mourners react next to the body of a Palestinian killed in Israeli strikes, amid the ongoing conflict between Israel and Hamas, at Al-Aqsa hospital, in Deir Al-Balah, in the central Gaza Strip, May 12, 2024.
REUTERS/Ramadan Abed

The UN is now playing cleanup, maintaining that the overall death toll has not changed and is roughly 35,000.

Putin visits Xi to continue "no-limit" relationship with China | Ian Bremmer | World In :60

Does Putin's upcoming visit with Xi Jinping signal a continuing “no-limits” partnership between China and Russia? Why is Europe alarmed with Georgia's “foreign agents” law? How will Biden respond if Israel continues to push into Rafah? Ian Bremmer shares his insights on global politics this week on World In :60.

Saudi Crown Prince and Prime Minister Mohammed bin Salman meets with U.S. Secretary of State Antony Blinken at the Al Yamamah Palace in Riyadh, Saudi Arabia, April 29, 2024.
REUTERS/Evelyn Hockstein

Saudi Arabia is reportedly showing fresh interest in a roadmap to peace in Yemen that was iced late last year in the wake of the Oct. 7 attacks in Israel.

EDITORS' NOTE: Reuters and other foreign media are subject to Iranian restrictions on leaving the office to report, film or take pictures in Tehran. A security personnel looks on at oil docks at the port of Kalantari in the city of Chabahar, 300km (186 miles) east of the Strait of Hormuz January 17, 2012.
REUTERS/Raheb Homavandi

On Monday, India signed a 10-year-long agreement to operate and develop Iran’s Chabahar port.

FILE PHOTO: Russian President Vladimir Putin and Chinese President Xi Jinping walk during a meeting at the Kremlin in Moscow, Russia March 21, 2023.
Sputnik/Grigory Sysoyev/Kremlin via REUTERS

Russian President Vladimir Putin will be in Beijing on Thursday for talks with Chinese President Xi Jinping, in a rare overseas trip to publicly underline strong relations.

Happy young couple hide behind paper hearts to kiss.
IMAGO/Pond5 Images via Reuters

ChatGPT is a prude. Try to engage with it about sex or other risqué topics, and it’ll turn you down. The OpenAI chatbot’s usage rules specify that even developers who build on the platform must be careful to design their applications so they’re age-appropriate for children, meaning no “sexually explicit or suggestive content,” except for scientific and educational purposes. But the company is reportedly now looking into its blue side.