Tech News

A Zelensky Deepfake Was Quickly Defeated. The Next One Might Not Be

Other conflicts and political leaders may be less fortunate, and could be more vulnerable to disruption by deepfakes, says Sam Gregory, who works on deepfakes policy at the nonprofit Witness.

Zelensky’s high profile helped Ukraine’s deepfake warning two weeks ago win international news coverage, and it also helped his quick response on Wednesday to spread rapidly. His prominence may also have prompted a quick response to the video from social networking companies. Meta spokesperson Aaron Simpson declined to say how it detected the video; so did YouTube’s Choi. The statement provided by Twitter’s Kennedy credited unspecified “external investigative reporting.”

Not all people targeted by deepfakes will be able to react as nimbly as Zelensky — or find their repudiation so widely trusted. “Ukraine was well positioned to do this,” Gregory says. “This is very different from other cases, where even a poorly made deepfake can create uncertainty about authenticity.”

Gregory points to a video that appeared in Myanmar last year, which appeared to show a former government minister held in detention saying he provided cash and gold to the country’s former leader Aung San Suu Kyi.

The military government that displaced Aung San Suu Kyi in a coup used that footage to accuse her of corruption. But in the video the former minister’s face and voice were distorted, causing many journalists and citizens to suggest the clip was faked.

Technical analysis has not resolved the mystery, in part because the video is of low quality, and because the former minister and others familiar with the truth do not speak as freely or to as large an audience as Zelensky could on Wednesday. While automatic deepfake detectors could someday help combat bad actors, they’re still a work in progress.

Deepfakes are still generally used more for titillation or harassment than grand deception, especially as they become easier to create. The Zelensky deepfake could represent a troubling new frontier. The quick and successful response to the clip highlights how, with a few tweaks and better timing, a deepfake attack could be an effective political weapon.

“If this was a more professional video and had been released early on in a more successful Russian advance on Kyiv, it could have created a lot of confusion,” says Samuel Bendett, who tracks Russian defense technology at the nonprofit CNA. As deepfake technology continues to get easier to access and more convincingZelensky is unlikely to be the last political leader targeted by fake video.


More Great WIRED Stories


Source link

Related Articles

Leave a Reply

Back to top button