You might have heard fears like this in the last few years: What if Donald Trump was woken up in the middle of the night and shown a fake video — indistinguishable from a real one — in which Kim Jong Un announced an imminent nuclear strike on the U.S.?
Today’s guest Nina Schick, author of Deepfakes: The Coming Infocalypse, thinks these concerns were the result of hysterical reporting, and that the barriers to entry in terms of making a very sophisticated ‘deepfake’ video today are a lot higher than people think.
But she also says that by the end of the decade, YouTubers will be able to produce the kind of content that's currently only accessible to Hollywood studios. So is it just a matter of time until we’ll be right to be terrified of this stuff?
Links to learn more, summary and full transcript.
Nina thinks the problem of misinformation and disinformation might be roughly as important as climate change, because as she says: “Everything exists within this information ecosystem, it encompasses everything.” We haven’t done enough research to properly weigh in on that ourselves, but Rob did present Nina with some early objections, such as:
• Won’t people quickly learn that audio and video can be faked, and so will only take them seriously if they come from a trusted source?
• If photoshop didn’t lead to total chaos, why should this be any different?
But the grim reality is that if you wrote “I believe that the world will end on April 6, 2022” and pasted it next to a photo of Albert Einstein — a lot of people would believe it was a genuine quote. And Nina thinks that flawless synthetic videos will represent a significant jump in our ability to deceive.
She also points out that the direct impact of fake videos is just one side of the issue. In a world where all media can be faked, everything can be denied.
Consider Trump’s infamous Access Hollywood tape. If that happened in 2020 instead of 2016, he would have almost certainly claimed it was fake — and that claim wouldn’t be obviously ridiculous. Malignant politicians everywhere could plausibly deny footage of them receiving a bribe, or ordering a massacre. What happens if in every criminal trial, a suspect caught on camera can just look at the jury and say “that video is fake”?
Nina says that undeniably, this technology is going to give bad actors a lot of scope for not having accountability for their actions.
As we try to inoculate people against being tricked by synthetic media, we risk corroding their trust in all authentic media too. And Nina asks: If you can't agree on any set of objective facts or norms on which to start your debate, how on earth do you even run a society?
Nina and Rob also talk about a bunch of other topics, including:
• The history of disinformation, and groups who sow disinformation professionally
• How deepfake pornography is used to attack and silence women activitists
• The key differences between how this technology interacts with liberal democracies vs. authoritarian regimes
• Whether we should make it illegal to make a deepfake of someone without their permission
• And the coolest positive uses of this technology
Producer: Keiran Harris.
Audio mastering: Ben Cordell.
Transcriptions: Sofia Davis-Fogel.
view more