That Time We Tried To Make Sense of a Statistic in a New York Times Story on Deepfakes
Deepfakes aren't nearly as dangerous as the tried-and-true technique of saying something misleading with the imprimatur of authority.

About a third of the way into a New York Times piece about the prospects for regulating "deepfake" technology, a striking statement appears:
Some experts predict that as much as 90 percent of online content could be synthetically generated within a few years.
It's striking because it's frustrating: The sentence grabs your attention with a huge number (90 percent!), but it doesn't convey what the number means. For one thing, there's that weasel-phrase "as much as"—just how much wiggle room is that masking? And what counts as "online content"? The passages before and after this line are explicitly about deepfakes, but strictly speaking even bot spam is "synthetically generated." What precisely are we talking about here?
Fortunately, the sentence had a hyperlink in it. When I clicked, I found myself at a report from the European law enforcement agency Europol. And on page 5, it says this:
Experts estimate that as much as 90% of online content may be synthetically generated by 2026. Synthetic media refers to media generated or manipulated using artificial intelligence (AI). In most cases, synthetic media is generated for gaming, to improve services or to improve the quality of life, but the increase in synthetic media and improved technology has given rise to disinformation possibilities, including deepfakes.
That helps a bit: We're not talking about spam, but we're not just talking about deepfakes either. But it still doesn't really say how we're defining the denominator here (90 percent of what?), and we still have that hazy "as much as."
Fortunately, Europol also cites a source: Nina Schick's 2020 book Deepfakes: The Coming Infocalypse. And using Amazon's "look inside" feature to search for the number 90, I was able to find the original source of the figure:
To understand more, I spoke to Victor Riparbelli, the CEO and Founder of Synthesia, a London-based start-up that is generating synthetic media for commercial use….At Synthesia, he and his team are already generating synthetic media for their clients, who he says are mostly Fortune 500 companies. Victor explains that they are turning to AI-generated videos for corporate communications, because it gives them a lot more flexibility. For example, the technology enables Synthesia to produce a video in multiple languages at once. As the technology improves, Victor tells me, it will become ubiquitous. He believes that synthetic video may account for up to 90 per cent of all video content in as little as three to five years.
And so:
1. We're not talking about all online content. We're talking about all video content.
2. It wasn't "some experts" who predicted this. It was one expert. And he's a guy who runs a company that generates synthetic media, so he has every reason to hype it.
3. He offered his forecast in a book published in August 2020, so we might—depending on when exactly Schick's interview with him took place—already be in that three-to-five-year zone. Technically, I suppose his prediction has already come true: The amount of online video content that is synthetically generated is indeed somewhere between 0 and 90 percent.
Does any of this tell us anything about the coming impact of deepfakes? I suppose it supports the general idea that these technologies are getting better and becoming more common, but I already knew that before I read any of this. Even if that sentence in the Times had been more accurate, it wouldn't have illuminated much; it's basically there as a prop.
But it does tell us something that this story discussing the dangers that deepfakes purportedly pose to the information environment would itself include a little misinformation. This misinformation is being spread not by some spooky new tech, but through the tried-and-true method of saying something misleading with the imprimatur of authority. For the average reader, that would be the authority of The New York Times; for the Times reporter, it was the authority of Europol.
If you talk with people who track the spread of rumors online, you'll often hear that it doesn't take an avant-garde technology to convince people of something that isn't true—just mislabel a photo or misquote a speech, and your deception might take off among the people who are already inclined to believe what you're telling them. Deepfakes themselves don't worry me tremendously. Society has adjusted to new forms of visual fakery before, and it can surely do so again. Better to worry about something as old as humanity: our capacity to find reasons to believe whatever we want.
Editor's Note: As of February 29, 2024, commenting privileges on reason.com posts are limited to Reason Plus subscribers. Past commenters are grandfathered in for a temporary period. Subscribe here to preserve your ability to comment. Your Reason Plus subscription also gives you an ad-free version of reason.com, along with full access to the digital edition and archives of Reason magazine. We request that comments be civil and on-topic. We do not moderate or assume any responsibility for comments, which are owned by the readers who post them. Comments do not represent the views of reason.com or Reason Foundation. We reserve the right to delete any comment and ban commenters for any reason at any time. Comments may only be edited within 5 minutes of posting. Report abuses.
Please
to post comments
Some experts predict that as much as 90 percent of online content could be synthetically generated within a few years.
Some experts note that as much as 90 percent of modern journalism buttresses claims by pointing to unnamed experts.
It's not just modern. This shit was happening back when I was a kid. The only difference is that no one gives a shit about objective journalism anymore, so everyone just making shit up with no one bothering to hold them accountable because literally no one cares. As long as journalists go through the proper forms ("unnamed sources say") no one cares.
Stephen Glass was not an anomaly, he was the forerunner.
It’s almost like they are the enemy of the people…
I blame the editors for refusing to send bad stories back for a rewrite, or demanding sources, etc.
Journalism has alwya been a bit leftist in this country, and mostly liberal, but only recently have they completely turned their backs on the journalistic standards of integrity and objectivity.
Some blame Woodward and Bernstein for becoming rich and famous for exposing Nixon, leading journalism students thinking they can change the world. But Woodward and Bernstein had sources and they didn't make up shit and they weren't out to explicitly hand the presidency over to their party. Woodward and Bernstein were actual journalists, not this new crowd who aren't even qualified to put out a high school newsletter.
Generally, it’s not the liberal bias that I take issue with (though it can get really bad sometimes). It’s their complete farce of “speaking truth to power” while simultaneously sucking that power off and running cover for it.
The bias has become more of a problem over the last 20 years, because it has led them to ignoring one parties abuses as much as possible while casting every other party as the worst human beings imaginable.
But Woodward and Bernstein had sources and they didn’t make up shit and they weren’t out to explicitly hand the presidency over to their party.
How do we explain their current public support for the recent Coup D'Etat? That they've (whichever is still alive and giving interviews) gotten stupid with age and can't see it for what it was?
I blame the editors for requiring that their churnalists parrot the correct line, sources be damned.
Remember, the biased journalists that were corrected yesterday are still fuming they didn't get their way yet are still the editors today.
Well when we have the capability to translate all media into other languages, the 90% figure does not seem surprising. Nor sinister.
I write a book, some bots translate it into nine other languages. Now 90% of my books are synthetic. So what? Baring copyright issues, so what?
I write a book, some bots translate it into nine other languages. Now 90% of my books are synthetic.
Yeah, I’m not sure if I would call your translated book… or put it in the category of ‘synthetic media’. Edit: The translation is synthetic, but assuming the translation provides a reasonable amount of fidelity to the original text, it's still your book, but possibly a crap translation.
AI has become a little better (if we can even call it AI) at doing language translation. At least it’s readable now. But the nuances of really translating one language into another, reliably, using an algorithm is still elusive.
Actual interview with a translator who did subtitles for movies:
I really struggled with “how do you like them apples” because there’s no expression in the language I was translating it to that really made sense. Finally, I decided to remove all references to fruit altogether and essentially write, “Are you pleased with yourself now?”
Now imagine an AI trying to work this out.
Unsure if manually translated or other but, Squid Game: "It reminds me of the moon where I'm from."
You mean Earth?
I do think that translations are a part of that 90% that the original source was referring to. Doesn't mean it's happening now, but it was what he foresaw happening at some time.
Let some reporters "translate it" into reviews for some media companies and you'll be lucky if you don't have a dozen psychos looking to hunt you down - like the 6/14/17 jackoff.
2. It wasn't "some experts" who predicted this. It was one expert. And he's a guy who runs a company that generates synthetic media, so he has every reason to hype it.
This is the bread and butter of NYT style reporting.
Faster and simpler to just label everything in the NYT as bullshit.
But I was amazed to see something in Reason that not followed the data, and cited sources along the way to confirm it.
"not followed" should be "not only followed"
I'm still waiting for the NYT ChatGPT Editorial Page. Can't be worse then what we get now.
You're looking in the wrong place. The NYT ChatGPT controls most of the Business section.
The best way to protect against misinformation is to be slow and deliberate in action. You are most susceptible to screwing something up if you’re willing to take drastic action based on little information.
Think of the Observe-Orient-Decide-Act (OODA). You want to collect a lot of information from a lot of resources, and orient yourself in that picture, before making a decision and acting on it, such that false information doesn’t corrupt the action.
You know: basically things democracy and politicians suck at.
I remember years ago when the 9/11 Troofers were running the Ron Paul campaign. One of them said "He's going to announce [major defection and endorsement] tomorrow! It's happening!" And I said "Nope, not gonna happen". And it didn't happen.
Whereupon she asked, "How did you know?!?!"
Well first off I've been through enough "libertarian moments" to know it wasn't going to happen. But more importantly, there were not real sources for the rumor, just the rumor. The will to believe is NOT evidence. The credulous types do not understand this. Not just wackadoodles like Troofers, but all the credulous types.
Sure I can get fooled by Fake News, but I tend not too because I take everything with a grain of salt to begin with. I trust those sources that have shown they are trustworthy. And I instinctively distrust those with motivated biases, and those just drumming up clicks and views.
Its' not hard. Just don't instinctively believe everything you hear. And I don't mean from the "other side", I mean from your own side as well. It's so fucking easy to believe someone when you think they are on your side and looking about for you. Cluestick: they are not.