Skip to main content

Sign up for our newsletter.

Quality journalism. Progressive values. Direct to your inbox.

screenshot of report cover

New York University published an report Tuesday entitled Disinformation and the 2020 Election: How the Social Media Should Prepare. (Image: screenshot of report)

New Report Warns of Deepfakes and Social Media Disinformation Campaigns That Aim to Undermine 2020 Election

The NYU expert who authored the report says social media companies "have to take responsibility for the way their sites are misused."

Jessica Corbett

A report published Tuesday by New York University warns that fake videos and other misleading or false information could be deployed by domestic and foreign sources in efforts influence the U.S. 2020 presidential election campaign and details strategies to combat such disinformation.

"We urge the companies to prioritize false content related to democratic institutions, starting with elections."
—Paul M. Barrett, report author

The report—entitled Disinformation and the 2020 Election: How the Social Media Should Prepare—predicts that for next year's election, so-called "deepfake" videos will be unleashed across the media landscape "to portray candidates saying and doing things they never said or did" and, as a result, "unwitting Americans could be manipulated into participating in real-world rallies and protests."

Deepfakes, as NPR reported Monday, are "computer-created artificial videos or other digital material in which images are combined to create new footage that depicts events that never actually happened." Manipulated videos like those of Democratic House Speaker Nancy Pelosi (Calif.) that spread virally online earlier this year—often called shallowfakes or cheapfakes—also pose a threat to democratic elections, the report says.

In terms of delivery of disinformation, the NYU report spotlights the messaging service WhatsApp and the video-sharing social media network Instagram—which are both owned by Facebook. A report commissioned by the Senate Intelligence Committee in the wake of the 2016 election accused Russia of "taking a page out of the U.S. voter suppression playbook" by using social media platforms including Facebook and Instagram to target African-American audiences to try to influence their opinions on the candidates in that race.

The NYU report predicts that governments such as Russia, China, and Iran may work to disseminate lies in attempts to sway public opinions regarding the next race for the White House, but "domestic disinformation will prove more prevalent than false content from foreign sources." Digital voter suppression, it warns, could "again be one of the main goals of partisan disinformation."

To combat disinformation from all sources, the NYU report offers nine recommendations for major social media companies:

  • Detect and remove deepfake videos: Realistic but fraudulent videos have the potential to undermine political candidates and exacerbate voter cynicism.
  • Remove provably false content in general: The platforms already remove hate speech, voter suppression, and other categories of content; we recommend that they add one more.
  • Hire a senior content overseer: Each company needs an executive with clout to supervise the process of guarding against disinformation.
  • Attend to Instagram's disinformation problem: The photo-posting platform needs the concerted attention of its parent, Facebook.
  • Limit the reach of WhatsApp: Users should be restricted to forwarding content to one chat group at a time.
  • Defend against for-profit disinformation: The companies must prepare for false content generated by hired-gun firms.
  • Back legislation regulating political ads and punishing voter suppression: Narrowly tailored bills pending on Congress could help curb some forms of disruption.
  • Improve industry-wide collaboration on disinformation: For example, when one platform takes down abusive accounts, others should do he same with affiliated accounts.
  • Teach social media literacy in a more direct, sustained way: Users have to take responsibility for recognizing false content, but they need more help to do it.

Paul M. Barrett, the report's author and deputy director of the NYU Stern Center for Business and Human Rights, told The Washington Post that social media companies "have to take responsibility for the way their sites are misused."

"We urge the companies to prioritize false content related to democratic institutions, starting with elections," he said. "And we suggest that they retain clearly marked copies of removed material in a publicly accessible, searchable archive, where false content can be studied by scholars and others, but not shared or retweeted."

While the removal of disinformation by social media giants is touted as a positive strategy by Barrett and others, such calls have sparked censorship concerns, especially as online platforms such as Facebook and YouTube have recently blocked content or shut down accounts that spread accurate information.

Michael Posner, director of NYU's Stern Center, said in a statement to The Hill that "taking steps to combat disinformation isn't just the right thing to do, it's in the social media companies' best interests as well." As he put it, "Fighting disinformation ultimately can help restore their damaged brand reputations and slows demands for governmental content regulation, which creates problems relating to free speech."

One example is what critics of the Trump administration have dubbed the 'Censor the Internet' executive order, which would give federal agencies certain powers to decide what internet material is acceptable. After a draft of that order leaked, as Common Dreams reported last month, "free speech and online advocacy groups raised alarm about the troubling and far-reaching implications of the Trump plan if it was put into effect by executive decree."

A Bloomberg report from Saturday which revealed that "fake news and social media posts are such a threat to U.S. security that the Defense Department is launching a project to repel 'large-scale, automated disinformation attacks'" with custom software sparked additional concerns about potential consequences of U.S. government actions to combat disinformation.

The NYU report does highlight the potential for legislation—particularly the Honest Ads Act, a bipartisan bill reintroduced earlier this year that aims to improve transparency around who is paying for political ads. However, Barrett told The Hill that he doesn't believe the bill "has much of a chance" of passing because of some Senate Republicans' positions on election security legislation.

"Congress would be making a huge contribution if there were hearings, particularly if there were bipartisan hearings... that educate people as to where we've been and what's likely to come," Barrett added. "We need more digital literacy, and Congress could use its position to provide that."

The NYU report adds to mounting concerns among tech experts, politicians, and voters about how disinformation could sway the 2020 election. John Feffer, director of Foreign Policy In Focus at the Institute for Policy Studies, wrote in June about the potential impact of deepfakes on next year's race:

Forget about October surprises. In this age of rapid dissemination of information, the most effective surprises happen in November, just before Election Day. In 2020, the election will take place on November 3. The video drops on November 2. The damage is done before damage control can even begin.

Feffer added that artificial intelligence (AI) systems which are "designed to root out such deepfake videos can't keep up with the evil geniuses that are employing other AI programs to produce them."

"It's an arms race," he wrote. "And the bad guys are winning."

Our work is licensed under Creative Commons (CC BY-NC-ND 3.0). Feel free to republish and share widely.

This is the world we live in. This is the world we cover.

Because of people like you, another world is possible. There are many battles to be won, but we will battle them together—all of us. Common Dreams is not your normal news site. We don't survive on clicks. We don't want advertising dollars. We want the world to be a better place. But we can't do it alone. It doesn't work that way. We need you. If you can help today—because every gift of every size matters—please do. Without Your Support We Simply Don't Exist.

'Political Malpractice': House Democrats' Bill Wouldn't Add Dental to Medicare Until 2028

"I don't want to see it drawn out to as far as the House has proposed," Sen. Bernie Sanders said during a recent press call.

Jake Johnson ·

'How Many More Deaths Must It Take?' Barbados Leader Rips Rich Nations in Fierce UN Speech

"How many more variants of Covid-19 must arrive, how many more, before a worldwide plan for vaccinations will be implemented?"

Jake Johnson ·

To Avert Debt Ceiling Calamity, Democrats Urged to Finally Kill the Filibuster

"The solution is to blow up the filibuster at least for debt limit votes, just as Mitch blew it up to pack the Supreme Court for his big donors."

Jake Johnson ·

Biden Decries 'Outrageous' Treatment of Haitians at Border—But Keeps Deporting Them

"I'm glad to see President Biden speak out about the mistreatment of Haitian asylum-seekers. But his administration's use of Title 42 to deny them the right to make an asylum claim is a much bigger issue."

Jessica Corbett ·

Global Peace Activists Warn of Dangers of US-Led Anti-China Pacts

"No to military alliances and preparation for catastrophic wars," anti-war campaigners from over a dozen nations write in a letter decrying the new AUKUS agreement. "Yes to peace, disarmament, justice, and the climate."

Brett Wilkins ·

Support our work.

We are independent, non-profit, advertising-free and 100% reader supported.

Subscribe to our newsletter.

Quality journalism. Progressive values.
Direct to your inbox.

Subscribe to our Newsletter.

Common Dreams, Inc. Founded 1997. Registered 501(c3) Non-Profit | Privacy Policy
Common Dreams Logo