z-logo
open-access-imgOpen Access
Negative Consequences of Informing Voters about Deepfakes: Evidence from Two Survey Experiments
Author(s) -
John Ternovski,
Joshua Kalla,
Peter M. Aronow
Publication year - 2022
Publication title -
journal of online trust and safety
Language(s) - English
Resource type - Journals
ISSN - 2770-3142
DOI - 10.54501/jots.v1i2.28
Subject(s) - distrust , politics , fake news , public opinion , misinformation , social psychology , political science , psychology , public relations , internet privacy , computer science , law
 Advances in machine learning have made possible “deepfakes,” or realistic, computer-generated videos of public figures saying something they have not actually said. Policymakers have expressed concern that deepfakes could mislead voters, but prior research has found that such videos have minimal effects. There has nevertheless been extensive media coverage of the dangers of deepfakes, urging voters to be critical consumers of political videos. We explore whether these well-intentioned messages have an unintended consequence: if voters are warned about deepfakes, they may begin to distrust all political videos. We conducted two online survey experiments, and found that informing participants about deepfakes did not enhance participants’ ability to successfully spot manipulated videos but consistently induced them to believe the videos they watched were fake, even when they were real. Our findings suggest that even if deepfakes are not themselves persuasive, information about deepfakes can nevertheless be weaponized to dismiss real political videos.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here