Abstract
Social media platforms often assume that users can self-correct against misinformation. However, social media users are not equally susceptible to all misinformation as their biases influence what types of misinformation might thrive and who might be at risk. We call “diverse misinformation” the complex relationships between human biases and demographics represented in misinformation. To investigate how users’ biases impact their susceptibility and their ability to correct each other, we analyze classification of deepfakes as a type of diverse misinformation. We chose deepfakes as a case study for three reasons: (1) their classification as misinformation is more objective; (2) we can control the demographics of the personas presented; (3) deepfakes are a real-world concern with associated harms that must be better understood. Our paper presents an observational survey (N = 2016) where participants are exposed to videos and asked questions about their attributes, not knowing some might be deepfakes. Our analysis investigates the extent to which different users are duped and which perceived demographics of deepfake personas tend to mislead. We find that accuracy varies by demographics, and participants are generally better at classifying videos that match them. We extrapolate from these results to understand the potential population-level impacts of these biases using a mathematical model of the interplay between diverse misinformation and crowd correction. Our model suggests that diverse contacts might provide “herd correction” where friends can protect each other. Altogether, human biases and the attributes of misinformation matter greatly, but having a diverse social group may help reduce susceptibility to misinformation.
Similar content being viewed by others
Introduction
There is a growing body of scholarly work focused on distributed harm in online social networks. From leaky data1, and group security and privacy2 to hate speech3, misinformation4 and detection of computer-generated content5. Social media users are not all equally susceptible to these harmful forms of content. Our level of vulnerability depends on our own biases. We define “diverse misinformation” as the complex relationships between human biases and demographics represented in misinformation. This paper explores deepfakes as a case study of misinformation to investigate how U.S. social media users’ biases influence their susceptibility to misinformation and their ability to correct each other. We choose deepfakes as a critical example of the possible impacts of diverse misinformation for three reasons: (1) their status of being misinformation is binary; they either are a deepfake or not; (2) the perceived demographic attributes of the persona presented in the videos can be characterized by participants; (3) deepfakes are a current real-world concern with associated negative impacts that need to be better understood. Together, this allows us to use deepfakes as a critical case study of diverse misinformation to understand the role individual biases play in disseminating misinformation at scale on social networks and in shaping a population’s ability to self-correct.
We present an empirical survey (N = 2016 using a Qualtrics survey panel6) observing what attributes correspond to U.S.-based participants’ ability to detect deepfake videos. Survey participants entered the study under the pretense that they would judge the communication styles of video clips. Our observational study is careful not to prime participants at the time of their viewing video clips so we could gauge their ability to view and judge deepfakes when they were not expecting them (not explicitly knowing if a video is fake or not is meant to emulate what they would experience in an online social media platform). Our survey also investigates the relationship between human participants’ demographics and their perception of the video person(a)’s features and, ultimately, how this relationship may impact the participant’s ability to detect deepfake content.
Our objective is to evaluate the relationship between classification accuracy and the demographic features of deepfake videos and survey participants. Further analysis of other surveyed attributes will be explored in future work. We also recognize that data used to train models that create deepfakes may introduce algorithmic biases in the quality of the videos themselves, which could introduce additional biases in the participant’s ability to guess if the video is a deepfake or not. The Facebook Deepfake Detection Challenge dataset that was used to create the videos we use in our survey was created to be balanced in diversity in several axes (gender, skin-tone, age). We suspect that if there are algorithmic-level biases in the model used resulting in better deepfakes for personas of specific demographics, we would expect to see poorer accuracy across the board for all viewer types when classifying these videos. We do see that viewer groups’ accuracy differs based on different deepfake video groups. However, our focus is on the perception of survey participants towards deepfakes’ identity and demographics to capture viewer bias based on their perception rather than the model’s bias and classification of the video persona’s racial, age, and gender identity. Our goal is to focus on viewers and capture what a viewer would experience in the wild (on a social media platform), where a user would be guessing the identity features of the deepfake and then interrogating if the video was real or not with little to no priming.
This paper adopts a multidisciplinary approach to answer these questions and understand their possible impacts. First, we use a survey analysis to explore individual biases related to deepfake detection. There is abundant research suggesting the demographics of observers and observed parties influence the observer’s judgment and sometimes actions toward the observed party7,8,9,10,11. In an effort to avoid assumptions about any demographic group, we chose four specific biases to analyze vis-à-vis deepfakes: (Question 1) Priming bias: How much does classification accuracy depend on participants being primed about the potential of a video being fake? Our participants are not primed on the meaning of deepfakes and are not told to be explicitly looking for them prior to beginning the survey. Importantly, we do not explicitly vary the priming of our participants but we compare their accuracy to a previous study with a similar design but primed participants5. Participants are debriefed after the completion of the survey questions and then asked to guess the deepfake status of the videos they watched. More information about our survey methodology and why the study was formulated as a deceptive survey can be seen in section 4.4. (Question 2) Prior knowledge: Does accuracy depend on how often the viewer uses social media and whether they have previously heard of deepfakes? Here, we ask participants to evaluate their own knowledge and use their personal assessment to answer this research question. (Question 3) Homophily bias: Are humans better classifiers of video content if the perceived demographic of the video persona matches their own identity? (Question 4) Heterophily bias: Inversely, are humans more accurate if the perceived demographic of the video persona does not match their own? We then use results from the survey to develop an idealized mathematical model to theoretically explore population-level dynamics of diverse misinformation on online social networks. Altogether, this allows us to hypothesize the mechanisms and possible impacts of diverse misinformation, as illustrated in Fig. 1.
Our paper is structured as follows. We outline the harms and ethical concerns of diverse misinformation and deepfakes in “Introduction.” We explore the possible effects through which demographics impact susceptibility to diverse misinformation through our observational study in “Results.” We then investigate the network-level dynamics of diverse misinformation using a mathematical model in “Mathematical Model.” We discuss our findings and their implications in “Discussion.” Our full survey methodology can be seen in “Methods.”
It is important to understand human biases as they impact the transmission and correction of misinformation and their potential impacts on polarization and degradation of the epistemic environment12. In social networks, it has been shown that there are human tendencies toward homophily bias13,14. Indeed, there are differences in user demographic groups’ abilities to detect deepfakes and misinformation (e.g., age)15. Previous work has also shown that biases impact people’s accuracy as an eyewitness through the own-race bias (ORB) phenomenon16,17,18. It is an open question whether deepfake detection also demonstrates the own-race bias (ORB) phenomenon.
Subsequently, these biases impact how social ties are formed and, ultimately, the shape of the social network. For example, in online social networks, homophily often manifests through triadic closures19 where friends in social networks tend to form new connections that close triangles or triads. Understanding individuals’ and groups’ biases will help understand the network’s structure and dynamics and how information and misinformation spread on the network depending on its level of diversity. For example, depending on the biases and the node-specific diversity of the connections it forms, one may have a system that may be more or less susceptible to widespread dissemination as it would in a Mixed Membership Stochastic Block Model (MMSBM)20. A Mixed Membership Stochastic Block Model is a Bayesian community detection method that segments communities into blocks but allows community members to mix with other communities. Assumptions in an MMSBM include a list of probabilities that determine the likelihood of communities interacting. We explore these topics in more detail in “Mathematical Model.”
Previous work has demonstrated that homophily bias towards content aligned with one’s political affiliation can impact one’s ability to detect misinformation21,22. Traberg et al. show that political affiliation can impact a person’s ability to detect misinformation about political content21. They found that viewers misclassified misinformation as being true more often when the source of information aligned with their political affiliation. Political homophily bias, in this case, made them feel as though the source was more credible than it was.
In this paper, we investigate the accuracy of deepfake detection based on multiple homophily biases in age, gender, and race. We also explore other bias types, such as heterophily bias, priming, and prior knowledge bias impacting deepfake detection.
Misinformation is information that imitates real information but does not reflect the genuine truth23. Misinformation has become a widespread societal issue that has drawn considerable recent attention. It circulates physically and virtually on social media sites24 and interacts with socio-semantic assortativity. In contrast, assortative social clusters will also tend to be semantically homogeneous25. For instance, misinformation promoting political ideology might spread more easily in social clusters based on shared demographics, further exacerbating political polarization and potentially influencing electoral outcomes26. This has sparked concerns about the weaponization of manipulated videos for malicious ends, especially in the political realm26. Those with higher political interests are more likely to share deepfakes inadvertently, and those with lower cognitive ability are also more likely to share deepfakes inadvertently. The relationship between political interest and deepfakes sharing is moderated by network size27.
Motivations vary broadly to explain why people disseminate misinformation, which we refer to as disinformation when specifically intended to deceive. Motivations include (1) purposefully trying to deceive people by seeding distrust in information, (2) believing the information to be accurate and spreading it mistakenly, and (3) spreading misinformation for monetary gain. In this paper, we will primarily focus on deepfakes as misinformation meaning the potential of a deepfake viewer getting duped and sharing a deepfake video. Disinformation is spreading misinformation with the intent to deceive. In this paper, we do not assume that all deepfakes are disinformation since we do not consider the intent of the creator. A deepfake could be made to entertain or showcase technology. We instead focus on deepfakes as misinformation meaning the potential of a deepfake viewer getting duped and sharing a deepfake video, regardless of intent.
There are many contexts where online misinformation is of concern. Examples include misinformation around political elections and announcements (political harms)28; such deepfake videos can, in theory, alter political figures to say just about anything, raising a series of political and civic concerns28; misinformation on vaccinations during global pandemics (health-related harms)29,30; false speculation to disrupt economies or speculative markets31; distrust in news media and journalism (harms to news media)4,32. People are more likely to feel uncertain than to be misled by deepfakes, but this resulting uncertainty, in turn, reduces trust in news on social media33; false information in critical informational periods such as humanitarian or environmental crises34; and propagation of hate speech online3 which spreads harmful false content and stereotypes about groups (harms related to hate speech).
Correction of misinformation: There are currently many ways to try to detect and mitigate the harms of misinformation online35. On one end of the spectrum are automated detection techniques that focus on the classification of content or on observing anomaly detection in the network structure context of the information or propagation patterns36,37. Conversely, crowd-sourced correction of misinformation leverages other users to reach a consensus or simply estimate the veracity of the content38,39,40. We will look at the latter form of correction in an online social network to investigate the role group correction plays in slowing the dissemination of diverse misinformation at scale.
Connection with deepfakes: The potential harms of misinformation can be amplified by computer-generated videos used to give fake authority to the information. Imagine, for instance, harmful messages about an epidemic conveyed through the computer-generated persona of a public health official. Unfortunately, deepfake detection remains a challenging problem, and the state-of-the-art techniques currently involve human judgment5.
Deepfakes are artificial images or videos in which the persona in the video is generated synthetically. Deepfakes can be seen as false depictions of a person(a) that mimics a person(a) but does not reflect the truth. Deepfakes should not be confused with augmented or distorted video content, such as using color filters or digitally-added stickers in a video. Creating a deepfake can involve complex methods such as training artificial neural networks known as generative adversarial networks (GANs) on existing media41 or simpler techniques such as face mapping. Deepfakes are deceptive tools that have gained attention in recent media for their use of celebrity images and their ability to spread misinformation across online social media platforms42.
Early deepfakes were easily detectable with the naked eye due to their uncanny visual attributes and movement43. However, research and technological developments have improved deepfakes, making them more challenging to detect4. There are currently several automated deepfake detection methods44,45,46,47,48. However, they are computationally expensive to deploy at scale. As deepfakes become ubiquitous, it will be necessary for the general audience to identify deepfakes independently during gaps between the development of automated techniques or in environments that are not always monitored by automated detection (or are offline). It will also be important to allow human-aided and human-informed deepfake detection in concert with automated detection techniques.
Several issues currently hinder automated methods: (1) they are computationally expensive; (2) there may be bias in deepfake detection software and training data—credibility assessments, particularly in video content, have been shown to be biased49; (3) As we have seen with many cybersecurity issues, there is a “cat-and-mouse” evolution that will leave gaps in detection methodology50.
Humans may be able to help fill these detection gaps. However, we wonder to what extent human biases impact the efficacy of detecting diverse misinformation. If human-aided deepfake detection becomes a reliable strategy, we need to understand the biases that come with it and what they look like on a large scale and on a network structure. We also posit that insights into human credibility assessments of deepfakes could help develop more lightweight and less computationally expensive automated techniques.
As deepfakes improve in quality, the harms of deepfake videos are coming to light51. Deepfakes raise several ethical considerations: (1) the evidentiary power of video content in legal frameworks4,52,53; (2) consent and attribution of the individual(s) depicted in deepfake videos54; (3) bias in deepfake detection software and training data49; (4) degradation of our epistemic environment, i.e., there is a large-scale disagreement between what community members believe to be real or fake, including an increase in misinformation and distrust4,32; and (5) possible intrinsic wrongs of deepfakes55.
It is important to understand who gets duped by these videos and how this impacts people’s interaction with any video content. The gap between convincing deepfakes and reliable detection methods could pose harm to democracy, national security, privacy, and legal frameworks4. Consequently, additional regulatory and legal frameworks56 will need to be adopted to protect citizens from harms associated with deepfakes and uphold the evidentiary power of visual content. False light is a recognized invasion of privacy tort that acknowledges the harms that come when a person has untrue or misleading claims made about them. We suspect that future legal protections against deepfakes might well be grounded in such torts, though establishing these legal protections is not trivial52,57.
The ethical implications of deepfake videos can be separated into two main categories: the impacts on our epistemic environment and people’s moral relationships and obligations with others and themselves. Consider the epistemic environment, which includes our capacity to take certain representations of the world as true and our taking beliefs and inferences to be appropriately justified. Audio and video are particularly robust and evocative representations of the world. They have long been viewed as possessing more testimonial authority (in the broader, philosophical sense of the phrase) than other representations of the world. This is true in criminal and civil contexts in the United States, where the admissibility of video recordings as evidence in federal trials is specifically singled out in Article X of the Federal Rules of Evidence58 (State courts have their own rules of evidence, but most states similarly have explicit rules that govern the admissibility of video recordings as evidence). The wide adoption of deepfake technology would strain these rules of evidence; for example, the federal rules of evidence reference examples of handwriting authentication, telephone conversation authentication, and voice authentication but do not explicitly mention video authentication. Furthermore, laws are notorious for lagging behind technological advances59, which can further complicate and limit how judges and juries can approach the existence of a deepfake video as part of a criminal or civil case.
Our paper asks four primary research questions regarding how human biases impact deepfake detection. (Q1) Priming: How important is it for an observer to know that a video might be fake? (Q2) Prior knowledge: How important is it for an observer to know about deepfakes, and how does social media usage affect accuracy? (Q3−Q4) Homophily and heterophily biases: Are participants more accurate at classifying videos whose persona they perceive to match (homophily) or mismatch (heterophily) their own demographic attributes in age, gender, and race?
To address our four research questions, we designed an IRB-approved survey (N = 2016) using video clips from the Deepfake Detection Challenge (DFDC) Preview Dataset60,61. Our survey participants entered the study under the pretense that they would judge the communication styles of video clips (they were not explicitly looking for deepfake videos in order to emulate the uncertainty they would experience in an online social network). After the consent process, survey participants were asked to watch two 10-second video clips. After each video, our questionnaire asked participants to rate the pleasantness of particular features (e.g., tone, gaze, likability, content) of the video on a 5-point Likert scale. They were also asked to state their perception of the person in the video by guessing the video persona’s gender identity, age, and whether they were white or a person of color.
After viewing both videos and completing the related questionnaire, the participants were then debriefed on the deception of the survey, given an overview of what deepfakes are, and then asked if they thought the videos they just watched were real or fake. After the debrief questions, we collected information on the participants’ backgrounds, demographics, and expressions of identity.
Our project investigates features or pairings of features (of the viewer or the person(a) in the video) that are the most important ones needed to determine an observer’s ability to detect deepfake videos and avoid being duped. Conversely, we also ask what pairings of features (of the viewer or the person(a) in the video) are important to determine an observer’s likelihood of being duped by a deepfake video.
Our null hypothesis asserts that none of the features or pairing of features we measure in our survey produce biases that show strong evidence of the importance of a user being duped by a deepfake video or being able to detect a deepfake video. We then measure our confidence in rejecting this null hypothesis by measuring a bootstrap credibility interval for a difference in means test between the accuracy of two populations (comparing Matthew’s Correlation Coefficient scores). In all tests, we use 10,000 bootstrap samples and consider a comparison significant (having strong evidence) if the difference is observed in 95% of samples (i.e., in 9500 pairs). With this method, our paper aims to better understand how potential social biases affect our ability to detect misinformation.
Results
Our results can be summarized as follows. (Q1) If not primed, our survey participants are not particularly accurate at detecting deepfakes (accuracy = 51%, essentially a coin toss). (Q3−Q4) Accuracy varies by some participants’ demographics and perceived demographics of video persona. In general, participants were better at classifying videos that they perceived as matching their own demographic.
Our results show that of the 4032 total videos watched, 49% were deepfakes, and 1429 of those successfully duped our survey participants. A confusion matrix showing the True Positive (TP), False Negative (FN), False Positive (FP), and True Negative (TN) rates can be seen in Fig. 2. We also note that the overall accuracy rate (where accuracy = (TP+TN)/(TP+FP+FN+TN)) of our participants was 51%. This translates to an overall Matthew’s Correlation Coefficient (MCC) score of 0.334 for all participant’s guesses vs. actual states of the videos. MCC62,63 is a simple binary correlation between the ground truth and the participant’s guess. Regardless of the metric, our participants performed barely better than a simple coin flip (credibility 94%). All summary statistics for our study and all confusion matrices for our primary and secondary demographic groups can be found in Appendix SI2 and Appendix SI3, respectively. Next, we explain our findings in detail.
Q1 Priming bias: Our results suggest that priming bias may play a role in a user’s ability to detect deepfakes. Compared with notable prior works5,64,65, our users were not explicitly told to look for deepfake videos while viewing the video content. Our survey takers participated in a deceptive study where they thought they answered questions about effective communication styles. They were debriefed only after the survey was completed and then asked if they thought the video clips were real or fake. Priming, on the contrary, would mean that when the user watched the two video clips, they would be explicitly looking for deepfakes.
Other works measured primed human deepfake detectors to compare them to machines and humans with machine aid. For example, in a study by ref. 64, humans were deployed as deepfake evaluators. The participants were explicitly asked to view images and look for fake images. Participants in the study were also required to pass a qualification test where they needed to correctly classify 65% real and fake images to participate in the study64 fully. In a more recent study by ref. 5, participants viewed video clips from the Facebook Deepfake Detection Challenge Dataset (DFCD), as in our study. They were asked to explicitly look for deepfake videos and then tested regarding how this compared to machines alone and machines aided by humans. Groh et al. reported an accuracy score of 66% for primed humans, 73% for a primed human with a machine helper, and 65% for the machine alone. In another study, ref. 65 also showed that hybrid systems that combine crowd-nominated and machine-extracted features outperform humans and machines alone.
In comparison, a previous study by ref. 5 uses the same benchmark video data but in their study subjects were informed beforehand and explicitly looked for deepfakes. We compare our participant’s accuracy to this study in Table 1. The section of the Groh et al. study where they gather human accuracy of deepfakes was conducted through a publicly available website (participant demographics were not gathered for this study). This website collected organic visitors from all over the world, the participants could view deepfakes from the DFCD dataset and guess if they could spot the deepfake or not (the specific question asked “Can you spot the deepfake video?”), the study participants were asked on a slider how confident they were in their answers as a percentage between 50% and 100%. In the human detection section of the study, they evaluated the accuracy of 882 individuals (only those who viewed at least 10 pairs of videos, note they did not find evidence that accuracy improves as the participants watch more videos) on 56 pairs of videos from the DFCD dataset. They compare the accuracy rate of participants (66% for humans alone) in this study with the accuracy of the leading model from the DFDC Kaggle challenge (65% accuracy for the leading model). In the second part of their experiment, they look at how the leading model (e.g., machine model) can help human accuracy. In this part of the study, after participants (N = 9492) submit their guesses regarding the state of the videos they are given the likelihood from the machine model and then told they can update their scores (resulting in a 73% accuracy score).
Our results show that the non-primed participants were only 51% accurate at detecting if a video was real or fake. One important takeaway from previous studies is that human-machine cooperation provides the best accuracy scores. The previously mentioned prior studies were performed with primed participants. We believe a more realistic reflection of how deepfake encounters would occur “in the wild” would be with observers who were not explicitly seeking out deepfakes. Ecological viewing conditions are important for this type of study66. Future work is needed to investigate how non-primed human deepfake detectors perform when aided by machines.
Q2 Prior knowledge effect: We also ask if participants are better at detecting a deepfake if they have prior knowledge about deepfakes or more exposure to social media.
Our results show that there was only weak evidence that prior knowledge or frequent social media usage impacts participants’ accuracy. Therefore, we cannot draw any strong conclusions as to the compatibility with our data for this particular question, given that our credibility score for this metric fell below 95% credibility.
We see that participants who are frequent social media users (i.e., use social media once a week or more) had a higher MCC score (MCC = 0.0396) than those who used social media less frequently (MCC = −0.0110). Participants who knew what a deepfake was before taking the survey (MCC = 0.0790) also had a higher score than those unfamiliar with deepfakes (MCC = 0.0175). However, in both comparisons, the difference was only deemed to have a weak effect given that bootstrap samples reject the null only with 83% and 94% credibility, respectively.
Q3-4 Homophily versus heterophily bias: We then focus on the potential impacts of heterophily and homophily biases on a participant’s ability to detect if a video is real or a deepfake. We look at the Matthew’s Correlation Coefficients (MCC) for all user groups and compare their guesses on videos that either match their identity (homophily) or do not match their own identity (heterophily). Results of these MCC scores related to homophily and heterophily bias can be seen in Fig. 3.
Our data shows strong evidence that one of our demographic subgroups, namely white participants, was more accurate when guessing the state of video personas that match their own demographic. We test our null hypothesis by comparing the answers given by a certain demographic of participants when looking at videos that match and do not match their identity. In doing so, we only observed evidence of a strong homophily bias for white participants, which can be seen in Table 2. In that case, the null hypothesis that they are equally accurate on videos of white personas and personas of color falls outside of a 99% credibility interval, which can be seen in Fig. 3.
We further break down this potential bias in two dimensions (overall demographic classes of the participants and video persona) in Table 2. We then see more evident results. Here we compare subgroups of our survey participants (e.g., male vs. female viewers, persons of color vs. white viewers, and young vs. old viewers) to see which groups perform better when watching videos of a specific sub-type (e.g., videos of men, videos of women, videos of persons of color, videos of white people, videos of young people, and videos of old people).
By gender, we find evidence that male participants are more accurate than female participants when watching videos with a male persona. Similarly, by race, we find strong evidence that participants of color are more accurate than white participants when watching videos that feature a persona who is a person of color. Lastly, young participants have the highest accuracy score overall for any of our demographic subgroups. Of course, these results may be confounded with other factors, such as social media usage, which can be more prominent in one group (e.g., young participants) than another (e.g., older participants). More work needs to be done to understand the mechanisms behind our results.
In summary, results that satisfy a threshold of credibility above 95% (rejecting the null hypothesis with 95% credibility) on human biases in deepfake detection are as follows.
-
We find strong evidence that white participants show a homophily bias, meaning they are more accurate at classifying videos of white personas than they are at classifying videos of personas of color.
-
We find strong evidence that when viewing videos of male personas, male participants in our survey are more accurate than female participants.
-
We find strong evidence that when viewing videos of personas of color, participants of color are more accurate than white participants.
-
We find strong evidence that when viewing videos of young personas, participants between the ages of 18−29 are more accurate than participants above the age of 30; surprisingly, participants aged 18−29 are also more accurate than participants aged 30−49 even when viewing videos of personas aged 30−49.
Mathematical model
In essence, the results shown in Table 2 illustrate how there is no single demographic class of participants that excels at classifying all demographics of video persona. Different participants can have different weaknesses. For example, a white male participant may be more accurate at classifying white personas than a female participant of color, but the female participant of color may be more accurate on videos of personas of colors. To consider the implications of this simple result, we take inspiration from our findings and formulate an idealized mathematical model of misinformation to better understand how deepfakes spread on social networks with diverse users and misinformation.
Models of misinformation spread often draw from epidemiological models of infectious diseases. This approach tracks how an item of fake news or a deepfake might spread, like a virus, from one individual to its susceptible network neighbors, duping them such that they can further spread misinformation67,68,69,70,71,72,73,74. However, unlike infectious diseases, an individual’s recovery does not occur on its own through its immune system. Instead, duped individuals require fact-checking or correction from their susceptible neighbors to return to their susceptible state75,76,77,78,79,80,81,82. In light of these previous modeling studies, it is clear that demographics can affect who gets duped by misinformation and who remains to correct their network neighbors. We therefore integrate these mechanisms with the core finding of our study: Not all classes of individuals are equally susceptible to misinformation.
Our model uses a network with a heterogeneous degree distribution and a structure inspired by the mixed-membership stochastic block model20. Previous models have shown the importance of community structure for the spread of misinformation71,74 and the stylized structure of the mixed-membership stochastic block model captures the known heterogeneity of real networks and its modular structure of echo chambers and bridge nodes with diverse neighborhoods83. We then track individuals based on their demographics. These abstract classes, such as 1 or 2, could represent a feature such as younger or older social media users. We also track their state, e.g., currently duped by a deepfake video (infectious) or not (susceptible). We also track the demographics of their neighbors to know their role in the network and exposure to other users in different states.
The resulting model has two critical mechanisms. First, inspired by our survey, individuals get duped by their duped neighbor at a rate λi dependent on their demographic class i. Second, as per previous models and the concept of crowd-sourced approaches to correction of misinformation based on the “self-correcting crowd”38,39,40, duped individuals can be corrected by their susceptible neighbors at a fixed rate γ. The dynamics of the resulting model are tracked using a heterogeneous mean-field approach84 detailed in Box 1 and summarized in Fig. 4.
This model has a simple interesting behavior in homogeneous populations and becomes much more realistic once we account for heterogeneity in susceptibility. In a fully homogeneous population, λi = λ ∀ i, if misinformation can, on average, spread from a first to a second node, it will never stop. The more misinformation spreads, the fewer potential fact-checkers remain. Therefore, misinformation invades the entire population for a correction rate γ lower than some critical value γc, whereas misinformation disappears for γ > γc.
The invasion threshold for misinformation is shown in Fig. 4a. In heterogeneous populations, where different nodes can feature different susceptibility λi, the discontinuous transition from a misinformation-free to a misinformation-full state is relaxed. Instead, a steady state of misinformation can now be maintained at any level depending on the parameters of misinformation and the demographics of the population. In this regime, we can then further break down the dynamics of the system by looking at the role of duped nodes in the network, as shown in Fig. 4b. The key result here is that very susceptible individuals with a homogeneous assortative neighborhood (e.g., an echo chamber) are at the highest risk of being duped. Conversely, nodes in the same demographic class but with a mixed or more diverse neighborhood are more likely to have resilient susceptible neighbors able to correct them if necessary.
Consider now that diverse misinformation spreads. We assume just two types of misinformation (say young or older personas in two deepfake videos) targeting each of our two demographic classes (say younger and older social media users). We show this thought experiment in Fig. 4c where we use two complementary types of misinformation: One with λ1 = λ2/2 = 1.0 and a matching type with \({\lambda }_{2}^{{\prime} }={\lambda }_{1}^{{\prime} }/2=1.0\). We run the dynamics of these two types of misinformation independently as we assume they do not directly interact, and, therefore simply combine the possible states of nodes after integrating the dynamical system. For example, the probability that a node of type 1 is duped by both pieces of misinformation would be the product of the probabilities that it is duped by the first and duped by the second. By doing so, we can easily study a model where multiple, diverse pieces of information spread in a diverse network population.
For diverse misinformation in Fig. 4c, we find two connectivity regimes where the role of network structure is critical. For low-degree nodes, a diverse neighborhood means more exposure to diverse misinformation than a homogeneous echo chamber, such that the misinformation that best matches the demographics of a low-degree user is more likely to find them if they have a diverse neighborhood. For high-degree nodes, however, we find the behavior of herd correction: A diverse neighborhood means a diverse set of neighbors that is more likely to contain users who are able correct you if you become misinformed34,85,86.
In the appendix, we analyze the robustness of herd correction to the parameters of the model. We show mathematically that the protection it offers is directly proportional to the homophily in the network (our parameter Q). By simulating the dynamics with more parameters, we also find that herd correction is proportional to the degree heterogeneity of the network. As we increase heterogeneity, we increase the strength of the friendship paradox. “Your friends have more friends than you do,”87 which means they get more exposed to misinformation than you do but also that they have more friends capable of correcting them when duped.
Our stylized model is meant to show how one can introduce biases in simple mathematical models of diverse misinformation. A first-order effect is that individuals with increased susceptibility should be preferentially duped, but this effect exists only if misinformation can spread (above a certain contagion threshold) but not saturate the population (below certain transmissibility such that the heterogeneity has impact). A second-order effect is that individuals with a diverse neighborhood are also more likely to have friends who can correct them should they be duped by misinformation.
Future modeling efforts should also consider the possible interactions between different kinds of misinformation88. These can be synergistic89, parasitic90, or antagonistic91; which all provide rich dynamical behaviors. Other possible mechanisms to consider are the adaptive feedback loops that facilitate the spread of misinformation in online social networks92.
Discussion
Understanding the structure and dynamics of misinformation is important as it can bring a great amount of societal harm. Misinformation has negatively impacted the ability to disseminate important information during critical elections, humanitarian crises, global unrest, and global pandemics. More importantly, misinformation degrades our epistemic environment, particularly regarding distrust of truths. It is necessary to understand who is susceptible to misinformation and how it spreads on social networks to mitigate its harm and propose meaningful interventions. Further, as deepfakes deceive viewers at greater rates, it becomes increasingly critical to understand who gets duped by this form of misinformation and how our biases and social circle impact our interaction with video content at scale. We hope this work will contribute to the critical literature on human biases and help to better understand their interplay with machine-generated content.
The overarching takeaways of our results can be summarized as follows. If not primed, humans are not particularly accurate at detecting deepfakes. Accuracy varies by demographics, but humans are generally better at classifying videos that match them. These results appear consistent with findings of the own-race bias (ORB) phenomenon18, where overall, we see that participants are better at detecting videos that match their own attributes. Consistent with ORB research93, our study results also show that white participants display a greater accuracy when presented with videos of white personas. We also see strong evidence that persons of color are more accurate than white participants when viewing deepfakes of personas of color and more accurate overall than white participants (see Supplementary Information). Our study adds several extra dimensions of demographic analysis by using gender and age. We see strong evidence that male participants are better at detecting videos of male personas than female viewers. With age, we see strong evidence that when viewing videos of young personas, participants between the ages of 18−29 are more accurate than participants above the age of 30; surprisingly, participants aged 18−29 are also more accurate than participants aged 30-49 even when viewing videos of personas aged 30−49. Combining these results, more work needs to be done to understand better how interventions such as education about deepfakes, cross-demographic experiences and exposure, and exposure to the technology impact a user’s ability to detect deepfakes.
In this observational study, we also explored the potential impacts of these results in a simple mathematical model and extrapolated from our survey to hypothesize that a diverse set of contacts might provide “herd correction” where friends can correct each other’s blind spots. Friends with different biases can better correct each other when duped. This modeling result is a generalization of the self-correcting crowd approach used in the correction of misinformation38.
In future work, we hope to investigate how non-primed human deepfake detectors perform when aided by machines. We want to investigate the mechanisms behind why some human viewers are better at guessing the state of videos that match their own identity. For example, do viewers have a homophily bias because they are more accustomed to images that match their own, or do they simply favor these images? We also would like to empirically investigate our survey via a more robust randomized controlled experiment and model results on real-world social networks with different levels of diversity to measure the spread of diverse misinformation in the wild. Consequently, we would be interested in testing possible educational or other intervention strategies to mitigate adversarial misinformation campaigns. Our simple observational study is a step towards understanding social biases’ role and potential impacts in an emerging societal problem with many multilevel interdependencies.
Methods
Survey methodology
We first ran a pilot stage of our observational study. We conducted a simple convenience sample of 100 participants (aged 18+) to observe the efficacy of our survey. We then ran phase 1 (April−May 2022) of the full survey using a Qualtrics survey panel of 1000 participants who matched the demographic distribution of U.S. social media users. We then ran phase 2 (September 2022) of the full survey, again using Qualtrics and the same sampling methodology. The resulting full study from phases 1 and 2 is a 2016-participant sample.
Towards ensuring that our experiment reflects the real-world context as closely as possible, survey participants did not know before the start of the survey that the videos could potentially be deepfakes. The survey was framed for participants as a study about different communication styles and techniques that help make video content credible. Participants were told that we were trying to understand how aspects of public speaking, such as tone of voice, facial expressions, and body language, contribute to the effectiveness and credibility of a speaker. The survey’s deceptiveness allowed us to ask questions about speaker attributes, likeability, and agreeableness naturally without priming the participants to look specifically for deepfakes94. We chose to make our survey deceptive not to prime the participants but also because this more closely replicates the deceptiveness that a social media user would encounter in the real world. Furthermore, Bröder95 argues that “in studies of cognitive illusions (e.g., hindsight bias or misleading postevent information effect), it is a necessity to conceal the true nature of the experiment.” We posit that our study clearly involves cognitive illusions, specifically in the form of deepfakes, and as such deception is an important tool.
We designed our survey using video clips (as seen in Fig. 5) from the Deepfake Detection Challenge (DFDC) Preview Dataset60,61. In our survey, we ask the participants to view two random video clips, which are approximately 10 s in length each. Each video clip may be viewed unlimited times before reading the questions but not again after moving to the questions. The information necessary to answer these questions relies solely on the previously shown video clip. A link to the full survey and survey questions is available in Appendix 1.
After viewing both videos, the participants are then asked to complete a related questionnaire about the communication styles and techniques of the videos. The questions ask about attributes of the video, such as pose, tone, and style and are asked to rate them on a Likert scale from very pleasant to very unpleasant. We also asked them to rate their agreement with the video content and credibility. We also ask participants to identify the perceived gender expression of the person(a) in the video, to identify what age group they belong to, and to ask if they perceive the person in the video to be a person of color or not.
In line with best practices in ethical research96,97, we debriefed the participants following the viewing of both videos and completion of the questionnaire on communication style and perceived demographics. The participants are debriefed on the deception of the survey, given a short explanation of deepfake technology, and then asked if they think the videos were real or fake (as seen in Fig. 6).
Lastly, we collect demographic information on the survey participants’ backgrounds and expressions of identity. We also ask participants how knowledgeable they already were on deepfakes, how often they use social media, and their political and religious affiliations. We also asked participants if they knew that the survey was about deepfakes before taking the survey (survey participants who were primed were subsequently dropped from the analysis).
Sampling
Survey responses from 2016 participants were collected through Qualtrics, an IRB-approved research panel provider, via traditional, actively managed, double-opt-in research panels6. Qualtrics’ participants for this study were randomly selected stratified samples from the Qualtrics panel membership pool that represents the average social media user in the U.S.98 Our survey respondents represent the following categories and demographic breakdown in Table 3.
Secondary Data
For this project, we use the publicly available Facebook AI Research Deepfake Detection Challenge (DFDC) Preview Dataset (N = 5000 video clips)60,61. For our purposes, we filtered out all videos from the dataset that featured more than one person(a). The video clips may be deepfake or real; see Table 4. Additionally, some of the videos have been purposefully altered in several ways. Here is the list of augmenters and distractors:
-
Augmenters: Frame-rate change, Quality level, Audio removal, Introduction of audio noise, Brightness or contrast level, Saturation, Resolution, Blur, Rotation, Horizontal flip.
-
Distractors: Dog filter, Flower filter, Introduction of overlaid images, shapes, or dots, Introduction of additional faces, Introduction of text.
A video’s deepfake status (deepfake or not) was not revealed to the respondents during or after the survey. Many augmenters and distractors were noticeable to the respondents but were not specifically revealed.
Original Data
We transformed all survey response variables of interest into numerical form to analyze our survey results. All Likert survey questions were converted from ‘Very unpleasant,’ ‘Unpleasant,’ ‘Neutral,’ ‘Pleasant,’ and ‘Very pleasant’ to an ordinal scale of 1,2,3,4,5.
Participants selected education levels from ‘Some high school,’ ‘High school diploma or equivalent,’ ‘Some college,’ Associate’s degree (e.g., A.A., A.E., A.F.A., AS, A.S.N.),’ ‘Vocational training,’ ‘Bachelor’s degree (e.g., B.A., BBA BFA, BS),’ ‘Some postgraduate work,’ ‘Master’s degree (e.g., M.A., M.B.A., M.F.A., MS, M.S.W.),’ ‘Specialist degree (e.g., EdS),’ ‘Applied or professional doctorate degree (e.g., M.D., D.D.C., D.D.S., J.D., PharmD), ‘Doctorate degree (e.g., EdD, Ph.D.)’ was transformed to an ordinal scale of 1-11 respectively.
Participants selected income levels from ‘Less than $30,000’, ‘$30,000−$49,999’, ‘$50,000−$74,999’, ‘$75,000+’ were transformed to an ordinal scale of 1−4 respectively.
Participants selected their social media usage levels from ‘I do not use social media,’ ‘I use social media but less than once a month,’ ‘Once a month,’ ‘A few times a month,’ ‘Once a week,’ ‘A few times a week,’ ‘Once a day,’ ‘More than once a day’ were transformed to an ordinal scale of 1−8 respectively. Variables were split into the category of frequent social media users 5−8 and infrequent social media users 1−4. We combined the ordinal scales into two categories in order to reduce the dimensionality of our data.
Participants selected their knowledge of deepfake from ‘I did not know what a deepfake was,’ ‘I somewhat knew what a deepfake was,’ ‘I knew what a deepfake was,’ ‘I consider myself knowledgeable about deepfakes’ was transformed to an ordinal scale of 1−4 respectively. Variables were split into users who are knowledgeable about deepfakes 3−4 and users who are not knowledgeable about deepfakes 1−2. We combined the ordinal scales into two categories in order to reduce the dimensionality of our data.
All nominal and categorical variables were transformed into binary variables. Categorical variables (some survey questions included write-in answers) were combined into coarser-grained categories for analysis, such as participant racial/ethnic identity (transformed to Person of Color or White), U.S. state of residence (transformed to U.S. regions), employment (transformed to occupational sectors), religious affiliation (transformed into religious affiliations), and political affiliation (transformed to major political affiliations).
We allowed survey participants to identify their gender identity, the results of which were largely binary. Unfortunately, our sample was insufficient to perform meaningful analysis on a larger non-binary gender identity spectrum. Primary variables with an N under 30 were dropped, meaning the participant’s responses were not included in the analysis (this was only applicable for non-binary gender responses where N = 13). Our survey participants were given two video clips to view and critique; in our analysis, we decided to analyze the first or second video in the same way.
Analytical methods
We use Matthews Correlation Coefficient to understand the relationship between the participant’s guesses on the status of the video (fake or real) and the actual state of the video (fake or real), we ran a Matthews Correlation Coefficient (MCC)62,63 to compare what variables show strong evidence to impact a participant’s ability to guess the actual state of the video correctly. MCC is typically used for classification models to observe the classifier’s performance. Here we treat human participant subgroups as classifiers and measure their performance with MCC. MCC takes the participant subgroup’s guesses and the actual answers and breaks them up into the following categories: number of true positives (TP), number of true negatives (TN), number of false positives (FP), and number of false negatives (FN). The MCC metric ranges from −1 to 1, where 1 indicates total agreement between participant guess about the video and the actual state of the video, −1 indicates complete disagreement between participant guess about the video and the actual state of the video, and 0 indicates something similar to a random guess. To calculate the MCC metric for our human classifiers, we then use the following formula:
MCC is considered a more balanced statistical measure than an F1, precision, or recall score because it is symmetric, meaning no class (e.g., TP, TN, FP, FN) is more important than another.
To compare MCC scores, we bootstrap samples from pairs of confusion matrices and compare their MCC scores. This process generates 10,000 bootstrapped samples of differences in correlation coefficients. We then compare the null hypothesis (difference equal to zero) to the bootstrapped distribution to measure the evidence level of biases and get a credibility interval on their strength.
Logistic regression
To understand the relationship between matching demographics and guess accuracy we run a Bayesian logistic regression on matching demographics (age matches, gender matches, race matches). Logistic regression is a statistical analysis method used to model and predict binary outcomes (the participant’s accuracy). Accuracy is equal to 1 if the participant’s guess about the video was correct and 0 if it was incorrect. It utilizes prior observations from a dataset to establish relationships and make predictions based on specific variables.
Accuracy rate
The performance metric we use to measure participant accuracy is the ratio of the correct guesses to the entire pool of guesses. The accuracy is thus equal to the sum of true positives and true negatives over the total number of guesses.
Data Availability
Our full survey questionnaire, code, data, and codebook can be found on our GitHub repository. https://github.com/juniperlovato/DiverseMisinformationPaper Due to the nature of this research, participants of this study did not consent for their personally identifiable data to be shared publicly, so the full survey’s raw individual level supporting data is not available. Aggregated and anonymized data needed for analysis can be found in our repository.
References
Bagrow, J. P., Liu, X. & Mitchell, L. Information flow reveals prediction limits in online social activity. Nat. Hum. Behav. 3, 122–128 (2019).
Lovato, J. L., Allard, A., Harp, R., Onaolapo, J. & Hébert-Dufresne, L. Limits of individual consent and models of distributed consent in online social networks. In 2022 ACM Conf. Fairness Account. Transpar., 2251–2262, https://doi.org/10.1145/3531146.3534640 (2022).
Garland, J., Ghazi-Zahedi, K., Young, J.-G., Hébert-Dufresne, L. & Galesic, M. Impact and dynamics of hate and counter speech online. EPJ Data Sci. 11, 3 (2022).
Chesney, R. & Citron, D. K. Deep fakes: a looming challenge for privacy, democracy, and national security. SSRN Electron. J. 107, 1753 (2018).
Groh, M., Epstein, Z., Firestone, C. & Picard, R. Deepfake detection by human crowds, machines, and machine-informed crowds. Proc. Natl. Acad. Sci. 119, e2110013119 (2021).
Boas, T. C., Christenson, D. P. & Glick, D. M. Recruiting large online samples in the united states and india: Facebook, mechanical turk, and qualtrics. Political Sci. Res. Methods 8, 232–250 (2018).
Ebner, N. C. et al. Uncovering susceptibility risk to online deception in aging. J. Gerontol.: B 75, 522–533 (2018).
Lloyd, E. P., Hugenberg, K., McConnell, A. R., Kunstman, J. W. & Deska, J. C. Black and white lies: race-based biases in deception judgments. Psychol. Sci. 28, 1125–1136 (2017).
Bond, J., Julion, W. A. & Reed, M. Racial discrimination and race-based biases on orthopedic-related outcomes. Orthop. Nurs. 41, 103–115 (2022).
Klaczynski, P. A., Felmban, W. S. & Kole, J. Gender intensification and gender generalization biases in pre-adolescents, adolescents, and emerging adults. Brit. J. Dev. Psychol. 38, 415–433 (2020).
Macchi Cassia, V. Age biases in face processing: The effects of experience across development. Brit. J. Psychol. 102, 816–829 (2011).
Dandekar, P., Goel, A. & Lee, D. T. Biased assimilation, homophily, and the dynamics of polarization. Proc. Natl. Acad. Sci. 110, 5791–5796 (2013).
Currarini, S. & Mengel, F. Identity, homophily and in-group bias. Eur. Econ. Rev. 90, 40–55 (2016).
Kossinets, G. & Watts, D. J. Origins of homophily in an evolving social network. Am. J. Sociol. 115, 405–450 (2009).
Nightingale, S. J., Wade, K. A. & Watson, D. G. Investigating age-related differences in ability to distinguish between original and manipulated images. Psychol. Aging 37, 326–337 (2022).
Bothwell, R. K., Brigham, J. C. & Malpass, R. S. Cross-racial identification. Pers. Soc. Psychol. B. 15, 19–25 (1989).
Brigham, J. C., Maass, A., Snyder, L. D. & Spaulding, K. Accuracy of eyewitness identification in a field setting. J. Pers. Soc. Psychol. 42, 673–681 (1982).
Meissner, C. A. & Brigham, J. C. Thirty years of investigating the own-race bias in memory for faces: a meta-analytic review. Psychol. Public Policy Law 7, 3–35 (2001).
Leskovec, J., Backstrom, L., Kumar, R. & Tomkins, A. Microscopic evolution of social networks. In Proc. 14th ACM SIGKDD int. conf. Knowl. discov. data min., 462–470, https://doi.org/10.1145/1401890.1401948 (2008).
Airoldi, E. M., Blei, D., Fienberg, S. & Xing, E. Mixed membership stochastic blockmodels. In Koller, D., Schuurmans, D., Bengio, Y. & Bottou, L. (eds.) Advances in Neural Information Processing Systems, Vol. 21, 1–8, https://proceedings.neurips.cc/paper_files/paper/2008/file/8613985ec49eb8f757ae6439e879bb2a-Paper.pdf (Curran Associates, Inc., 2008).
Traberg, C. S. & van der Linden, S. Birds of a feather are persuaded together: Perceived source credibility mediates the effect of political bias on misinformation susceptibility. Pers. Indiv. Differ. 185, 111269 (2022).
Calvillo, D. P., Garcia, R. J., Bertrand, K. & Mayers, T. A. Personality factors and self-reported political news consumption predict susceptibility to political fake news. Pers. Indiv. Differ. 174, 110666 (2021).
Lazer, D. M. J. et al. The science of fake news. Science 359, 1094–1096 (2018).
Watts, D. J., Rothschild, D. M. & Mobius, M. Measuring the news and its impact on democracy. Proc. Natl. Acad. Sci. 118, e1912443118 (2021).
Roth, C., St-Onge, J. & Herms, K. Quoting is not citing: Disentangling affiliation and interaction on twitter. In Benito, R. M. et al. (eds.) Complex Networks & their Applications X, Studies in Computational Intelligence, 705–717, https://doi.org/10.1007/978-3-030-93409-5_58 (Springer Int. Publ., 2022).
Appel, M. & Prietzel, F. The detection of political deepfakes. J. Comput.-Mediat. Commun. 27, zmac008 (2022).
Ahmed, S. Who inadvertently shares deepfakes? analyzing the role of political interest, cognitive ability, and social network size. Telemat. Inform. 57, 101508 (2021).
Jacobsen, B. N. & Simpson, J. The tensions of deepfakes. Inf. Commun. & Soc. 1–15, https://doi.org/10.1080/1369118x.2023.2234980 (2023).
Chou, W.-Y. S., Oh, A. & Klein, W. M. P. Addressing health-related misinformation on social media. JAMA 320, 2417 (2018).
Tasnim, S., Hossain, M. M. & Mazumder, H. Impact of rumors and misinformation on COVID-19 in social media. J. Prev. Med. Pub. Health 53, 171–174 (2020).
Kimmel, A. J. Rumors and the financial marketplace. J. Behav. Finance 5, 134–141 (2004).
Rini, R. Deepfakes and the epistemic backstop. Philos. Impr. 20, 1–16 (2020).
Vaccari, C. & Chadwick, A. Deepfakes and disinformation: exploring the impact of synthetic political video on deception, uncertainty, and trust in news. Soc. Media Soc. 6, 205630512090340 (2020).
Walter, N., Brooks, J. J., Saucier, C. J. & Suresh, S. Evaluating the impact of attempts to correct health misinformation on social media: A meta-analysis. Health Commun. 36, 1776–1784 (2020).
Wu, L., Morstatter, F., Carley, K. M. & Liu, H. Misinformation in social media. ACM SIGKDD Explor. Newsl. 21, 80–90 (2019).
Starbird, K., Maddock, J., Orand, M., Achterman, P. & Mason, R. M. Rumors, false flags, and digital vigilantes: Misinformation on twitter after the 2013 Boston marathon bombing. IConference 2014 proc. (2014).
Sedhai, S. & Sun, A. HSpam14. In Proc. 38th Int. ACM SIGIR Conf. Res. Dev. Inf. Retr., 223–232, https://doi.org/10.1145/2766462.2767701 (ACM, 2015).
Arif, A. et al. A closer look at the self-correcting crowd. In Proc. 2017 ACM Conf. Comput. Support. Coop. Work Soc. Comput., Cscw ’17, 155–168, https://doi.org/10.1145/2998181.2998294 (ACM, New York, NY, USA, 2017).
Micallef, N., He, B., Kumar, S., Ahamad, M. & Memon, N. The role of the crowd in countering misinformation: A case study of the COVID-19 infodemic. In 2020 IEEE Int. Conf. Big Data (Big Data), 748–757, https://doi.org/10.1109/bigdata50022.2020.9377956. Ieee (IEEE, 2020).
Allen, J., Arechar, A. A., Pennycook, G. & Rand, D. G. Scaling up fact-checking using the wisdom of crowds. Sci. Adv. 7, eabf4393 (2021).
Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A. & Ortega-Garcia, J. Deepfakes and beyond: a survey of face manipulation and fake detection. Inform. Fusion 64, 131–148 (2020).
Roose, K. Here come the fake videos, too. The New York Times 4 (2018).
Mori, M. The uncanny valley: The original essay by masahiro Mori. IEEE Spectr. (1970).
Verdoliva, L. Media forensics and DeepFakes: An overview. IEEE J. Sel. Top. Signal Process. 14, 910–932 (2020).
Jung, T., Kim, S. & Kim, K. DeepVision: Deepfakes detection using human eye blinking pattern. IEEE Access 8, 83144–83154 (2020).
Guera, D. & Delp, E. J. Deepfake video detection using recurrent neural networks. In 2018 15th IEEE Int. Conf. Adv. Video Signal Based Surveill. (AVSS), 1–6, https://doi.org/10.1109/avss.2018.8639163. IEEE (IEEE, 2018).
Zotov, S., Dremliuga, R., Borshevnikov, A. & Krivosheeva, K. DeepFake detection algorithms: A meta-analysis. In 2020 2nd Symp. Signal Process. Syst., 43–48, https://doi.org/10.1145/3421515.3421532 (ACM, 2020).
Blue, L. et al. Who are you (I really wanna know)? detecting audio DeepFakes through vocal tract reconstruction. In 31st USENIX Secur. Symp. (USENIX Secur. 22), 2691–2708 (Boston, MA, 2022).
Ng, J. C. K., Au, A. K. Y., Wong, H. S. M., Sum, C. K. M. & Lau, V. C. Y. Does dispositional envy make you flourish more (or less) in life? an examination of its longitudinal impact and mediating mechanisms among adolescents and young adults. J. Happiness Stud. 22, 1089–1117 (2020).
Shillair, R. & Dutton, W. H. Supporting a cybersecurity mindset: Getting internet users into the cat and mouse game. Soc. Sci. Res. Netw. (2016).
Greengard, S. Will deepfakes do deep damage? Commun. ACM 63, 17–19 (2019).
Schwartz, G. T. Explaining and justifying a limited tort of false light invasion of privacy. Case W. Res. L. Rev. 41, 885 (1990).
Fallis, D. The epistemic threat of deepfakes. Philos. & Technol. 34, 623–643 (2020).
Harris, D. Deepfakes: False pornography is here and the law cannot protect you. Duke Law & Technol. Rev. 17, 99 (2018).
de Ruiter, A. The distinct wrong of deepfakes. Philos. & Technol. 34, 1311–1332 (2021).
115th Congress (2017–2018), S. –. Malicious deep fake prohibition act of 2018 (2018).
Citron, D. K.The fight for privacy: Protecting dignity, identity, and love in the digital age (W.W. Norton & Company, 2022), first edn.
on the Judiciary House of Representatives, T. C. Federal rules of evidence (2019).
Solove, D. J. Conceptualizing privacy. Calif. Law Rev. 90, 1087 (2002).
Dolhansky, B., Howes, R., Pflaum, B., Baram, N. & Ferrer, C. The deepfake detection challenge (DFDC) preview dataset. Preprint at https://arxiv.org/abs/1910.08854 (2019).
Dolhansky, B. et al. The DeepFake detection challenge dataset. Preprint at https://arxiv.org/abs/2006.07397 (2020).
Matthews, B. Comparison of the predicted and observed secondary structure of t4 phage lysozyme. Biochim. Biophys. Acta (BBA) - Protein Struct. 405, 442–451 (1975).
Boughorbel, S., Jarray, F. & El-Anbari, M. Optimal classifier for imbalanced data using Matthews correlation coefficient metric. PLoS One 12, e0177678 (2017).
Azur, M. J., Stuart, E. A., Frangakis, C. & Leaf, P. J. Multiple imputation by chained equations: What is it and how does it work? Int. J. Method. Psych. 20, 40–49 (2011).
Cheng, J. & Bernstein, M. S. Flock. In Proc. 18th ACM Conf. Comput. Support. Coop. Work & Soc. Comput., CSCW ’15, 600–611, https://doi.org/10.1145/2675133.2675214 (ACM, New York, NY, USA, 2015).
Josephs, E., Fosco, C. & Oliva, A. Artifact magnification on deepfake videos increases human detection and subjective confidence. J. Vision 23, 5327 (2023).
Aliberti, G., Di Pietro, R. & Guarino, S. Epidemic data survivability in unattended wireless sensor networks: New models and results. J. Netw. Comput. Appl. 99, 146–165 (2017).
Jin, F., Dougherty, E., Saraf, P., Cao, Y. & Ramakrishnan, N. Epidemiological modeling of news and rumors on twitter. In Proc. 7th Workshop Soc. Netw. Min. Anal., 1–9, https://doi.org/10.1145/2501025.2501027 (ACM, 2013).
Kimura, M., Saito, K. & Motoda, H. Efficient estimation of influence functions for SIS model on social networks. In Twenty-First Int. Jt. Conf. Artif. Intell. (2009).
Di Pietro, R. & Verde, N. V. Epidemic theory and data survivability in unattended wireless sensor networks: Models and gaps. Pervasive Mob. Comput. 9, 588–597 (2013).
Shang, J., Liu, L., Li, X., Xie, F. & Wu, C. Epidemic spreading on complex networks with overlapping and non-overlapping community structure. Physica A 419, 171–182 (2015).
Scaman, K., Kalogeratos, A. & Vayatis, N. Suppressing epidemics in networks using priority planning. IEEE Trans. Network Sci. Eng. 3, 271–285 (2016).
van der Linden, S. Misinformation: Susceptibility, spread, and interventions to immunize the public. Nat. Med. 28, 460–467 (2022).
Weng, L., Menczer, F. & Ahn, Y.-Y. Virality prediction and community structure in social networks. Sci. Rep. 3, 1–6 (2013).
Bao, Y., Yi, C., Xue, Y. & Dong, Y. A new rumor propagation model and control strategy on social networks. In Proc. 2013 IEEE/ACM Int. Conf. Adv. Soc. Netw. Anal. Min., 1472–1473, https://doi.org/10.1145/2492517.2492599 (ACM, 2013).
Zhang, N., Huang, H., Su, B., Zhao, J. & Zhang, B. Dynamic 8-state ICSAR rumor propagation model considering official rumor refutation. Physica A 415, 333–346 (2014).
Hong, W., Gao, Z., Hao, Y. & Li, X. A novel SCNDR rumor propagation model on online social networks. In 2015 IEEE Int. Conf. Consum. Electron. - Taiwan, 154–155, https://doi.org/10.1109/icce-tw.2015.7216829. IEEE (IEEE, 2015).
Tambuscio, M., Ruffo, G., Flammini, A. & Menczer, F. Fact-checking effect on viral hoaxes. In Proc. 24th Int. Conf. World Wide Web, 977–982, https://doi.org/10.1145/2740908.2742572 (ACM, 2015).
Xiao, Y. et al. Rumor propagation dynamic model based on evolutionary game and anti-rumor. Nonlinear Dynam. 95, 523–539 (2018).
Zhang, Y., Su, Y., Weigang, L. & Liu, H. Rumor and authoritative information propagation model considering super spreading in complex social networks. Physica A 506, 395–411 (2018).
Kumar, K. K. & Geethakumari, G. Information diffusion model for spread of misinformation in online social networks. In 2013 Int. Conf. Adv. Comput. Commun. Inform. (ICACCI), 1172–1177, https://doi.org/10.1109/icacci.2013.6637343. IEEE (IEEE, 2013).
King, K. K., Wang, B., Escobari, D. & Oraby, T. Dynamic effects of falsehoods and corrections on social media: A theoretical modeling and empirical evidence. J. Manage. Inform. Syst. 38, 989–1010 (2021).
Red, V., Kelsic, E. D., Mucha, P. J. & Porter, M. A. Comparing community structure to characteristics in online collegiate social networks. SIAM Rev. 53, 526–543 (2011).
Pastor-Satorras, R. & Vespignani, A. Epidemic spreading in scale-free networks. Phys. Rev. Lett. 86, 3200–3203 (2001).
Bode, L. & Vraga, E. K. In related news, that was wrong: The correction of misinformation through related stories functionality in social media. J. Commun. 65, 619–638 (2015).
Vraga, E. K. & Bode, L. Using expert sources to correct health misinformation in social media. Sci. Commun. 39, 621–645 (2017).
Feld, S. L. Why your friends have more friends than you do. Am. J. Sociol. 96, 1464–1477 (1991).
Chang, H.-C. H. & Fu, F. Co-diffusion of social contagions. New J. Phys. 20, 095001 (2018).
Hébert-Dufresne, L. & Althouse, B. M. Complex dynamics of synergistic coinfections on realistically clustered networks. Proc. Natl. Acad. Sci. 112, 10551–10556 (2015).
Hébert-Dufresne, L., Mistry, D. & Althouse, B. M. Spread of infectious disease and social awareness as parasitic contagions on clustered networks. Phys. Rev. Research 2, 033306 (2020).
Fu, F., Christakis, N. A. & Fowler, J. H. Dueling biological and social contagions. Sci. Rep. 7, 1–9 (2017).
Törnberg, P. Echo chambers and viral misinformation: Modeling fake news as complex contagion. PLoS One 13, e0203958 (2018).
Anthony, T., Copper, C. & Mullen, B. Cross-racial facial identification: A social cognitive integration. Pers. Soc. Psychol. B. 18, 296–301 (1992).
Barrera, D. & Simpson, B. Much ado about deception. Sociol. Methods & Res. 41, 383–413 (2012).
Bröder, A. Deception can be acceptable. Am. Psychol. 53, 805–806 (1998).
Greene, C. M. et al. Best practices for ethical conduct of misinformation Research. Eur. Psychol. 28, 139–150 (2023).
Boynton, M. H., Portnoy, D. B. & Johnson, B. T. Exploring the ethics and psychological impact of deception in psychological research. IRB 35, 7 (2013).
Center, P. R. Social media fact sheet. Pew Research Center: Washington, DC, USA (2021).
Acknowledgements
Institutional Review Board Approval: The survey in this project is CHRBSS (Behavioral) STUDY00001786, approved by the University of Vermont I.R.B. on 12/6/2021. The authors would like to thank Anne Marie Stupinski, Nana Nimako, Austin Block, and Alex Friedrichsen for their feedback on early drafts and Jean-Gabriel Young and Maria Sckolnick for comments on our analysis. The authors would also like to thank Engin Kirda and Wil Robertson for their contributions to an early survey prototype. This work is supported by the Alfred P. Sloan Foundation, The UVM OCEAN Project, and MassMutual under the MassMutual Center of Excellence in Complex Systems and Data Science. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the aforementioned financial supporters.
Author information
Authors and Affiliations
Contributions
Author contributions: Conceptual: J.L., J.O., R.H., L.H-D.; Survey Development: J.L., J.O., R.H.; Survey Implementation: J.L., I.U.H., J.S-O., S.P.R., G.S.L.; Wrangling and Analysis: J.L., J.S-O., G.S.L., S.P.R., L.H-D.; Mathematical Model: L.H-D., J.L.; All authors drafted the manuscript, revised it critically for important intellectual content, gave final approval of the completed version, contributed to the conception of the work, and are accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.
Corresponding author
Ethics declarations
Competing interests
The authors declare no Competing Financial Interests but the following Competing Non-Financial Interests: the author, Laurent Hébert-Dufresne, is the Editor-in-Chief for npj Complexity and was not involved in the journal’s review of, or decisions related to, this manuscript.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lovato, J., St-Onge, J., Harp, R. et al. Diverse misinformation: impacts of human biases on detection of deepfakes on networks. npj Complex 1, 5 (2024). https://doi.org/10.1038/s44260-024-00006-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s44260-024-00006-y