key: cord-1048758-p96djq79 authors: Kim, Hye Kyung; Tandoc, Edson C. title: Consequences of Online Misinformation on COVID-19: Two Potential Pathways and Disparity by eHealth Literacy date: 2022-02-14 journal: Front Psychol DOI: 10.3389/fpsyg.2022.783909 sha: 4a9b8ab3e79681ebbe2303e6f2c9103ac4f398fe doc_id: 1048758 cord_uid: p96djq79 The COVID-19 pandemic poses an unprecedented threat to global human wellbeing, and the proliferation of online misinformation during this critical period amplifies the challenge. This study examines consequences of exposure to online misinformation about COVID-19 preventions. Using a three-wave panel survey involving 1,023 residents in Singapore, the study found that exposure to online misinformation prompts engagement in self-reported misinformed behaviors such as eating more garlic and regularly rinsing nose with saline, while discouraging evidence-based prevention behaviors such as social distancing. This study further identifies information overload and misperception on prevention as important mechanisms that link exposure to online misinformation and these outcomes. The effects of misinformation exposure differ by individuals’ eheath literacy level, suggesting the need for a health literacy education to minimize the counterproductive effects of misinformation online. This study contributes to theory-building in misinformation by addressing potential pathways of and disparity in its possible effects on behavior. In a publicly televised press briefing sometime in April 2020, the former United States President Donald Trump suggested looking into injecting disinfectant as a potential way to treat the novel coronavirus disease 2019 (here after referred to as . This was during the time when COVID-19 had already infected more than 800,000 and killed more than 40,000 in the United States. President Trump later clarified he was being sarcastic, but that did not stop some individuals from trying it. Right after the President's remarks, New York City's poison control center reported receiving calls related to exposure to bleach and disinfectant as doctors scrambled to warn people not to ingest disinfectants (Slotkin, 2020) . Indeed, during the COVID-19 pandemic, medical experts around the world found themselves having to control the spread of not only the virus, but also misinformation about it. After declaring a global pandemic, the World Health Organization (WHO) also warned about an "infodemic" as misinformation on causes, remedies, and prevention spread online (Thomas, 2020) . Fact-checking groups had to debunk viral posts, such as those claiming that eating garlic, consuming bananas, and gargling with saltwater can prevent or cure COVID-19. Although studies and reports have documented ways that misinformation affected a range of behaviors, especially in the political context (Allcott and Gentzkow, 2017; Valenzuela et al., 2019) , the effects of misinformation are particularly crucial to understand in the context of a pandemic, when individuals need reliable information to make critical decisions that can affect their own as well as others' well-being. Governments around the world have asked their citizens to protect themselves and one another by engaging in preventive behaviors, such as handwashing and practicing safe distancing; and yet viral messages had also recommended preventive behaviors that medical experts had debunked as ineffective or even harmful. Exposed to a myriad of recommended behaviors-both reliable and unreliable-how did individuals behave in terms of prevention? Using data from a three-wave panel survey conducted during the early stages of the COVID-19 pandemic, this current study documented adverse consequences of online misinformation exposure for people's self-reported adoption of preventive behaviors. In doing so, this study advanced theoretical understanding of misinformation effects by addressing two potential mechanisms of how online misinformation exposure leads to negative outcomes, guided by the protection motivation theory (Rogers, 1975) , and work on information overload (Schneider, 1987) . To offer practical insights in communicating risk during a pandemic, this study also addressed ehealth literacy as an important boundary condition that determines an individual's vulnerability to the negative impact of online misinformation. Misinformation refers to the presence of objectively incorrect or false information, which is "not supported by clear evidence and expert opinion (Nyhan and Reifler, 2010) ." Others use it as a term to encompass information that is completely fabricated as well as information that mixes truths and falsehoods to mislead others; this can be either "deliberately promoted or accidentally shared" (Southwell et al., 2018, p. 1) . It is somewhat complicated to decide what is scientifically true or false, especially in a pandemic, as new scientific evidence emerges over time that may contradict previous beliefs and findings about the disease. In the current study, we classify as misinformation claims that have been debunked or corrected by expert scientific consensus as either completely false or misleading at the time of the study. Misinformation can be either deliberately promoted or accidentally shared by individuals (Southwell et al., 2018) and the likelihood of propagation depends on a variety of factors that range from individual to community level factors, as well as to content specific factors (Southwell, 2013) . Scholars have also scrutinized the role of social media platforms in the propagation of misinformation-not only do these platforms provide easy and accessible channels for misinformation to spread, but the logics behind social networking, such as prioritizing social relationship and mutual exchange, also provide an incentive to social media users to exchange humorous, outrageous, and popular content, sometimes at the expense of information quality (Duffy et al., 2019) . The swarm of misinformation online, however, is not only due to humans sharing them-the misinformation ecosystem is also characterized by bots designed to spread falsehoods, hoaxes, and conspiracy theories (Wang et al., 2018; Jones, 2019) . While many studies have focused on political misinformation (Allcott and Gentzkow, 2017; Bakir and McStay, 2018; Farhall et al., 2019) , another popular topic for misinformation is health, which can be seen in the volume of conspiracy theories and hoaxes about vaccination and cancer treatment, among others, being spread on social media (Oyeyemi et al., 2014; Dredze et al., 2016; Wang et al., 2019; Rodgers and Massac, 2020) . A classic example of health-related misinformation is the false claim that the measles, mumps, rubella (MMR) vaccine causes autism. After the claim was initially published in 1998 (Wakefield et al., 1998) , the scientific community refuted it (Taylor et al., 1999) and the author eventually retracted the publication. However, the false claim still influences many parents' decision on child vaccination as evidenced by public health emergencies due to measles outbreak in the US and European countries in 2019 (The City of NY, 2019; World Health Organization, 2019). In the context of COVID-19, an analysis of messages being forwarded on the messaging app WhatsApp in Singapore found that 35% of the forwarded messages were based on falsehoods, while another 20% mixed true and false information (Tandoc and Mak, 2020) . Among different types of false claims on COVID-19, information concerning the prevention and treatment of COVID-19 poses a great threat to public health by misleading people to engage in ineffective and potentially harmful remedies. To counter the negative impact of misinformation, the WHO and government agencies across countries have monitored and clarified false or still not scientifically tested claims about COVID-19 prevention and cure. Studies have examined the effects of misinformation on perceptions and information behaviors (e.g., sharing) (e.g., Guess et al., 2019; Valenzuela et al., 2019) , as well as on self-reported preventive behaviors (Lee et al., 2020; Roozenbeek et al., 2020; Loomba et al., 2021) . In the context of COVID-19, a cross-sectional survey of 1,049 South Koreans found an association between receiving misinformation and misinformation belief as well as between misinformation belief and the number of preventive behaviors performed (Lee et al., 2020) . Using five national samples, Roozenbeek et al. (2020) also found a negative association between belief in COVID-19 misinformation and self-reported compliance with health guideline. A randomized controlled trial also found that exposure to online misinformation on COVID-19 vaccine reduced vaccination intention in the United Kingdom and United States (Loomba et al., 2021) . Studies point to possible effects of online misinformation exposure; however, the indirect effects of online misinformation were not tested, when doing so can help to identify psychological mechanisms of misinformation effect. More importantly, there is a dearth of studies that address engagement in the behaviors advocated by misinformation (i.e., misinformed behaviors, such as eating garlic) beyond compliance to health guidelines (e.g., social distancing). One possibility is that the mechanism of how misinformation shapes misinformed behaviors differs from how misinformation affects evidence-based practices. To fill these important gaps, we examined whether exposure to online misinformation on COVID-19 prevention influences self-reported engagement in misinformed behaviors as well as evidence-based practices advocated by health authorities. Individuals may not be as motivated to engage in evidencebased practices when they learn about alternative, misinformed remedies to protect themselves against COVID-19. A recent cross-sectional survey found that belief in the prevalence of misinformation on COVID-19 motivates people to comply with official guidelines (Hameleers et al., 2020) . However, given that some individuals often do not evaluate the veracity of (mis) information online (Pennycook and Rand, 2019) , misinformation exposure may negatively influence their evidence-based practices without them realizing that what they had considered as "evidence" was unreliable information (Loomba et al., 2021) . Understanding psychological mechanisms underlying the effect of online misinformation exposure on self-reported behaviors is important for theory development as well as for mitigating the negative effects of online misinformation. Informed by theories of behavior change and cognitive load, respectively, we explored two potential pathways: misperception on prevention and information overload. Misperception refers to incorrect or false beliefs, which could be formed because of exposure to misinformation (Southwell et al., 2018) . In the current context, misperception refers to incorrect response efficacy associated with a misinformed behavior as to whether adopting the behavior will be effective at reducing the risk of COVID-19. Researchers have addressed the challenge in correcting misinformation once misperceptions are formed (Lewandowsky et al., 2012; Pluviano et al., 2017; Walter and Tukachinsky, 2020) . Individuals find messages consistent with their prior beliefs as more persuasive than counterattitudinal messages, sometimes at the expense of accuracy (Lazer et al., 2018) . According to McGuire (1989) , to change a belief or attitude, individuals need to be first exposed to relevant information, pay attention, and process the information. When the information is accepted and stored in memory, individuals retrieve that information when they make behavioral decisions. Research has shown that people often do not think critically about the veracity of information they encounter online (Pennycook and Rand, 2019) . During a pandemic marked by uncertainty, individuals may be motivated to believe whatever possible measures they can take to protect themselves regardless of the veracity of the information they find. The Protection Motivation Theory (Rogers, 1975) theorizes that believing in the effectiveness of a protective action in reducing the disease's threat promotes actual engagement in the protective action. If individuals believe that misinformed behaviors such as eating garlic and regularly gargling with mouthwash can help reduce their risk of COVID-19, they will be more likely to engage in those behaviors. Indeed, one study found that believing in cancer misinformation (e.g., "indoor tanning is more dangerous than tanning outdoors") reduced intention to perform the behavior (e.g., indoor tanning) (Tan et al., 2015) . In this study, we examined whether perceived efficacy of prevention measures against COVID-19, regardless of whether they are scientifically proven or not, increases subsequent behavior. Furthermore, early research on cognitive load has also shown that individuals have limitations in the amount of information they can process (Miller, 1960 (Miller, , 1994 . Information overload occurs when the processing requirements exceed this processing capacity (Schneider, 1987) . Information overload is defined as the state when an individual cannot properly process and utilize all the given information (Jones et al., 2004) . It often involves the subjective feelings of stress, confusion, pressure, anxiety or low motivation (O'Reilly, 1980) . The quality of decisions correlates positively with the amount of information one receives but only up to a certain point (Chewning and Harrell, 1990) . Beyond this critical point, additional information will no longer be integrated into making decisions and information overload will take place (O'Reilly, 1980) , resulting in confusion and difficulty in recalling prior information (Schick et al., 1990) . Besides the amount of information, the level of ambiguity and uncertainty also contribute to the likelihood of information overload (Keller and Staelin, 1987; Schneider, 1987) . That is because individuals can more easily process high-quality information than low-quality or unclear information (Schneider, 1987) . Given that online misinformation contains uncertain and novel information regarding the prevention of COVID-19, exposure to such information can increase information overload. Relatedly, a study found that exposure to misinformation led to information avoidance and less systematic processing of COVID-19 information (Kim et al., 2020) . With information overload, objective facts become less influential in shaping people's perceptions and behaviors as people find it difficult to identify and select important information (Eppler and Mengis, 2004; Vogel, 2017) . In health contexts, a study found that higher levels of information overload predicted higher levels of self-reported stress and poorer health status at a 6-weeks follow-up (Misra and Stokols, 2011) . In another study, cancer information overload significantly decreased cancer screening behavior at 18-month follow-up (Jensen et al., 2014) . As such, information overload may prevent individuals from finding and processing information regarding evidence-based practices on COVID-19 prevention. Scholars have sought to explain why some individuals believe or act on inaccurate information online (Tandoc, 2019) . A study suggested that the effect of COVID-19 misinformation varies across sociodemographic groups, based on gender, ethnicity, or employment status (Loomba et al., 2021) . In another study, those who are less capable of processing basic numerical concepts (i.e., low numeracy) were found to be more susceptible to misinformation on COVID-19 . Researchers suggested that technological advancement has exacerbated health disparity, especially among more vulnerable groups (Parker et al., 2003) , partly due to limited navigation skills required for those technologies (Cline and Haynes, 2001) . Southwell (2013) also pointed out that individuals with more exposure to health misinformation have fewer chances or capabilities to correct such incorrect information. Building upon this line of work, we propose ehealth literacy as an important boundary condition for the effect of online misinformation on subsequent health behavioral decisions. eHealth literacy refers to the capability to obtain, process, and understand health information from electronic sources to make appropriate health decisions (Swire- Thompson and Lazer, 2020) . eHealth literacy is viewed as a process-oriented skill that advances overtime with the introduction of new technologies and changes in personal, social, and environmental contexts (Norman and Skinner, 2006b) . Studies have shown that low health literacy leads to delay or not getting health care, poorer overall health status and knowledge, less health promoting behaviors, and higher mortality rates (Institute of Medicine, 2004; Berkman et al., 2011) . Compared with high ehealth literates, those with low ehealth literacy level are less capable of discerning or analyzing the quality of online health information and its source (Norman and Skinner, 2006b) , making them more likely to believe in misinformation. Prior studies also emphasized the role of cognitive ability-those who have low levels of analytical thinking or numeracy were more vulnerable to being misinformed (De keersmaecker and Roets, 2017; Pennycook and Rand, 2019; Roozenbeek et al., 2020) . Also, with a smaller capacity to process information, low ehealth literates may suffer more from information overload upon exposure to misinformation online. Indeed, researchers have emphasized the importance of improving health information literacy to cope with information overload (Kim et al., 2007) . More importantly, low ehealth literates may find it more difficult to verify misinformation from other information sources; they may also have lower capability to obtain the needed health information online. These would make them more likely to engage in misinformed behaviors. We aimed to investigate the effects of online misinformation exposure on two outcomes: self-reported adoption of evidencebased practices as well as misinformed behaviors during the COVID-19 pandemic. Specifically, we predicted that exposure to online misinformation on COVID-19 prevention will (H1a) decrease evidence-based prevention practices and (H1b) increase misinformed behaviors. Guided by the protection motivation theory (Rogers, 1975) , and work on information overload (Schneider, 1987) , we proposed two specific mechanisms for this impact. First, we predicted that (H2a) exposure to online misinformation will increase misperception on prevention; and that (H2b) misperception will increase engagement in misinformed behaviors. Second, we hypothesized that (H3a) exposure to online misinformation will increase information overload and that (H3b) information overload will reduce engagement in evidence-based prevention practices. Finally, this study posed a research question to examine the moderating role of ehealth literacy between the paths proposed in H1-H3: "Does the effect of online misinformation exposure differ based on levels of ehealth literacy?" (RQ1) Figure 1 summarizes our hypotheses. A three-wave online panel survey was conducted with Singaporeans recruited from an online panel managed by Qualtrics. Singapore, an urban, developed country in Southeast Asia, is one of the countries hit hard by COVID-19 in the early stages of the pandemic. The Wave 1 survey (N = 1,023 out of 1,182 who initiated) was conducted in February 2020 while the Wave 2 survey was conducted in March 2020 (N = 767 out of 827 who started; retention rate of 75.0%). In April 2020, 540 participants completed the Wave 3 survey (retention rate of 70.4% from Wave 2). The survey was administered in English and took about 15 minutes to complete. Each wave collected data for 2 weeks, and Qualtrics sequentially sent out email invitations to keep the time interval consistent across participants. Participants were incentivized based on Qualtrics' remuneration system. During the data collection period, the local cases of COVID-19 increased from 96 to 8,014 in Singapore, recording among the highest number of cases in Asia at that time. We employed quota sampling based on age, gender, and ethnicity to achieve a sample that resembles the national profile. This study was approved by the Institutional Review Board of the hosting university in Singapore. To ensure the temporal ordering among study variables, we assessed exposure to misinformation at Wave 1, information overload and misperception at Wave 2 and engagement in social distancing and misinformed behaviors at Wave 3. We identified four false claims on how to prevent COVID-19: (1) gargling with mouthwash, (2) eating garlic, (3) vaccination against pneumonia, and (4) regularly rinsing nose with saline. These claims spread online during the early stages of the pandemic and were clarified as false by World Health Organization (2020). Based on prior research on exposure to scanned health (mis) information (Niederdeppe et al., 2007; Tan et al., 2015) , the participants were asked how often they had heard in the past few weeks that each of the four behaviors can prevent COVID-19. For each false claim, we measured exposure from the following three online sources, separately (1 = not at all; 2 = a few times; 3 = several times: 4 = a lot of times): news app or website, social media app or website, medical or health websites. Following the approach taken in an earlier study on misinformation exposure (Tan et al., 2015) , responses were averaged to create a scale of exposure to each of the four pieces of misinformation ( To assess the feelings of being overwhelmed with information on COVID-19, we used eight items derived from Jensen et al. (2014) . Participants reported their agreement with each statement, such as, "There is not enough time to do all of the things recommended to prevent the COVID-19, " and "No one could actually do all of the COVID-19 recommendations that are given" (1 = strongly disagree; 5 = strongly agree; α = 0.87, M = 2.73, SD = 0.74). Misperception focused on the perceived efficacy of four misinformed behaviors. We thus asked to what extent For evidence-based practices, we focused on social distancing as this particular behavior was most actively promoted between Wave 2 and Wave 3 with the implementation of lockdown measures in Singapore (Yong, 2020) . Participants self-reported how often they engaged in social distancing in the past 2 weeks on a 5-point scale (1 = never; 5 = very often). We used two items based on the advisories released by Ministry of Health in Singapore (gov.sg, 2020): (1) staying home as much as you can, and (2) keeping a safe distance from others (Spearman-Brown coefficient = 0.71, M = 4.67, SD = 0.54). For each of the misinformation items, we assessed the corresponding behavior: (1) gargling with mouthwash, (2) eating garlic, (3) vaccination against pneumonia, and (4) regularly rinsing nose with saline. Participants self-reported whether they had engaged in each of the behaviors in the past 2 weeks to prevent COVID-19 (0 = no, 1 = maybe, 2 = yes; α = 0.81, M = 0.28, SD = 0.48). When participants answered "yes, " they were asked to report the total number of times they did the respective behavior (except vaccination): gargling with mouthwash (M = 8.93, SD = 9.53, range = 1-60), eating garlic (M = 5.94, SD = 5.77, range = 1-24), and rinsing nose with saline (M = 4.26, SD = 6.38, range = 1-28). We used an 8-item ehealth literacy scale from Norman and Skinner (2006a) . Participants reported their level of agreement with eight statements about their experience using the Internet for health information (1 = strongly disagree; 5 = strongly agree). Sample statements include: "I know what health resources are available on the Internet, " and "I know how to use the health information I find on the internet to help me" (α = 0.94, M = 3.80, SD = 0.68). We used structural equation modeling (SEM; AMOS 25) to estimate path coefficients while accounting for measurement errors by using latent variables (Aiken et al., 1994) . We followed a two-step process of latent path modeling, which examines a measurement model first and then a structural path model. SEM is also recommended for group comparisons with latent variables as it allows statistical testing for group differences in path coefficients (Cole et al., 1993; Aiken et al., 1994) . For group comparisons, we used the median split to categorize participants into either the high or low ehealth literacy group. All constructs were treated as latent variables with respective measurements. We used item parceling with a random algorithm, for a latent factor with more than six indicators, because parceling helps to remove theoretically unimportant noise (Matsunaga, 2008) . We employed the full information maximum likelihood (FIML) method to address missing data (Graham, 2009 ). We controlled for age, gender, education, and income in model testing. Participants were restricted to those aged 21 or older and they were on average 44 years old (SD = 12.43) and 51.7% male at Wave 1 ( Table 1 ). The majority of respondents were ethnic Chinese (80.4%), followed by 10.5% Malay, 6.4% Indian, 0.8% Eurasian and other race (2.1%). The median education attainment was university graduate and median household monthly income was in the range of SGD 6, [0] [1] [2] [3] [4] [5] [6] [7] 999 (equivalent to USD 4, 711) . The attrition rate differed between genders from Wave 2 to Wave 3, p = 0.03 and among those who received different levels of formal education from Wave 1 to Wave 2, p = 0.01. The robustness check using the balanced samples of who completed all waves found the same results as those using the imputed data employing FIML estimation. Using confirmatory factory analysis (CFA), we first validated the measurement model with all latent factors in the proposed model. A good model has a root mean square error of approximation (RMSEA) ≤ 0.06, a comparative fit index (CFI) ≥ 0.95, and a standardized root mean square residual (SRMR) < 0.08 (Hu and Bentler, 1999 ). The CFA model fitted satisfactorily (χ 2 /df = 3.28, CFI = 0.97, RMSEA = 0.047, SRMR = 0.029). As presented in Table 2 , standardized loadings for all factors ranged from 0.60 to 0.93 (Kline, 2011) . The composite reliabilities (CRs) of latent variables were all above 0.7 and the average variance extracted (AVE) values of the latent factors were all above 0.5 (Hair et al., 2010) . The square root of each construct's AVE was also greater than its correlation with other latent factors. Thus, the CFA model had sufficient reliability, as well as convergent and discriminant validity (see Table 3 ). The baseline model with all groups adequately explained patterns of association between latent constructs (χ 2 /df = 2.85, To compare the path coefficients in the model between high and low ehealth literacy groups (RQ1), we conducted a structural model invariance test with cross-validation. The model with metric invariance constraints (i.e., keeping the measurement loadings constant across groups) had a good fit: χ 2 /df = 2.20, CFI = 0.95, RMSEA = 0.03, SRMR = 0.046. Thus, the CFA model is valid across the two groups, allowing us to compare the strength of path coefficients using these measures between groups (Byrne, 2001) . Next, we compared between the baseline model and the models with cross-group equality constraints (i.e., keeping constant path coefficients between groups) using chi-square change as a statistical test. This procedure tests if model fit improves with a model that allows a path coefficient to change between groups. Thus, significant model fit improvement would be interpreted as having different path coefficient strengths between the low and high ehealth literacy groups. The effect of online misinformation exposure at Wave 1 on social distancing at Wave 3 was stronger in the low ehealth literacy group (β = −0.20, p < 0.001) than the high ehealth literacy group (β = −0.12, p = 0.07), χ 2 (1) = 6.85, p = 0.001. Also, misperception at Wave 2 had a stronger effect on the performance of misinformed behaviors at Wave 3 in the low ehealth literacy group (β = 0.34, p < 0.001) than in the high ehealth literacy group (β = 0.17, p = 0.001). However, this group difference did not reach the statistical significance, χ 2 (1) = 3.34, p = 0.068. No other paths in the proposed model differed by ehealth literacy (Table 4) . This study found that exposure to online misinformation reduced self-reported engagement in social distancing and increased misinformed behaviors. This effect was partly explained by greater misperception and information overload triggered by online misinformation exposure. Misperception increased subsequent misinformed behaviors, while information overload reduced social distancing. Moreover, the effects of misinformation exposure differed by individuals' eheath literacy level. We discuss theoretical and practical implications of these key findings. Exposure to online misinformation on COVID-19 prevention not only increased self-reported engagement in behaviors advocated in the misinformation, but it also discouraged engagement in social distancing, which had been reinforced by health and government authorities during the pandemic. This suggests that misinformation exposure can discourage the adoption of evidence-based practices even when the content is not specifically on those practices. It is noteworthy that the two behavioral measures were not correlated (p = 0.70), indicating that they are independent from one another and that engaging in one behavior type may not compensate nor negate the other behavior. This echoes the differential mechanisms that led to the two different behavioral types found in this study. Consistent with prior research (e.g., Nyhan and Reifler, 2010; Kishore et al., 2011; Wang et al., 2019) , we found that exposure to misinformation can lead to misperception; in this specific context, individuals frequently exposed to online misinformation about preventive behaviors were more likely to believe in the efficacy of these misinformed behaviors at preventing COVID-19. Belief may be a default position when people encounter new information, and that disbelief requires mental effort, which people often do not bother to exercise when using online platforms (Pennycook and Rand, 2019) . This is of concern given that some misinformation on COVID-19 promote fake remedies that can bring about detrimental health outcomes as seen in the cases of ingesting methanol and disinfectant to prevent COVID-19 (Associated Press, 2020; Slotkin, 2020) . We conceptualized misperception as perceived response efficacy of preventive measures based on PMT (Rogers, 1975) , Standardized coefficients **denotes p < 0.01, ***p < 0.001, † p < 0.10. and, as the theory predicts, we found that misperception on COVID-19 prevention subsequently increased the likelihood of misinformed behaviors. Furthermore, a significant indirect effect of online misinformation exposure was found on prompting self-reported misinformed behaviors via cultivating misperception. Similarly, Tan et al. (2015) found that misperception on cancer causes, conceptualized as outcome expectancies, predicted intention to engage in cancer-related behaviors. Both response efficacy and outcome expectancies have been suggested as key motivators for health behaviors (Floyd et al., 2006; Fishbein and Ajzen, 2010) . This highlights the importance of correcting misperception by addressing its scientifically ungrounded nature before individuals act on their incorrect beliefs. However, considering the difficulty in rectifying misperception (Lewandowsky et al., 2012; Pluviano et al., 2017) , it would be also critical to minimize the public's exposure to false claims on the internet. It may as well be important to build a critical mindset in dealing with online health information via inoculation or other prebunking strategies (van der Linden et al., 2020). We also found that online misinformation exposure increased information overload. The uncertain and novel nature of misinformation on COVID-19 prevention might have been difficult to process, which made individuals feel more overloaded with information (Schneider, 1987) . Information overload also reduced self-reported compliance to social distancing measures in line with prior research that found poorer health status and less engagement in cancer screening due to information overload (Misra and Stokols, 2011; Jensen et al., 2014) . Objective information on preventive measures may become less influential when people experience information overload with the difficulty in identifying and selecting important information (Eppler and Mengis, 2004) . Because failing to implement evidence-based practices can exacerbate the spread of a disease, we call for more research on the link between misinformation and information overload and other possible mechanisms to better explain the effects of online misinformation. This study also explored the possible moderating role of ehealth literacy in how online misinformation exposure leads to self-reported preventive behaviors. We found one path in the proposed model that significantly differed between the low and high ehealth literacy groups. Specifically, the negative effects of online misinformation exposure on engaging in social distancing was stronger among those with lower levels of ehealth literacy than higher ehealth literates. Similarly, misperception also had a stronger, albeit marginal, effect on self-reported misinformed behaviors in the low ehealth literacy group. Collectively, it appears that online misinformation cultivates misperception regardless of ehealth literacy; however, high ehealth literates (vs. low literates) are less likely to act on their misperceptions perhaps because they are more capable of verifying and correcting their misperceptions over the course. This points to the value of ehealth literacy in empowering individuals in navigating an online information environment that has been polluted by different forms of healthrelated misinformation. There are several limitations of this study worth noting. First, we focused on examining the impact of exposure to misinformation, but not exposure to correct information nor exposure to mixed messages while individuals often encounter information that includes both true and false claims (Brennen et al., 2020; Tandoc and Mak, 2020) . Also, given that new scientific evidence and misinformation emerge as the pandemic develops, future work should examine other types of misinformation (e.g., on COVID-19 vaccines) to replicate our findings and to identify the most critical types of online misinformation during the pandemic. Second, we relied on self-reported measures of exposure to online misinformation and behaviors, which means our data relied heavily on the capability and willingness of participants to correctly recall and self-report their exposure levels and behavior. Given the lack of empirical evidence on how misinformation influences the way people actually behave, future studies should employ observational designs to assess preventive as well as misinformed behaviors. While our exposure measures had good reliability and convergent/discriminant validity in the current study, future studies could develop and adopt alternative measures or manipulation of information exposure to avoid issues with self-report. Because we averaged misinformation exposure across three different online sources for the sake of model parsimony, future work could also consider treating this factor differently, for example, by modeling misinformation exposure separately by source type. Third, we cannot confirm causal relations between study variables even with the temporal ordering established with the three-wave panel design. We collected data in the middle of a rapidly changing pandemic, which also involved constant updates on preventive measures advocated by health authorities. Such changes made it difficult for us to constantly assess study variables across three waves and control for factors measured in earlier waves, which would have provided more convincing causal evidence on misinformation effect. Also, our data only reflects the earlier stages of the outbreak in Singapore; nonetheless, we believe that it is crucial to examine misinformation effects during early stages of an actual pandemic, especially one that involves a novel disease, given that the spread of online misinformation is more frequent in such case as the public and scientific community struggle to figure out what it is and how to deal with the novel outbreak. Using a three-wave panel survey, this study offered some evidence that online misinformation exposure can lead to the public's maladaptive behaviors during a disease pandemic. Furthermore, we addressed two types of behavioral responses to the pandemic with differential mechanisms through which exposure to online misinformation could prompt those behavioral responses. This study also provided initial evidence on the impact of online misinformation on information overload beyond misperception; thus, this study informs further theory development in online misinformation exposure and effects. Lastly, we identified ehealth literacy as a potential boundary condition for the adverse consequences of online misinformation exposure, which highlights the importance of health literacy education to fight the growing problem of misinformation online. The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation. The studies involving human participants were reviewed and approved by the Institutional Review Board of Nanyang Technological University. The participants provided their written informed consent to participate in this study. Structural equation analyses of clinical subpopulation differences and comparative treatment outcomes: characterizing the daily lives of drug addicts Social media and fake news in the 2016 election Coronavirus: In Iran, The False Belief That Toxic Methanol Fights Covid-19 Kills Hundreds Fake news and the economy of emotions Low health literacy and health outcomes: an updated systematic review Types, Sources, And Claims of COVID-19 Misinformation Structural Equation Modeling With AMOS: Basic Concepts, Applications, And Programming The effect of information load on decision makers' cue utilization levels and decision quality in a financial distress decision task Consumer health information seeking on the Internet: the sate of the art Multivariate group comparisons of variable systems: MANOVA and structural equation modeling Fake news': incorrect, but hard to correct. The role of cognitive ability on the impact of false information on social impressions Zika vaccine misperceptions: a social media analysis Too good to be true, too good not to share: the social utility of fake news The concept of information overload: a review of literature from organization science, accounting, marketing, MIS and related disciplines Political elites' use of fake news discourse across communications platforms Predicting And Changing Behavior: The Reasoned Action Approach COVID-19 Circuit Breaker: Heightened Safe-Distancing Measures To Reduce Movement Missing data analysis: making it work in the real world Less than you think: prevalence and predictors of fake news dissemination on Facebook Multivariate Data Analysis, 7th Edn Feeling "disinformed" lowers compliance with COVID-19 guidelines: evidence from the US, UK, Netherlands and Germany Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives Health Literacy: A prescription to end confusion The cancer information overload (CIO) scale: establishing predictive and discriminant validity Propaganda, fake news, and fake trends: the weaponization of twitter bots in the gulf crisis Information overload and the message dynamics of online interaction spaces: a theoretical model and empirical exploration Effects of quality and quantity of information on decision effectiveness Effects of COVID-19 misinformation on information seeking, avoidance, and processing: a multicountry comparative study Predictors of cancer information overload Myths, beliefs and perceptions about mental disorders and health-seeking behavior in Delhi Principles And Practice Of Structural Equation Modeling The science of fake news Associations Between COVID-19 misinformation exposure and belief with COVID-19 knowledge and preventive behaviors: cross-sectional online study Misinformation and its correction continued influence and successful debiasing Measuring the impact of COVID-19 vaccine misinformation on vaccination intent in the UK and USA Item parceling in structural equation modeling: a primer Theoretical foundations of campaigns The magical number seven, plus or minus two: some limits on our capacity for processing information Information input overload and psychopathology Psychological and health outcomes of perceived information overload Examining the dimensions of cancer-related information seeking and scanning behavior EHealth literacy: essential skills for consumer health in a networked world EHEALS: the eHealth literacy scale When corrections fail: the persistence of political misperceptions Individuals and information overload in organizations: is more necessarily better? Acad Ebola, twitter, and misinformation: a dangerous combination Health literacy: a policy challenge for advancing high-quality health care Lazy, not biased: susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning Misinformation lingers in memory: failure of three pro-vaccination strategies Misinformation: a threat to the public's health and the public health system A protection motivation theory of fear appraisals and attitude change Susceptibility to misinformation about COVID-19 around the world Information overload: a temporal approach Information overload: causes and consequences NYC Poison Control Sees Uptick In Calls After Trump's Disinfectant Comments Social Networks And Popular Understanding Of Science And Health: Sharing Disparities Misinformation among mass audiences as a focus for inquiry Public health and online misinformation: challenges and recommendations Exposure to health (mis)information: lagged effects on young adults' health behaviors and potential pathways The facts of fake news: a research review Forwarding a WhatsApp Make Sure You Don't Spread MisinformationMessage on COVID-19 news? How to . Channel News Asia Autism and measles, mumps, and rubella vaccine: no epidemiological evidence for a causal association De Blasio Administration's Health Department Declares Public Health Emergency Due To Measles Crisis WHO Says Fake Coronavirus Claims Causing 'Infodemic The paradox of participation versus misinformation: social media, political engagement, and the spread of misinformation Inoculating against fake news about COVID-19 Viral misinformation threatens public health RETRACTED: leallymphoid-nodular hyperplasia, non-specific colitis, and pervasive developmental disorder in children A meta-analytic examination of the continued influence of misinformation in the face of correction: how powerful is it, why does it happen, and how to stop it? Is this the era of misinformation yet? Combining social bots and fake news to deceive the masses Systematic literature review on the spread of health-related misinformation on social media European Region Loses Ground In Effort To Eliminate Measles Coronavirus disease (COVID-19) Advice For The Public: Myth busters Coronavirus: More People Following Safe Distancing Measures, Fewer Than 70 Caught Flouting These Rules. The Straits Times HK and ET contributed to conception and design of the study. HK performed the statistical analysis and wrote the first draft of the manuscript. ET wrote sections of the manuscript. Both authors contributed to manuscript revision, read, and approved the submitted version. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.Publisher's Note: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.Copyright © 2022 Kim and Tandoc. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.