key: cord-0806271-ae89tpt2 authors: MacFarlane, Douglas; Tay, Li Qian; Hurlstone, Mark J.; Ecker, Ullrich K.H. title: Refuting Spurious COVID-19 Treatment Claims Reduces Demand and Misinformation Sharing date: 2020-12-29 journal: J Appl Res Mem Cogn DOI: 10.1016/j.jarmac.2020.12.005 sha: 16d6b70666cb1cce829aa76a54c7a7d27afd8fcf doc_id: 806271 cord_uid: ae89tpt2 The COVID-19 pandemic has seen a surge of health misinformation, which has had serious consequences including direct harm and opportunity costs. We investigated (N = 678) the impact of such misinformation on hypothetical demand (i.e., willingness-to-pay) for an unproven treatment, and propensity to promote (i.e., like or share) misinformation online. This is a novel approach, as previous research has used mainly questionnaire-based measures of reasoning. We also tested two interventions to counteract the misinformation, contrasting a tentative refutation based on materials used by health authorities with an enhanced refutation based on best-practice recommendations. We found prior exposure to misinformation increased misinformation promotion (by 18%). Both tentative and enhanced refutations reduced demand (by 18% and 25%, respectively) as well as misinformation promotion (by 29% and 55%). The fact that enhanced refutations were more effective at curbing promotion of misinformation highlights the need for debunking interventions to follow current best-practice guidelines. the media, reducing both demand for the product and the propensity to share related misinformation online. Sharing Misinformation-defined here as any information that is false-represents a threat to societies that value evidence-based practice and policy making (Lazer et al., 2018; . Smear tactics and political fake news have affected voter attitudes (Landon-Murray, Mujkic, & Nussbaum, 2019; Vaccari & Morini, 2014) , vaccine myths have contributed to the re-emergence of measles (Poland, Jacobson, & Ovsyannikova, 2009) , and climate change misinformation continues to pose a barrier to mitigative action (Cook, Ellerton, & Kinkead, 2018) . COVID-19 is another case in point. At a time when demand for high-quality information has been pronounced, online misinformation has proliferated-both as a result of the rapidly-developing nature of the pandemic, where even the best available evidence of today may be invalidated tomorrow, and due to the actions of ill-informed or malicious disinformants. COVID-19 misinformation has included claims that the virus originated as a bioweapon, or that health supplements or even onions could be effective treatments (e.g., Dupuy, 2020) . The spread of such misinformation can fuel hostility towards groups perceived to be responsible for the pandemic (Devakumar, Shannon, Bhopal, & Abubakar, 2020) ; it can also draw people towards remedies that are unproven, ineffective, or even harmful, and prevent uptake of superior interventions (e.g., using a vitamin supplement to prevent viral J o u r n a l P r e -p r o o f infection while abstaining from social distancing, mask use, or vaccination), at variance with official health advice Van Bavel et al., 2020) . Accordingly, the World Health Organization (WHO) noted that the pandemic has been accompanied by an infodemic, characterized by a deluge of false information amplified by modern communication technologies (Zarocostas, 2020) . Given the scale of the problem, considerable efforts are being undertaken to fact-check common claims (e.g., Therapeutic Goods Administration, 2020; Weule, 2020) . However, it is unclear how successful these efforts will be, and to what extent individuals initially misled by misinformation will adjust their beliefs and behaviours after receiving a refutation. Cognitive psychology research suggests refuting misinformation is no easy task. Even clearly corrected misinformation often continues to influence memory, reasoning, and decision making--the so-called continued influence effect (Chan, Jones, Hall Jamieson, & Albarracín, 2017; Lewandowsky, Ecker, Seifert, Schwarz, & Cook, 2012; Walter & Tukachinsky, 2020) . Post-correction reliance on misinformation can occur despite demonstrable memory for the correction; it can even arise with fictitious or trivial information. This suggests that continued influence is a cognitive effect that does not necessarily depend on individuals' motivation-it can emerge from failures of memory updating or retrieval processes alone (Rich & Zaragoza, 2016; Swire, Ecker, & Lewandowsky, 2017) . Much of the literature on continued influence has focused on how misinformation is best counteracted (Ecker, O'Reilly, Reid, & Chang, 2019; Paynter et al., 2019; Walter & Tukachinsy, 2020) . Based on this research, several best-practice recommendations have been identified. Specifically, refutations should: (a) come from a trustworthy source (Ecker & Antonio, 2020; Guillory & Geraci, 2013) and, if applicable, discredit the disinformant by exposing their hidden agenda (Walter & Tukachinsky, 2020) ; (b) make salient the J o u r n a l P r e -p r o o f discrepancy between false and factual information, which has been shown to facilitate knowledge revision (Ecker, Hogan, & Lewandowsky, 2017; Kendeou, Butterfuss, Kim, & van Boekel, 2019) ; (c) explain why the misinformation is false, providing factual information to replace the false information in people's mental models (Paynter et al., 2019; Swire et al., 2017) ; and (d) draw attention towards any misleading strategies employed by misinformants (Cook et al., 2018; MacFarlane, Hurlstone, & Ecker, 2018 , 2020a . Regrettably, real-world refutation attempts often do not take into account these recommendations. For example, it is common for scientists and official bodies to adopt a tentative, "diplomatic" approach in their communication (MacFarlane et al., 2018 (MacFarlane et al., , 2020a Paynter et al., 2019) . To illustrate, to counter misinformation regarding the use of vitamin C for COVID-19, the Australian Therapeutic Goods Administration (TGA) wrote, "there is no robust scientific evidence to support the usage of [high dose vitamin C] in the management of COVID-19" (TGA, 2020) . Such a diplomatic approach to refutation may sometimes be called for to express nuance, and this may be of particular concern during a pandemic characterized by much uncertainty (e.g., Chater, 2020) . However, it is important to recognize that tentative refutations leave open the possibility that misleading information may turn out to be true, which can then be exploited by individuals or organizations with ulterior motives (MacFarlane, Hurlstone, & Ecker, 2020b; Oreskes & Conway, 2010) . It is therefore clear that refutations of health misinformation need to be well-designed to both provide an accurate reflection of the best-available evidence and achieve the desired outcome, while also taking into account the cognitive biases that disinformants-such as health fraudsters trying to sell a product-seek to exploit. Arguably the greatest weakness of existing research into the continued influence effect is its near-universal reliance on questionnaire measures of reasoning. This is an issue because some of the most harmful effects of misinformation may be on behaviours. For this J o u r n a l P r e -p r o o f reason, a more consequential indicator of the success of a particular refutation strategy is consumer behaviour. The need for objective behavioural indicators is particularly acute in the context of COVID-19, where the combination of high anxiety levels and need for effective treatment may foster misinformation-driven demand for remedies that are ineffective and often dangerous, including bleach, methylated spirits, or essential oils (e.g., Spinney, 2020; note, however, that for ethical reasons we selected a product for the current study that is unproven as a COVID-19 treatment but relatively safe). Given the well-established attitudebehaviour gap (e.g., McEachan, Conner, Taylor, & Lawton, 2011) , it is critical that studies take steps towards investigating how misinformation affects behaviour, and to what extent corrections can undo those effects (Hamby, Ecker, & Brinberg, 2020) . Against this backdrop, and building on previous theoretical work (MacFarlane et al., 2020b), our aim was to investigate the following two questions: (1) How does exposure to COVID-19 misinformation influence people's (hypothetical) willingness to pay for unproven health products, as well as their subsequent propensity to spread misinformation? (2) Are enhanced refutations based on best-practice psychological insights better suited to counteract the behavioural impacts of misinformation, compared to the tentative refutations often used by authorities? To this end, we exposed participants to misinformation 1 suggesting that high-dose vitamin E can reliably prevent and cure COVID-19. Although vitamins are essential for health, there is ample evidence that vitamin supplementation is contra-indicated in individuals with no pre-existing deficiencies (Guallar 2016) , and the present claim that vitamin E will reliably cure COVID-19 is demonstrably false (Shakoor et al., 2021) . The misinformation employed a number of deceptive techniques used by real-life disinformants (MacFarlane et al., 2020b) . We then provided participants with either no refutation, a tentative refutation based on reputable real-world sources, or an enhanced refutation based on best-practice psychological insights (MacFarlane et al., 2020a (MacFarlane et al., , 2020b Paynter et al., 2019; Walter & Tukachinsky, 2020 ). An additional control group received only an article that provided general information about vitamin E. Instead of using the questionnaire approach typical of misinformation debunking studies, we examined participants' willingness to pay for a vitamin E supplement in a hypothetical auction, and their propensity to promote a mock social media message endorsing the use of vitamin E to treat COVID-19. We hypothesized that misinformation exposure would increase willingness to pay and misinformation promotion, and that the enhanced refutation would more effectively reverse this effect than the tentative refutation. The experiment adopted a between-participants design with four conditions (control vs. misinformation vs. tentative refutation vs. enhanced refutation), with two dependent variables: "willingness-to-pay"-the amount bid on the vitamin E supplement-and "misinformation-promotion"-a score derived from a participant's engagement with the misleading social media post. Participants' general attitudes towards health supplements and alternative medicines, as well as their concern about the COVID-19 pandemic were used as covariates; additional predictors included frequency of past vitamin E supplement consumption, belief in the effectiveness of routine vitamin E supplementation, and current state of health. A sample of 680 U.S.-based adult participants were recruited via Amazon's Mechanical Turk (MTurk). To ensure data quality, participants had to meet eligibility criteria of 97% approval rate and a minimum of 5,000 prior tasks completed. An a-priori power J o u r n a l P r e -p r o o f analysis suggested a minimum sample size of 309 to detect an effect of f = .16 (based on the contrast between control and "full-contingency plus" conditions in MacFarlane et al., 2018) in our primary contrast (misinformation vs. enhanced refutation) with df = 1, α = .05, 1 -β = .80. Total sample size of 680 was calculated by extrapolating this to incorporate four conditions and an estimated 10% exclusion rate. Participants were randomly allocated to one of the four intervention conditions, subject to the constraint of approximately equal cell sizes. Two a-priori (though not formally pre-registered) exclusion criteria were applied to remove careless responders. One participant was excluded for giving non-differentiated answers to every question in a survey block, and another for responding erratically, with overly inconsistent responses between pairs of equivalent questions (i.e., an odd/even threshold of > 2 Likert points apart). Final sample size was thus N = 678 (344 females, 328 males, 3 non-binary, 3 participants of undisclosed gender; age M = 42.81, SD = 12.64). Predictors. Prior beliefs can moderate responses to health messaging (Myers, 2014) . Thus, participants' general attitudes towards health supplements and alternative medicines were assessed using an 18-item questionnaire, following MacFarlane et al. (2018) . Each item consisted of a declarative statement relating to a motivation for consuming alternative health products (e.g., "Vitamins are natural, and supplements are therefore safe"). Participants were asked to rate how much they agreed with each statement using a 5-point Likert scale (1 = strongly disagree; 5 = strongly agree). A composite score was calculated for each participant indicating their general attitude to health supplements and alternative medicines (hereafter, "general-attitude"). To measure response consistency, each item was paired with a reversephrased statement of similar meaning (i.e., 9 pairs of items). The order of items in the scale was randomized. The general-attitude scale was found to have very good internal consistency J o u r n a l P r e -p r o o f reliability (Cronbach's α = .87). The full scale can be found in the Supplement, which is available at https://osf.io/p89bm/. Participants' concern about the COVID-19 pandemic was measured as a composite score based on participants' responses to four questions (e.g., "How severe do you think novel coronavirus will be in the U.S. general population as a whole?" on a 5point Likert scale [1 = very mild; 5 = very severe]), following Ecker, Butler, Cook, Hurlstone, Kurz, & Lewandowsky (2020) . All items are provided in the Supplement. 2 Interventions. Dependent on condition, participants were presented with one or two articles (all articles are provided in Supplement). The control condition used a news article emphasizing the importance of a healthy immune system for fighting COVID-19. The article was modelled on two pieces (Collins, 2020; Weule, 2020) from reputable and independent media outlets, The Conversation and the Australian Broadcasting Corporation. It highlighted the importance of maintaining a balanced diet, regular exercise, and good sleep. To provide a background as to why participants were bidding on a vitamin E supplement, the article included a quote from the U.S. National Institute of Health "Vitamins and Minerals" website, "Vitamins and minerals are…essential for health. The lack of a specific micro-nutrient may cause…disease. Vitamin E, for example, acts as an antioxidant and is involved in immune function. It helps to widen blood vessels and keep blood from clotting. In addition, cells use Vitamin E to interact with each other and to carry out many important functions." The control article was therefore not fully neutral but contained some allusion that vitamin E supplementation might provide some protection from viral infection. The article also noted that whilst some people believe that vitamin supplements are essential, dietitians say that taking supplements is not necessary unless you have a specific nutrient deficiency or dietary need. The article in the misinformation condition additionally contained false claims suggesting vitamin E can protect against COVID-19. These were modelled on misinformation spread via social media during the early stages of the pandemic (Bogle, 2020; RMIT Fact Check, 2020; Weule, 2020) . We invented a fake expert, Dr. Avery Clarke, and used several deceptive techniques employed by real-life fraudsters (MacFarlane et al., 2020b; see Table 1 for details). In the tentative-refutation condition, the misinformation article was followed by a second article that highlighted the lack of evidence for claims about vitamin E and COVID-19. It was based on a tentative, "diplomatic" refutation style commonly used in real-world attempts to debunk claims about ineffective or unproven health products; specifically, it was modelled on an article featuring interviews with a disease expert and a dietitian (Weule, 2020) . We invented a public health expert, Prof. Simon Corner, and outlined his response to Dr. Avery's claims, "There is no conclusive evidence that vitamin [E] supplements can delay the onset of an infection or treat respiratory infections, such as COVID-19. Furthermore, vitamin and mineral supplements are not recommended for the general population." The final paragraph noted that there were some exceptions (e.g., people with specific vitamin deficiencies) and recommended people talk to a doctor, pharmacist, or accredited dietitian. In the enhanced-refutation condition, the misinformation article was followed by a refutational article that, instead of simply noting lack of evidence, drew attention to deceptive and misleading techniques used in the misinformation article and explained how those techniques were used to deceive readers. This enhanced refutation was based on psychological insights regarding the effective debunking of misinformation (MacFarlane et al., 2020a (MacFarlane et al., , 2020b Paynter et al., 2019; Walter & Tukachinsky, 2020 ; see Table 2 for details). J o u r n a l P r e -p r o o f Dependent measures. We included two novel tasks to measure participants' hypothetical willingness to pay for a spurious treatment and their propensity to spread related misinformation. Willingness-to-pay. To assess willingness-to-pay, we used an experimental auction (MacFarlane et al., 2018 ; also see Becker, DeGroot, & Marschak, 1964; Kagel & Levin, 2011; Thrasher, Rousu, Hammond, Navarro, & Corrigan, 2011) . The auction was hypothetical but previous experiments have shown that results obtained using this hypothetical auction format are comparable to those under fully-incentivized conditions (i.e., when bidding using real money for real products; MacFarlane et al., 2020a). The decision to use a hypothetical format was made for both practicality reasons and ethical concerns about selling an unproven COVID-19 treatment during the pandemic. First, participants were asked to imagine they had been given a $5 endowment. They then had an opportunity to place a bid on a bottle of 100 vitamin E capsules. Participants were shown a plain-packaged picture of the product and a generic text (e.g., "…The supplement is designed to be taken once a day."). It was explained that the auction was different from other auctions in that participants could only bid once, and it was in their best interest to bid the amount they were willing to pay for the product. Participants entered their bid amount b in cents, with b ∈ (0, 500). They knew this amount would be compared against a random number r ∈ (0, 500) drawn from a uniform distribution, and that if b ≥ r, they would win the auction and purchase the product for amount b but keep 500b of their endowment; otherwise they would lose the auction but keep the full hypothetical endowment. There was an initial practice auction to ensure participants understood the procedure. At the end of the auction, they were informed about the outcome, and were shown their bid amount, the random comparison amount, and their "take-home" amount. To mitigate hypothetical bias, we employed two established biasreduction techniques, namely "cheap-talk"-asking participants to behave as they would if J o u r n a l P r e -p r o o f the auction was real-and consequentiality-reminding participants the results of this study would have implications for significant public health issues. To assess participants' propensity to promote misinformation via social media, we developed a novel measure. Participants were presented with four social media posts, and they were asked to indicate how they would engage with each post using one of three options, namely "share", "like", or "flag". Each option was accompanied by a brief explanation: Participants were told they could "Share the post publicly, so more people could read it"; "Like the post, so your contacts see that you agree with it"; or "Flag the post if you think it is inappropriate (e.g., offensive, inaccurate, misleading, or promoting illegal activity)". Participants were also told they could choose not to interact with a post, thus enabling "pass" as a fourth response option. The posts are shown in Figure 1 . Three posts were on topics generally related to health and/or COVID-19, whereas the fourth contained an endorsement of misinformation, namely that vitamin E can treat COVID-19. Interest centred on responses to this target post, which was always presented last. The general posts were presented first (in random order) as decoys to help manage demand characteristics. To ensure plausibility, two of the decoy posts were real tweets from the Wall Street Journal and the World Health Organization. To avoid a slight confound that only the target post was from an individual doctor, the third decoy post (relating to a tweet from New Scientist) was also from an invented doctor. Ethics approval was granted by the Human Research Ethics Office of the University of Western Australia. The experiment was run using Qualtrics software (Qualtrics, Provo, UT). Participants were initially given an ethics-approved information sheet and provided informed consent. They then responded to questions on demographics, vitamin E consumption, the general-attitude scale, and the COVID-19 questions. Participants then read J o u r n a l P r e -p r o o f the article(s) associated with the condition they were assigned to (i.e., control, misinformation, tentative-refutation, or enhanced-refutation). This was followed by the hypothetical auction and the mock social media engagement task, in which participants bid for a bottle of vitamin E tablets and interacted with a series of posts (response options: share, like, flag, pass). Finally, participants were debriefed; this included a detailed refutation of the misinformation participants had encountered (following the procedure in the enhancedrefutation condition), and participants had to demonstrate their understanding by correctly answering a question regarding the gist of the debriefing. The experiment took approximately 17 minutes; participants were paid US$2.50. On a descriptive level, willingness-to-pay means across control (C), misinformation We conducted an analysis of covariance (ANCOVA) with a four-level condition factor, willingness-to-pay as the dependent variable, and general attitude towards health supplements (hereafter: general attitude) and COVID-19 concern as covariates. There was a significant main effect of condition, F(3,672) = 5.76, p < .001, ω 2 = .02 (see Figure 2) . There was also a significant effect of general attitude, F(1,672) = 150.45, p < .001, ω 2 = .18, indicating that greater support for supplements was associated with greater willingness-topay. There was no significant effect of COVID-19 concern, F < 1. Tukey post-hoc tests revealed mean willingness-to-pay was greater in the misinformation condition than the tentative-refutation condition (Δ = 43.86, 95% CI 4.73 -82.98, t = 2.89, p = .021, d = 0.28) and the enhanced-refutation condition (Δ = 59.46, t = 3.92, p < .001, d = 0.37) . There were no significant differences between the other conditions (see Table 3 for details). 3 In terms of participant proportions across response categories and conditions, 50.0% of participants in the control condition engaged in liking or sharing of the misinformationendorsing social media post. This increased to 59.1% in the misinformation condition (i.e., an 18.1% increase) but dropped to 41.7% and 26.6% in tentative-refutation and enhancedrefutation conditions, respectively (i.e., 29.5% and 54.9% reductions relative to the misinformation condition). See Table 4 for further details. To formally evaluate participants' propensity to promote the misinformation post, we converted the four response options into a single measure ("misinformation-promotion"). We re-coded each response option to reflect the progression towards an increasing propensity to spread misinformation: flag = -1 (actively combatting the spread of misinformation); pass = 0 (not contributing to the spread of misinformation); like = 1 (signalling approval of misinformation and increasing its salience); share = 2 (actively spreading misinformation). An asymptotic Kruskal-Wallis test found that condition significantly influenced misinformation-promotion, H(3) = 54.25, p < .001 (see Figure 3 ). 4 To assess the locus of condition differences, post-hoc Mann-Whitney U tests with Bonferroni adjustment were run. Misinformation-promotion was significantly greater for participants in the misinformation condition, Mdn = 1, compared to control, Mdn = 0.5, Ws = 12434, p = .017, r = -.09, the tentative-refutation condition, Mdn = 0, Ws = 17712, p < .001, r = -.15, and the enhanced-refutation condition, Mdn = -1, Ws = 20566, p < .001, r = -.27. Participants in the enhanced-refutation condition showed significantly lower misinformation-promotion than those in the tentative-refutation condition, Ws = 17386, p < .001, r = -.15, and the control condition Ws = 18494, p < .001, r = -.19. There was no significant difference between the control condition and the tentative-refutation condition, Ws = 15473, p = .168, r = -.05. The present study is one of the first to incorporate behavioural measures into the continued influence paradigm. Specifically, we tested whether exposure to misinformation that presented vitamin E as a potent remedy to prevent and cure COVID-19 affected (hypothetical) product demand and misinformation promotion behaviour. To test the efficacy of interventions counteracting the misinformation, we contrasted a tentative refutation based on real-world sources and an enhanced refutation based on best-practice recommendations. We included general attitude towards supplements and COVID-19 concern as covariates. We found that pre-existing attitudes predicted demand and propensity to promote misinformation, whereas we found no effects of COVID-19 concern. Even though this is perhaps surprising, we refrain from drawing strong conclusions based on this null finding; it is of course possible that degree of concern has little influence on the processing of related misinformation and subsequent decisions (similarly, Ecker, Lewandowsky, & Apai, 2010, reported limited impact of emotionality on the continued influence effect). More importantly, with regards to our experimental manipulation, we found that exposure to misinformation significantly increased participants' subsequent propensity to promote a mock social media post endorsing the misinformation. Although willingness to pay J o u r n a l P r e -p r o o f was not reliably affected by the misinformation relative to control, this is best explained by the control article's allusion that vitamins provide some protection from viral infection, meaning that the control condition arguably did not provide a fully-neutral comparison baseline. Compared to the misinformation condition, both refutation types substantially reduced willingness-to-pay and misinformation-promotion, which underscores the general utility of refutations beyond the realm of inferential reasoning measures. The enhanced refutation was more effective than the tentative refutation in reducing misinformation promotion, 5 reinforcing the best-practice recommendations used. Our results demonstrate the persuasive appeal of (COVID-19) misinformation and its potential to influence behaviours (Zarocostas, 2020 ; see also Pennycook, McPhetres, Zhang, Lu, & Rand, 2020) . They also demonstrate the value of an evidence-based approach to debunking. This is in line with previous research (Paynter et al., 2019; Swire et al., 2017) but one of the first demonstrations of this kind with behavioural measures (also see Hamby et al., 2020) . This is of particular relevance given (1) the need for studies to take steps towards addressing the attitude-behaviour gap (McEachan et al., 2011) , and (2) previous work questioning the utility of refutations when it comes to behaviours (Swire-Thompson, DeGutis, & Lazer, 2020; Swire-Thompson, Ecker, Lewandowsky, & Berinsky, 2020) . The practical implications are clear. Individuals exposed to refutations are less likely to waste money on spurious products. Also, individuals exposed to misinformation may be more likely to later on spread it to those around them, but this harmful sharing behaviour can be mitigated through refutations. Although sharing decisions may also be affected by factors other than veracity (e.g., inattention, worldview; Mercier; 2020; Pennycook et al., 2020) , the present study suggests that well-designed refutations are an indispensable tool. The fact that our enhanced refutation was more effective at reducing misinformation promotion illustrates the need for debunking practices to be based on insights from psychology (MacFarlane et al., 2020b ). This will be especially critical during crises such as the current infodemic. Nonetheless, limitations must be acknowledged. First, our enhanced refutation included many elements and it is unclear which ones contributed to the enhanced refutation's success. There may also be nuisance factors such as perceived confidence in evidence, memorability, and article length that differentiated tentative and enhanced refutations. Further research should thus include component analyses to aid the construction of specific best-practice guidelines (see also Lewandowsky et al., 2020; Paynter et al., 2019) . This would also help add theoretical nuance beyond the scope of the current study, furthering understanding of why certain interventions work better than others. Second, we relied only on a hypothetical auction and a mock social media scenario to assess behaviours. Although an important first step, willingness-to-pay methods may display substantial heterogeneity despite bias-reduction techniques (e.g., Kanya, Sanghera, Lewin, & Fox-Rushby, 2019; Voelckner, 2006) , and the same may apply to our social media scenario. It is therefore critical that future studies use measures that correspond with behaviours of interest even more closely (e.g., incentivized auctions, social-media simulations) and validate results using real-world data (e.g., Mosleh, Pennycook, & Rand, 2020) . Third, we did not find a difference in willingness-to-pay between the enhanced and tentative refutation conditions, suggesting that the auction measure may not have been sufficiently sensitive. Potentially related to this, we also did not find a reliable willingness-topay difference between misinformation and control conditions. However, consistent with effects of implied and subtle misinformation (e.g., Ecker, Lewandowsky, Chang, & Pillai, 2014; Powell, Keil, Brenner, Lim, & Markman, 2018; Rich & Zaragoza, 2016) , this may have also been driven by the insinuation in the control article that vitamins can provide some J o u r n a l P r e -p r o o f infection protection. This suggests that the framing of "neutral" educational communications by impartial media sources needs to be carefully considered. The fact that the control condition was non-neutral means there was no neutral no-misinformation baseline; this makes it impossible to conclude with certainty whether there was continued influence (i.e., above-baseline reliance on refuted misinformation) in either of the refutation conditions. Furthermore, there was no time lag between misinformation and correction, nor between interventions and behavioural tasks. This may constrain applicability of current findings for real-world scenarios; future research should explore the impact of longer time delays (see Paynter et al., 2019) . Finally, we recommend that communicators dealing with misinformation consider additional strategies beyond refutations. The speed of misinformation generation and spread presents a challenge to fact-checkers and their capacity to refute misleading content (Lazer et al., 2018) . Future work should thus examine inoculation strategies, which pre-emptively explain misleading techniques common to misinformation campaigns and thus increase resilience to misdirection (e.g., Comptom, 2013; Maertens, Roozenbeek, Basol, & van der Linden; van der Linden, Leiserowitz, Rosenthal, & Maibach, 2017) . In particular, more research is needed to assess the impact of inoculations on behaviours. Both the data and a Supplement containing the materials and supplementary analyses are available at https://osf.io/p89bm/. Author Contributions: All authors jointly developed the study concept and contributed to the study design. DM prepared the experimental survey, and performed data collection and analysis, with input from all other authors. LQT and UKHE performed the J o u r n a l P r e -p r o o f J o u r n a l P r e -p r o o f Debunk the appeal to nature "Using terms such as 'natural' is designed to get people to associate the proposed treatment with being harmless" Highlight key overlooked risks "high doses of Vitamin E supplements have been linked to adverse side-effects…and some more serious diseases" Debunk the appeal to morality "sharing such bogus remedies with your family would put them at risk of serious side-effects, while providing no benefit" Counteract the illusion of causality "the majority of people will recover from COVID-19 thanks to their own immune systems, irrespective of any dietary supplements or unproven remedies they may or may not have taken" J o u r n a l P r e -p r o o f J o u r n a l P r e -p r o o f Good News about Bad News: Gamified Inoculation Boosts Confidence and Cognitive Immunity Against Fake News Measuring utility by a singleresponse sequential method As coronavirus fears grow, family group chats spread support but also misinformation Debunking: A metaanalysis of the psychological efficacy of messages countering misinformation Facing up to the uncertainties of COVID-19 5 ways nutrition could help your immune system fight off the coronavirus. The Conversation Inoculation theory. The Sage Handbook of Persuasion: Developments in Theory and Practice Deconstructing climate misinformation to identify reasoning errors Neutralizing misinformation through inoculation: Exposing misleading argumentation techniques reduces their influence Racism and discrimination in COVID-19 responses No evidence onions can cure ailments or kill viruses Can you believe it? An investigation into the impact of retraction source credibility on the continued influence effect Using the COVID-19 economic crisis to frame climate change as a secondary issue reduces mitigation support Reminders and repetition of misinformation: Helping or hindering its retraction Terrorists brought down the plane!-No, actually it was a technical fault: Processing corrections of emotive information The effects of subtle misinformation in news headlines The effectiveness of shortformat refutational fact-checks Enough is enough: Stop wasting money on vitamin and mineral supplements Correcting erroneous inferences in memory: The role of source credibility How stories in memory perpetuate the continued influence of false information Auctions: A survey of experimental research The criterion validity of willingness to pay methods: A systematic review and meta-analysis of the evidence Knowledge revision through the lenses of the three-pronged approach Disinformation in contemporary U.S. foreign policy: Impacts and ethics in an era of fake news, social media, and artificial intelligence The science of fake news The Debunking Handbook Beyond misinformation: Understanding and coping with the post-truth era Misinformation and its correction: Continued influence and successful debiasing Reducing demand for ineffective health remedies: Overcoming the illusion of causality Countering demand for ineffective health remedies: Do consumers respond to risks, lack of benefits, or both? Psychology and Health Protecting consumers from fraudulent health claims: A taxonomy of psychological drivers, interventions, barriers, and treatments Long-term effectiveness of inoculation against misinformation: Three longitudinal experiments Prospective prediction of health-related behaviours with the theory of planned behaviour: A meta-analysis Not Born Yesterday: The Science of Who We Trust and What We Believe Self-reported willingness to share political news articles in online surveys correlates with actual sharing on Twitter Changing smokers' risk perceptions-for better or worse Defeating the merchants of doubt Evaluation of a template for countering misinformation-Real-world Autism treatment myth debunking Misinformation on social media: Experimental evidence for a scalable accuracynudge intervention Trends affecting the future of vaccine development and delivery: The role of demographics, regulatory science, the antivaccine movement, and vaccinomics Misleading health consumers through violations of communicative norms: A case study of online diabetes education The continued influence of implied and explicitly stated misinformation in news reports Susceptibility to misinformation about COVID-19 around the world Facebook post spruiking vitamin C as a coronavirus cure -Fact Check Bleach baths and drinking hand sanitiser: poison centre cases rise under Covid-19. The Guardian Immune-boosting role of vitamins D, C, E, zinc, selenium and omega-3 fatty acids: Could they help against COVID-19? The role of familiarity in correcting inaccurate information Searching for the backfire effect: Measurement and design considerations They might be a liar but they're my liar: Source evaluation and the prevalence of misinformation No evidence to support intravenous high-dose vitamin C in the management of COVID-19 Estimating the impact of pictorial health warnings and "plain" cigarette packaging: evidence from experimental auctions among adult smokers in the United States The power of smears in two American presidential campaigns Using social and behavioural science to support COVID-19 pandemic response An empirical comparison of methods for measuring consumers' willingness to pay. Marketing Letters A meta-analytic examination of the continued influence of misinformation in the face of correction: How powerful is it, why does it happen, and how to stop it? Can you boost your immune system to help fight coronavirus? Australian Broadcasting Corporation How to fight an infodemic. The Lancet