ArticlePDF Available

Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online

Authors:

Abstract and Figures

Recent research suggests that shifting users’ attention to accuracy increases the quality of news they subsequently share online. Here we help develop this initial observation into a suite of deploy-able interventions for practitioners. We ask (i) how prior results generalize to other approaches for prompting users to consider accuracy, and (ii) for whom these prompts are more versus less effec-tive. In a large survey experiment examining participants’ intentions to share true and false head-lines about COVID-19, we identify a variety of different accuracy prompts that su¬ccessfully increase sharing
Harvard Kennedy School Misinformation Review1
May 2021, Volume 2, Issue 3
Creative Commons Attribution 4.0 International (CC BY 4.0)
Reprints and permissions: misinforeview@hks.harvard.edu
DOI: https://doi.org/10.37016/mr-2020-71
Website: misinforeview.hks.harvard.edu
Research Article
Developing an accuracy-prompt toolkit to reduce COVID-19
misinformation online
Recent research suggests that shifting users’ attention to accuracy increases the quality of news they
subsequently share online. Here we help develop this initial observation into a suite of deployable
interventions for practitioners. We ask (i) how prior results generalize to other approaches for prompting
users to consider accuracy, and (ii) for whom these prompts are more versus less effective. In a large survey
experiment examining participants’ intentions to share true and false headlines about COVID-19, we
identify a variety of different accuracy prompts that successfully increase sharing discernment across a
wide range of demographic subgroups while maintaining user autonomy.
Authors: Ziv Epstein (1,2), Adam J. Berinsky (3), Rocky Cole (4), Andrew Gully (4), Gordon Pennycook (5,6), David G. Rand
(2,7)
Affiliations: (1) Media Lab, Massachusetts Institute of Technology, USA, (2) Sloan School of Management, Massachusetts Institute
of Technology, USA, (3) Department of Political Science, Massachusetts Institute of Technology, USA, (4) Google, USA, (5)
Hill/Levene Schools of Business, University of Regina, Canada, (6) Department of Psychology, University of Regina, Canada, (7)
Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, USA
How to cite: Epstein Z., Berinsky A. J., Cole, R., Gully, A., Pennycook, G., & Rand, D. G. (2021). Developing an accuracy-
prompt toolkit to reduce COVID-19 misinformation online. Harvard Kennedy School (HKS) Misinformation Review, 2(3).
Received: February 25th, 2021. Accepted: April 22nd, 2021. Published: May 18th, 2021.
Research questions
There is mounting evidence that inattention to accuracy plays an important role in the spread of
misinformation online. Here we examine the utility of a suite of different accuracy prompts aimed at
increasing the quality of news shared by social media users.
Which approaches to shifting attention towards accuracy are most effective?
Does the effectiveness of the accuracy prompts vary based on social media user characteristics?
Assessing effectiveness across subgroups is practically important for examining the
generalizability of the treatments and is theoretically important for exploring the underlying
mechanism.
Essay summary
Using survey experiments with N = 9,070 American social media users (quota-matched to the
1
A publication of the Shorenstein Center on Media, Politics, and Public Policy at Harvard University, John F. Kennedy School of
Government.
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 2
national distribution on age, gender, ethnicity, and geographic region), we compared the effect
of different treatments designed to induce people to think about accuracy when deciding what
news to share. Participants received one of the treatments (or were assigned to a control
condition), and then indicated how likely they would be to share a series of true and false news
posts about COVID-19.
We identified three lightweight, easily-implementable approaches that each increased sharing
discernment (the quality of news shared, measured as the difference in sharing probability of true
versus false headlines) by roughly 50%, and a slightly more lengthy approach that increased
sharing discernment by close to 100%. We also found that another approach that seemed
promising ex ante (descriptive norms) was ineffective. Furthermore, we found that gender, race,
partisanship, and concern about COVID-19 did not moderate effectiveness, suggesting that the
accuracy prompts will be effective for a wide range of demographic subgroups. Finally, helping to
illuminate the mechanism behind the effect, the prompts were more effective for participants
who were more attentive, reflective, engaged with COVID-related news, concerned about
accuracy, college-educated, and middle-aged.
From a practical perspective, our results suggest a menu of accuracy prompts that are effective in
our experimental setting and that technology companies could consider testing on their own
services.
Implications
The spread of inaccuracies on social media including political “fake news” (Lazer et al., 2018; Pennycook
& Rand, 2021) and COVID-19 misinformation (Pennycook, McPhetres et al., 2020) is a topic of great
societal concern and focus of academic research. Of particular importance is identifying approaches that
technology companies could directly use to combat online misinformation. Here we focus on reducing the
sharing of misinformation because simply being exposed to misinformation can increase subsequent
belief (Pennycook et al., 2018). Thus, it is especially important to prevent initial exposure. To that end, we
explore the effectiveness of shifting users’ attention toward accuracy. Recent work suggests that
misinformation sharing often occurs, not because people purposefully share news they know is
inaccurate, but because people are distracted or focused on other elements when deciding what to share.
As a result, shifting attention towards accuracy can help people attend to their existing but often latent
capacity and desire to discern truth from falsehood, for both political misinformation (Fazio, 2020;
Jahanbakhsh et al., 2021; Pennycook et al., 2021) and COVID-19 misinformation (Pennycook, McPhetres
et al., 2020). This approach is particularly appealing because it does not require technology companies to
decide (e.g., via machine learning or human moderators) what is true versus false. Rather, emphasizing
accuracy helps users exercise the (widely held; see Pennycook et al., 2021) desire to avoid sharing
inaccurate content, preserving user autonomy. This approach is also appealing from a practical
perspective because accuracy prompts are scalable (unlike, for example, professional fact-checking, which
is typically slow and only covers a small fraction of all news content; Pennycook, Bear et al., 2020).
Here, we advance the applicability of accuracy prompt interventions by asking which such
interventions are effective. We focused on COVID-19 misinformation, and began by replicating prior
findings (Pennycook et al., 2021; Pennycook, McPhetres et al., 2020) that, in the absence of any
intervention, headline veracity has little impact on sharing intentions despite participants being fairly
discerning when asked to judge the accuracy of the headlines. That is, there is a substantial disconnect
between accuracy judgments and sharing intentions.
We then compared the relative effectiveness of several different ways to induce people to think about
accuracy, all delivered immediately prior to the sharing task. We found that (i) asking participants to judge
Epstein; Berinsky; Cole; Gully; Pennycook; Rand 3
the accuracy of a non-COVID-19 related headline, (ii) providing minimal digital literacy tips, and (iii) asking
participants how important it was to them to share only accurate news, all increased sharing discernment
by roughly 50% (3 percentage points); and that (iv) asking participants to judge the accuracy of a series of
4 non-COVID-19related headlines (and providing corrective feedback on their responses) increased
sharing discernment by roughly 100% (6 percentage points), all by decreasing the sharing of false but not
true headlines. Conversely, (v) informing participants that other people thought it was important to share
only accurate news (providing “descriptive norm” information) was ineffective on its own but may have
increased the effectiveness of other approaches when implemented together.
From a practical perspective, these findings demonstrate that accuracy prompt effects are not unique
to the particular implementations used in prior work and provide platform designers with a menu of
effective accuracy prompts to choose from (and ideally cycle through, to reduce adaption and banner
blindness) when creating user experiences to increase the quality of information online. Our results also
suggest important directions for future research, such as assessing how long the effects last, how quickly
users become insensitive to repeated treatment, and how our results would generalize cross-culturally.
Additionally, further work is required to test the ecological validity of these interventions across different
online services (querying a search engine versus browsing a social media feed, for example) and other
content verticals where accuracy might be important.
From a theoretical perspective, we found that the treatments operate by reducing the disconnect
between accuracy judgments and sharing intentions observed in the baseline condition: By shifting
participants’ attention to accuracy, we increased the link between a headline’s perceived accuracy and its
likelihood of being shared. The observation that approaches (i), (ii), and (iii) were equally effective sheds
further light on the mechanism driving these effects. In particular, the tips (ii) were no more effective than
merely having participants make an accuracy judgment (i), suggesting that the treatments work almost
entirely by priming accuracy, rather than via knowledge transmission per se. Explicitly stating a
commitment to accuracy (iii) was no more effective than the simple accuracy prime (i), indicating that the
mechanism is not commitment per se, either. This is promising from an applied perspective, as it suggests
that simply shifting attention to accuracy can be effective in and of itself, and that this can be done in a
wide variety of ways.
We also examined whether the effect of these treatments varied based on numerous user
characteristics. We found various moderators that support our proposed mechanism. Foremost, the
treatment was more effective for participants who were more attentive, consistent with the idea that one
must notice the treatments in order to be affected. The effect was also stronger among people who placed
greater importance on sharing only accurate news, consistent with the idea that shifting attention to
accuracy should increase sharing discernment only insofar as the user actually cares about accuracy (as
formalized by a limit-attentioned utility model in Pennycook et al. [2021]). Finally, the effect was stronger
among those who engaged in more analytic thinking, self-reported engaging with COVID-related news to
a greater extent, and were college-educated, all of which are consistent with the idea that these users
seem likely to have a greater knowledge base upon which to draw when evaluating the headlines (or a
greater ability to do so); although importantly the treatment still significantly increased sharing
discernment among less analytic, less COVID-19-news engaged, and non-college educated participants.
These all support our interpretation that our treatment effect, across conditions, is driven by increased
attention to accuracy as does a headline-level analysis finding that the effect of the treatments on
sharing of a given headline is strongly correlated with the perceived accuracy of that headline.
Furthermore, we found no evidence that the treatment effect on sharing discernment varied
significantly based on gender, race, partisanship, or concern about COVID-19; and although the treatment
effect did vary non-linearly with age (such that it was most effective for middle-aged participants), sharing
discernment was significantly increased for all age groups (1834, 3550, 5064, 65+). From a practical
perspective, these results suggest that accuracy prompts are likely to work for diverse groups of users,
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 4
even those who may approach sharing in a motivated way (e.g., those who are not concerned about
COVID-19 or who may be motivated to downplay COVID risks). This indicates that the intervention is
widely applicable.
Together, our results help to inform social media platforms, civil society organizations, and policy
makers about how to most effectively prompt users to consider accuracy online. We hope that our work
will help guide efforts to reduce the spread of misinformation online.
Findings
A disconnect exists between accuracy judgments and sharing intentions
Consistent with past work, we found that sharing intentions were much less discerning than accuracy
judgments; see Figure 1. When asked to judge the accuracy of 20 COVID-19 related headlines (half true,
half false), participants displayed a fair level of discernment: True headlines were much more likely to be
rated as accurate (67.7%) than false headlines (40.9%; p < .0001). When a separate group of participants
recruited at the same time were instead asked if they would share the same set of headlines online, the
results were strikingly different: Veracity had no significant impact on sharing intentions (True headlines
54.1%, False headlines 48.3%, p = .19; interaction between condition and veracity, p <.0001).
Figure 1. Disconnect between accuracy judgments and sharing intentions. Fraction of true versus false headlines rated as
accurate in the Accuracy-Only condition (“To the best of your knowledge, is the claim in the above headline accurate?”) and
considered for sharing in the Sharing condition (“Would you consider sharing this story online (for example, through Facebook or
Twitter)?”) of survey wave 1. Error bars indicate 95% confidence intervals.
Various accuracy prompts increase sharing discernment
We then assessed the impact of eight different experimental treatments on sharing intentions of true
versus false headlines (sample materials are shown in Figure 2). Each treatment was administered prior
to the beginning of the news sharing task, as follows:
1. In the "Evaluation" treatment, as in Pennycook et al. (2021) and Pennycook, McPhetres et al.
(2020), participants were asked to evaluate the accuracy of a single non-COVID-related headline
0%
10%
20%
30%
40%
50%
60%
70%
80%
Accuracy
N=208 Sharing
N=195
"Yes" Responses
Condition
True
False
Epstein; Berinsky; Cole; Gully; Pennycook; Rand 5
thereby priming the concept of accuracy when the participants continued on to the sharing
task.
2. In the "Long Evaluation" treatment, participants evaluated the accuracy of eight non-COVID-
related headlines (half true, half false). After each headline, they were (accurately) informed
about whether their answer was correct or incorrect and whether the preceding headline was “a
real news headline” or “a fake news headline. No source was provided for the answers we
provided to them.
3. In the "Importance" treatment, as in Pennycook et al. (2021), participants were asked "How
important is it to you that you share only news articles on social media (such as Facebook and
Twitter) if they are accurate?"
4. In the "Tips" treatment, participants were provided with four simple digital literacy tips, taken
from an intervention developed by Facebook (Guess et al., 2020).
5. In the "Generic Norms" treatment, participants were informed that over 80% of past survey
respondents said it was important to think about accuracy before sharing.
6. In the "Partisan Norms" treatment, participants were informed that 8 out of 10 past survey
respondents said it was ``very important'' or ``extremely important'' to share only accurate news
online, and that this was true of both Democrats and Republicans.
7. In the "Tips+Norms" treatment, participants were shown both the "Partisan Norms" treatment
and the "Tips" treatment, in that order.
8. In the "Importance+Norms" treatment, participants were shown both the "Importance"
treatment and the "Partisan Norms" treatment, in random order.
Figure 2. Sample materials for the evaluation treatment, importance treatment, tips treatment, and partisan norms
treatment. The generic norms treatment had the same image as the partisan norms treatment but used the text “Did
you know, over 80% of past survey respondents say it’s important to think about accuracy before sharing news on
social media?”
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 6
Figure 3. Effect of accuracy prompt interventions. (A) Percentage point change in sharing probability of true and false headlines
caused by each treatment relative to the control. (B) Percent change in discernment (sharing probability of true headlines minus
sharing probability of false headlines) caused by each treatment relative to the control. Sample size for each treatment shown
(control condition included 2,100 participants). Error bars indicate 95% confidence intervals.
For each treatment, the percentage point change relative to the control in sharing of true and false
headlines is shown in Figure 3(A), and the percent change relative to the control in sharing discernment
is shown in Figure 3(B). As can be seen, the Evaluation, Tips, and Importance treatments significantly
increased sharing discernment (p < 0.001 for all except p = 0.010 for Importance), whereas the two norms-
-15
-13
-11
-9
-7
-5
-3
-1
1
3
5
Evaluation
N=935
Long
Evaluation
N=410
Importance
N=1046
Tips
N=906
Generic
Norms
N=510
Partisan
Norms
N=949
Tips
+ Norms
N=934
Importance
+ Norms
N=1072
Percentage Point Change
in Sharing Probability
Treatment
True
False
-50%
-30%
-10%
10%
30%
50%
70%
90%
110%
130%
150%
Evaluation
N=935
Long
Evaluation
N=410
Importance
N=1046
Tips
N=906
Generic
Norms
N=510
Partisan
Norms
N=949
Tips
+ Norms
N=934
Importance
+ Norms
N=1072
Change in Sharing Discernment
Treatment
B
Epstein; Berinsky; Cole; Gully; Pennycook; Rand 7
only treatments did not significantly affect sharing discernment (p = 0.54 for Generic Norms, p = 0.22 for
Partisan Norms). The increase in sharing discernment for the effective treatments was largely driven by a
decrease in the probability of sharing false news (p < 0.05 for all, except for Evaluation at p = 0.075 and
Importance+Norms at p = 0.081; aggregating all treatments, p < 0.001), rather than a change in the
probability of sharing true news (p > 0.22 for all). The effective treatments led to significantly larger
increases in sharing discernment than the (ineffective) norms treatments (p < 0.05 for all comparisons,
except for Importance versus Partisan Norms, p = 0.31). Among the effective treatments, Long Evaluation
was significantly more effective than most of the other treatments (p < 0.05 for all except for Long
Evaluation versus Tips+Norms at p = 0.152), while there were few significant differences between the
other effective treatments (p > 0.05 for all, except for Importance being significantly less effective than
Tips+Norms, p = 0.002 and Importance+Norms, p = 0.046).
The treatments operate by narrowing the gap between accuracy perceptions and sharing intentions
Next, we examined the mechanism underlying the observed treatment effect. To do so, we conducted a
headline-level analysis. For each headline, we calculated its perceived accuracy (breaking down the
accuracy data in Figure 1 by headline) and the overall impact of the effective treatments on sharing
intentions for that headline (average sharing intention among participants in any of the effective
treatments minus average sharing intention among control participants; excludes participants in the two
norms treatments). As shown in Figure 4, we find a strong positive correlation (r(18) = 0.742, p < 0.001),
such that the more inaccurate a headline seemed to participants, the more its sharing was decreased by
the treatments. In other words, the disconnect between perceived accuracy and sharing intentions was
smaller in the treatments than in the control.
Figure 4. Treatments reduce sharing of headlines to the extent that they seem inaccurate. Shown is one point per headline,
with perceived accuracy among participants in the Accuracy-Only condition on the x-axis, and average change in sharing
intentions caused by the treatments on the y-axis.
Does the treatment effect size vary based on user characteristics?
By looking at the three-way interaction between headline veracity, treatment, and various covariates of
interest, we can assess heterogeneous treatment effects based on each covariate. Because higher order
-7
-6
-5
-4
-3
-2
-1
0
1
2
3
0.25 0.35 0.45 0.55 0.65 0.75
Treatment Effect on Sharing
(Percentage points)
Perceived Accuracy
True
False
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 8
interactions require substantially more statistical power (i.e., larger sample size) to detect, we do not
investigate differences between treatments, but instead compare the control to have received any of the
effective treatments (and exclude the two norms treatments) for this analysis. Because these analyses are
exploratory and involve a substantial number of tests, we adjust for the possibility of false positives that
arises from conducting multiple comparisons by applying the HolmBonferroni correction to the reported
p-values (Holm, 1979).
We find significant positive 3-way interactions between veracity, treatment, and the following
variables (such that each variable is associated with a larger treatment effect): the number of attention
checks passed by the participant (pholm < 0.001), self-reported importance placed on sharing only accurate
content (pholm = 0.039), self-reported tendency to seek out news about COVID-19 (pholm = 0.008), the
tendency to stop and think rather than going with one's intuitive first responses (pholm = 0.007) as
measured behaviorally using the Cognitive Reflection Test (a series of questions with intuitively
compelling but wrong answers), education (college degree or higher, pholm = 0.026), and age (pholm = 0.045,
joint significance test of linear and quadratic terms).
Conversely, we find no significant moderation based on participants’ preference for the Republican
versus Democratic party (pholm = 0.746), concern about COVID-19 (pholm= 0.555), race (pholm = 0.179), or
gender (pholm = 0.895).
Methods
From April 27 to May 21, 2020, we conducted five waves of data collection. A total of 11,237 participants
began the study, of which 9,070 participants completed the survey (and thus constitute our final sample);
mage = 44.5; 44.9% male, 54.2% female, 0.9% other; 72.4% White/Caucasian.
Participants were recruited using Lucid, an aggregator of survey platforms that provides samples that
are quota matched to the U.S. on age, gender, race, and geographic region. In all waves, participants were
shown 20 online news cards (e.g., the combination of a headline, an image, and a source) pertaining to
COVID-19. Half were true and half were false. For a list of the headlines used, see Table 2; full experimental
materials are available at https://osf.io/hu4k2/. In all conditions except for the Accuracy-Only condition,
participants were asked “Would you consider sharing this story online (for example, through Facebook or
Twitter)?” (No/Yes binary response); in the Accuracy-Only condition, they were asked “To the best of your
knowledge, is the claim in the above headline accurate?” The sharing intentions input screen for a sample
true and false headline are shown in Figure 4. Although we measure sharing intentions rather than actual
sharing, some evidence in support of the validity of this self-report measure comes from the observation
that news headlines that Mechanical Turk participants report a higher likelihood of sharing indeed
received more shares on Twitter (Mosleh et al., 2020); as well as the observation that accuracy prompts
increase sharing discernment for political headlines using both self-report sharing intentions and actual
sharing in a Twitter field experiment (Pennycook et al., 2021).
As described in detail above, in addition to the Accuracy-Only condition and a control condition, we
ran 8 treatments that consisted of short accuracy prompts administered immediately prior to the sharing
task (see Table 1 for a list of which treatments were included in each wave of data collection).
Prior to the main task, participants indicated their level of concern about COVID-19 and the extent to
which they had been following COVID-19related news. Following the main task, participants completed
the Cognitive Reflection Test (Frederick, 2005), the question about the importance of sharing only
accurate news (except for participants in the Importance treatment, who completed this question at the
study outset), and demographics. We measured ethnicity with the following labels (White/Caucasian,
Asian, Black or African American, American Indian or Alaska Native, Native Hawaiian or Pacific Islander,
other), and gender with the following labels (Male, Female, Transgender Male, Transgender Female,
Epstein; Berinsky; Cole; Gully; Pennycook; Rand 9
Trans/non-binary, not listed, prefer not to answer). Ethnicity and gender are important to consider for
this study to check that the intervention works for a diverse set of participants.
All p-values are calculated using linear regression where the dependent variable is the participant's
choice (share/dont share for all conditions except for the Accuracy-Only condition where the choice was
true/false), with two-way standard errors clustered on participant and headline to account for the non-
independence of repeated choices from the subject, and repeated choices for the same headline. The
independent variables were an indicator variable for headline veracity (0 = false, 1 = true), indicator
variables for each treatment, and interactions between veracity and each treatment indicator (capturing
the effect of each treatment on sharing discernment), as well as indicators for wave (to account for
variation across waves in baseline sharing rates). For the two treatments that were administered in
multiple waves (Evaluation and Tips), we included separate indicator variables for each treatment-wave
combination and used an average of the two coefficients weighted by the treatment sample size in each
wave (using bootstrapping to calculate confidence intervals for plotting). For the moderation analyses, we
used a single “treated” indicator (0 = control, 1 = any of the effective treatments), and interacted the
veracity indicator, the treated indicator, and the veracity X treated interaction with the covariate of
interest.
Table 1. Details of the five waves of data collection. Sharing discernment in the control did not differ
significantly across waves.
Wave
Date
N
Conditions
1
4/27/20
403
Control, Accuracy-Only
2
4/28/20
1192
Control, Evaluation, Long Evaluation
3
5/1/20
2081
Control, Evaluation, Tips, Generic Norms
4
5/55/6
2778
Control, Tips, Partisan Norms, Tips+Norms
5
5/205/21
2616
Control, Importance, Importance+Norms
Table 2. Headlines used in the study.
Veracity
Headline
False
Clinton-Owned Medical Supply Company Quadruples Price for Ventilators and Masks
False
Coconut oil’s history in destroying viruses, including Coronaviruses.
False
COVID-19 is Now Mutating into Something Indescribable That is Now Found in Brazil!
False
Governor Cuomo Signs Law Using Coronavirus as an Excuse to Take ‘Temporary’ Dictatorial
Powers - Blunt Truth
False
Mark of the Beast: Gates Wants Coronavirus Patients Implanted with Quantum-Dot
Tattoos
False
Covid-19’s meant to be the new Black Death, but in Britain no more people are than
NORMAL. What does this say...
False
HUGE! MN Senator and Dr. Reveals HHS Document Coached Him on How to Overcount
COVID-19 Cases -- WITH COPY O….
False
Roberston Blames Coronavirus On Oral Sex, ‘Lady Chemicals’ - The Business Standard News
False
Teen Who Died of Covid-19 Was Denied Treatment Because He Didn’t Have Health
Insurance
False
Vitamin C protects against Coronavirus.
True
Amazon plans to prosecute sellers for price gouging during coronavirus outbreak
True
CDC: Coronavirus spread may last into 2021, but impact can be blunted
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 10
True
Hear what Barack Obama said in 2014 about pandemics - CNN Video
True
Ohio man dies from COVID-19 after criticizing governor’s coronavirus lockdown as
‘madness’
True
Vegas invited homeless to parking lot after coronavirus closed shelter
True
‘There are more important things than living’ says Texas Lieutenant Governor
True
Trump letter attacking Schumer is sent as President says ‘this is not the time for politics’
True
Trump spent the last 2 years slashing the government agencies responsible for handling
the coronavirus outbreak
True
In unprecedented move, Treasury orders Trump’s name printed on stimulus checks
True
U.S. sent millions of fake masks to China early this year, ignoring pandemic warning signs
Figure 5. Sharing intentions input screen for a sample true (left) and false (right) headline.
Bibliography
Fazio, L. (2020). Pausing to consider why a headline is true or false can help reduce the sharing of false
news. Harvard Kennedy School (HKS) Misinformation Review, 1(2). https://doi.org/10.37016/mr-
2020-009
Frederick, S. (2005). Cognitive reflection and decision making. Journal of Economic Perspectives, 19(4),
2542. https://doi.org/10.1257/089533005775196732
Guess, A. M., Lerner, M., Lyons, B., Montgomery, J. M., Nyhan, B., Reifler, J., & Neelanjan, S. (2020). A
digital media literacy intervention increases discernment between mainstream and False News
in the United States and India. Proceedings of the National Academy of Sciences, 117(27),
1553615545. https://doi.org/10.1073/pnas.1920498117
Holm, S. (1979). A simple sequentially rejective multiple test procedure. Scandinavian Journal of
Statistics, 6(2), 6570. http://www.jstor.org/stable/4615733?origin=JSTOR-pdf
Epstein; Berinsky; Cole; Gully; Pennycook; Rand 11
Jahanbakhsh, F., Zhang, A. X., Berinsky, A. J., Pennycook, G., Rand, D. G., & Karger, D. R. (2021). Exploring
lightweight interventions at posting time to reduce the sharing of misinformation on social
media. Proceedings of the ACM on Human-Computer Interaction, 5 (CSCW1).
https://doi.org/10.1145/3449092
Lazer, D. M. J., Baum, M., Benkler, Y., Berinsky, A. J., Greenhill, K. M., Menczer, F. Metzger, M., Nyhan,
B., Pennycook, G., Rothschild, D., Schudson, M., Sloman, S. A., Sunstein, C., Thorson, E. A.,
Watts, D. J., & Zittrain, J. L. (2018). The science of fake news. Science, 359(6380), 10941096.
https://doi.org/10.1126/science.aao2998
Mosleh, M., Pennycook, G., & Rand, D. G. (2020). Self-reported willingness to share political news
articles in online surveys correlates with actual sharing on Twitter. PLOS ONE, 15(2), e0228882.
https://doi.org/10.1371/journal.pone.0228882
Pennycook, G., & Rand, D. G. (2021). The psychology of fake news. Trends in Cognitive Sciences, 25(5),
388402. https://doi.org/10.1016/j.tics.2021.02.007
Pennycook, G., Bear, A., Collins, E. T., & Rand, D. G. (2020). The implied truth effect: Attaching warnings
to a subset of fake news headlines increases perceived accuracy of headlines without warnings.
Management Science, 66(11), 49444957. https://doi.org/10.1287/mnsc.2019.3478
Pennycook, G., Cannon, T. D., & Rand, D. G. (2018). Prior exposure increases perceived accuracy of fake
news. Journal of Experimental Psychology: General, 147(12), 18651880.
https://doi.org/10.1037/xge0000465
Pennycook, G., Epstein, Z., Mosleh, M., Arechar, A. A., Eckles, D., & Rand, D. G. (2021). Shifting attention
to accuracy can reduce misinformation online. Nature, 592, 590595.
https://doi.org/10.1038/s41586-021-03344-2
Pennycook, G., McPhetres, J., & Zhang, Y. (2020). Fighting COVID-19 misinformation on social media:
Experimental evidence for a scalable accuracy-nudge intervention. Psychological Science, 31(7),
770780. https://doi.org/10.1177%2F0956797620939054
Developing an accuracy-prompt toolkit to reduce COVID-19 misinformation online 12
Funding
The authors gratefully acknowledge funding from the Ethics and Governance of Artificial Intelligence
Initiative of the Miami Foundation, the William and Flora Hewlett Foundation, the Reset Initiative of
Luminate (part of the Omidyar Network), the John Templeton Foundation, the TDF Foundation, the
Canadian Institutes of Health Research, the Social Sciences and Humanities Research Council of Canada,
and Google.
Competing interests
AB and DR received research support through gifts from Google; RC and AG work for Google.
Ethics
This research was deemed exempt by the MIT Committee on the Use of Humans as Experimental Subjects,
#1806400195.
Copyright
This is an open access article distributed under the terms of the Creative Commons Attribution License,
which permits unrestricted use, distribution, and reproduction in any medium, provided that the original
author and source are properly credited.
Data availability
All materials needed to replicate this study are available via the Harvard Dataverse:
https://doi.org/10.7910/DVN/18SHLJ
... One possible explanation is that the educational intervention was too brief and simple, which did not lead to any additional effect. Feedback, which is itself a type of educational intervention (Epstein et al., 2021), appeared to be sufficient to improve accuracy. These findings underscore the value of feedback and suggest that more robust or long-term educational interventions may be needed to enhance participants' detection abilities. ...
... It focused on images from a specific online community (/r/photoshopbattles), and the intervention's effectiveness for politically motivated or other specialized manipulations remains uncertain. Notably, there was no significant difference in accuracy between participants receiving feedback alone and those receiving feedback plus an educational intervention, suggesting that reflecting on the accuracy of their classifications during feedback may enhance critical thinking (Epstein et al., 2021). Additionally, as prior research has shown, media literacy interventions often have short-lived effects (Maertens et al., 2021). ...
Article
Full-text available
This study investigates the impact of educational interventions and feedback on users' ability to detect manipulated images on social media, addressing a gap in research that has primarily focused on algorithmic approaches. Through a pre-registered randomized and controlled experiment, we found that feedback and educational content significantly improved participants' ability to detect manipulated images on social media. However, the educational content did not result in a significantly greater improvement compared to feedback alone. These findings underscore feedback as a powerful tool for enhancing digital literacy, with practical implications for combating misinformation.
... It is therefore not surprising that the inattention-based account has proven to be extremely influential in guiding research and discussion on how to combat inaccurate information on social media [19][20][21][22] . The account is also supported by findings showing that the discrimination between accurate and inaccurate statements in sharing decisions is affected by situational factors such as accuracy prompts 17,18,23 and individual factors such as the tendency for cognitive reflection 17,24 . ...
... Given the extensive evidence for the influence of verbal framing on behavior [32][33][34][35][36][37] , it is important to examine whether the framing of the questions is the cause of the dissociation between the responses to the accuracy questions and the responses to the sharing questions. Testing how question framing affects the responses to accuracy and sharing questions is important, not the least because questions with the framing scrutinized here have been used by the original authors in several other studies 23,24,[38][39][40][41][42] as well as by other research groups 22,[43][44][45][46][47] . According to the inattention-based account 17,18 , people should be well able to discriminate between accurate and inaccurate statements in response to accuracy questions but should discriminate less between accurate and inaccurate statements in response to sharing questions. ...
Article
Full-text available
Previous research suggests that even when people are capable of judging to the best of their knowledge whether claims are accurate or inaccurate, they do not sufficiently discriminate between accurate and inaccurate information when asked to consider whether they would share stories on social media. However, question framing (“To the best of your knowledge…”, “Would you consider…?”) differed between the questions about accuracy and the questions about sharing. Here we examine the effects of question framing on responses to accuracy questions and responses to sharing questions. The framing of accuracy questions had no effect on accurate-inaccurate discrimination. In contrast, accurate-inaccurate discrimination in response to sharing questions increased when participants were asked to respond, to the best of their knowledge, whether they would share claims compared to when they were asked whether they would consider sharing stories. At a theoretical level, the findings support the inattention-based account, according to which contextual cues shifting the focus toward accuracy can enhance accurate-inaccurate discrimination in sharing responses. At a methodological level, these findings suggest that researchers should carefully attend to the verbal framing of questions about sharing information on social media, as the framing may significantly influence participants’ focus on accuracy.
... Willingness to Liking or Sharing headline [25] This metric measures the likelihood that participants will engage with the content by liking or sharing it. Sharing intention/Likelihood to Share [37] This measures the likelihood that a participant would share the information on social media. ...
Preprint
Full-text available
Misinformation has become a widespread issue in the 21st century, impacting numerous areas of society and underscoring the need for effective intervention strategies. Among these strategies, user-centered interventions, such as warning systems, have shown promise in reducing the spread of misinformation. Many studies have used various metrics to evaluate the effectiveness of these warning interventions. However, no systematic review has thoroughly examined these metrics in all studies. This paper provides a comprehensive review of existing metrics for assessing the effectiveness of misinformation warnings, categorizing them into four main groups: behavioral impact, trust and credulity, usability, and cognitive and psychological effects. Through this review, we identify critical challenges in measuring the effectiveness of misinformation warnings, including inconsistent use of cognitive and attitudinal metrics, the lack of standardized metrics for affective and emotional impact, variations in user trust, and the need for more inclusive warning designs. We present an overview of these metrics and propose areas for future research.
... These teachers' training in the sciences equips them with scientific reasoning skills and an appreciation for evidence-based information [2,41]. This can enhance their ability to discern accurate from inaccurate content, consistent with previous research [40,42]. This suggests that the scientific literacy and cognitive reflection innate to their profession could serve as a buffer against misinformation. ...
Article
Full-text available
Introduction: The COVID-19 pandemic underscored the importance of health literacy, where pre-service Life Sciences teachers can promote safe behavioral practices such as vaccination. However, the health information sources influencing their behavioral intentions toward vaccination are poorly understood. This study examines how different sources impact vaccination attitudes and behaviors among pre-service Life Sciences teachers. Methods: A cross-sectional study was conducted involving pre-service Life Sciences teachers (n = 87) from a South African university. Data were collected using a closed-ended questionnaire to explore the types of information sources they use and how these sources influence their attitudes, subjective norms, and perceived behavioral control related to COVID-19 vaccination. Results: Participants utilized a diverse array of information sources, including government communiques, traditional media, social media, research publications, and personal interactions. Although the susceptibility to misinformation from unreliable sources such as social media was identified, participants' scientific training served as a buffer against such misinformation. Notably, the type of information source did not significantly predict subjective norms and perceived behavioral control. However, information sources significantly influenced attitudes and normative beliefs toward vaccination. Discussion: The study underscores the complex relationship between information sources, cognitive processes, and behavioral intentions in the context of health literacy among pre-service Life Sciences teachers. The findings suggest that while their scientific background provides a defense against misinformation, targeted strategies are essential to ensure the reliability and validity of information. Further research is warranted to understand additional factors influencing this demographic's vaccination intentions and to inform more effective health literacy initiatives. This research bears significant implications for public health communication strategies, particularly in a pandemic.
Technical Report
Full-text available
There is widespread concern that misinformation poses dangerous risks to health, well-being, and civic life. Despite a growing body of research on the topic, significant questions remain about (a) psychological factors that render people susceptible to misinformation, (b) the extent to which it affects real-world behavior, (c) how it spreads online and offline, and (d) intervention strategies that counter and correct it effectively. This report reviews the best available psychological science research to reach consensus on each of these crucial questions, particularly as they pertain to health-related misinformation. In addition, the report offers eight specific recommendations for scientists, policymakers, and health professionals who seek to recognize and respond to misinformation in health care and beyond.
Article
Full-text available
The global reach of misinformation has exacerbated harms in low- and middle-income countries faced with deficiencies in funding, platform engagement, and media literacy. These challenges have reiterated the need for the development of strategies capable of addressing misinformation that cannot be countered using popular fact-checking methods. Focusing on Kenya’s contentious 2022 election, we evaluate a novel method for democratizing debunking efforts termed “social truth queries” (STQs), which use questions posed by everyday users to draw reader attention to the veracity of the targeted misinformation in the aim of minimizing its impact. In an online survey of Kenyan participants ( N ~ 4,000), we test the efficacy of STQs in reducing the influence of electoral misinformation which could not have been plausibly fact-checked using existing methods. We find that STQs reduce the perceived accuracy of misinformation while also reducing trust in prominent disseminators of misinformation, with null results for sharing propensities. While effect sizes are small across conditions, assessments of the respondents most susceptible to misinformation reveal larger potential effects if targeted at vulnerable users. These findings collectively illustrate the potential of STQs to expand the reach of debunking efforts to a wider array of actors and misinformation clusters.
Article
Full-text available
We synthesize a burgeoning literature investigating why people believe and share false or highly misleading news online. Contrary to a common narrative whereby politics drives susceptibility to fake news, people are ‘better’ at discerning truth from falsehood (despite greater overall belief) when evaluating politically concordant news. Instead, poor truth discernment is associated with lack of careful reasoning and relevant knowledge, and the use of heuristics such as familiarity. Furthermore, there is a substantial disconnect between what people believe and what they share on social media. This dissociation is largely driven by inattention, more so than by purposeful sharing of misinformation. Thus, interventions can successfully nudge social media users to focus more on accuracy. Crowdsourced veracity ratings can also be leveraged to improve social media ranking algorithms.
Article
Full-text available
In recent years, there has been a great deal of concern about the proliferation of false and misleading news on social media1, 2, 3–4. Academics and practitioners alike have asked why people share such misinformation, and sought solutions to reduce the sharing of misinformation5, 6–7. Here, we attempt to address both of these questions. First, we find that the veracity of headlines has little effect on sharing intentions, despite having a large effect on judgments of accuracy. This dissociation suggests that sharing does not necessarily indicate belief. Nonetheless, most participants say it is important to share only accurate news. To shed light on this apparent contradiction, we carried out four survey experiments and a field experiment on Twitter; the results show that subtly shifting attention to accuracy increases the quality of news that people subsequently share. Together with additional computational analyses, these findings indicate that people often share misinformation because their attention is focused on factors other than accuracy—and therefore they fail to implement a strongly held preference for accurate sharing. Our results challenge the popular claim that people value partisanship over accuracy8,9, and provide evidence for scalable attention-based interventions that social media platforms could easily implement to counter misinformation online.
Article
Full-text available
Significance Few people are prepared to effectively navigate the online information environment. This global deficit in digital media literacy has been identified as a critical factor explaining widespread belief in online misinformation, leading to changes in education policy and the design of technology platforms. However, little rigorous evidence exists documenting the relationship between digital media literacy and people’s ability to distinguish between low- and high-quality news online. This large-scale study evaluates the effectiveness of a real-world digital media literacy intervention in both the United States and India. Our largely encouraging results indicate that relatively short, scalable interventions could be effective in fighting misinformation around the world.
Article
Full-text available
There is an increasing imperative for psychologists and other behavioral scientists to understand how people behave on social media. However, it is often very difficult to execute experimental research on actual social media platforms, or to link survey responses to online behavior in order to perform correlational analyses. Thus, there is a natural desire to use self-reported behavioral intentions in standard survey studies to gain insight into online behavior. But are such hypothetical responses hopelessly disconnected from actual sharing decisions? Or are online survey samples via sources such as Amazon Mechanical Turk (MTurk) so different from the average social media user that the survey responses of one group give little insight into the on-platform behavior of the other? Here we investigate these issues by examining 67 pieces of political news content. We evaluate whether there is a meaningful relationship between (i) the level of sharing (tweets and retweets) of a given piece of content on Twitter, and (ii) the extent to which individuals (total N = 993) in online surveys on MTurk reported being willing to share that same piece of content. We found that the same news headlines that were more likely to be hypothetically shared on MTurk were also shared more frequently by Twitter users, r = .44. For example, across the observed range of MTurk sharing fractions, a 20 percentage point increase in the fraction of MTurk participants who reported being willing to share a news headline on social media was associated with 10x as many actual shares on Twitter. We also found that the correlation between sharing and various features of the headline was similar using both MTurk and Twitter data. These findings suggest that self-reported sharing intentions collected in online surveys are likely to provide some meaningful insight into what content would actually be shared on social media.
Article
Full-text available
In an online experiment, participants who paused to explain why a headline was true or false indicated that they were less likely to share false information compared to control participants. Their intention to share accurate news stories was unchanged. These results indicate that adding “friction” (i.e., pausing to think) before sharing can improve the quality of information shared on social media.
Preprint
Full-text available
There is an increasing imperative for psychologists and other behavioral scientists to understand how people behave on social media. However, it is often very difficult to execute experimental research on actual social media platforms, or to link survey responses to online behavior in order to perform correlational analyses. Thus, there is a natural desire to use self-reported behavioral intentions in standard survey studies to gain insight into online behavior. But are such hypothetical responses hopelessly disconnected from actual sharing decisions? Or are online survey samples via sources such as Amazon Mechanical Turk (MTurk) so different from the average social media user that the survey responses of one group give little insight into the on-platform behavior of the other? Here we investigate these issues by examining 67 pieces of political news content. We evaluate whether there is a meaningful relationship between (i) the level of sharing (tweets and retweets) of a given piece of content on Twitter, and (ii) the extent to which individuals (total N = 993) in online surveys on MTurk reported being willing to share that same piece of content. We found that the same news headlines that were more likely to be hypothetically shared on MTurk were actually shared more frequently by Twitter users, r = .44. For example, across the observed range of MTurk sharing fractions, a 20 percentage point increase in the fraction of MTurk participants who reported being willing to share a news headline on social media was associated with 10x as many actual shares on Twitter. This finding suggests that self-reported sharing intentions collected in online surveys are likely to provide some meaningful insight into what participants would actually share on social media.
Article
Full-text available
The 2016 U.S. presidential election brought considerable attention to the phenomenon of “fake news”: entirely fabricated and often partisan content that is presented as factual. Here we demonstrate one mechanism that contributes to the believability of fake news: fluency via prior exposure. Using actual fake-news headlines presented as they were seen on Facebook, we show that even a single exposure increases subsequent perceptions of accuracy, both within the same session and after a week. Moreover, this “illusory truth effect” for fake-news headlines occurs despite a low level of overall believability and even when the stories are labeled as contested by fact checkers or are inconsistent with the reader’s political ideology. These results suggest that social media platforms help to incubate belief in blatantly false news stories and that tagging such stories as disputed is not an effective solution to this problem. It is interesting, however, that we also found that prior exposure does not impact entirely implausible statements (e.g., “The earth is a perfect square”). These observations indicate that although extreme implausibility is a boundary condition of the illusory truth effect, only a small degree of potential plausibility is sufficient for repetition to increase perceived accuracy. As a consequence, the scope and impact of repetition on beliefs is greater than has been previously assumed.
Article
Full-text available
Addressing fake news requires a multidisciplinary effort
Article
When users on social media share content without considering its veracity, they may unwittingly be spreading misinformation. In this work, we investigate the design of lightweight interventions that nudge users to assess the accuracy of information as they share it. Such assessment may deter users from posting misinformation in the first place, and their assessments may also provide useful guidance to friends aiming to assess those posts themselves. In support of lightweight assessment, we first develop a taxonomy of the reasons why people believe a news claim is or is not true; this taxonomy yields a checklist that can be used at posting time. We conduct evaluations to demonstrate that the checklist is an accurate and comprehensive encapsulation of people's free-response rationales. In a second experiment, we study the effects of three behavioral nudges---1) checkboxes indicating whether headings are accurate, 2) tagging reasons (from our taxonomy) that a post is accurate via a checklist and 3) providing free-text rationales for why a headline is or is not accurate---on people's intention of sharing the headline on social media. From an experiment with 1668 participants, we find that both providing accuracy assessment and rationale reduce the sharing of false content. They also reduce the sharing of true content, but to a lesser degree that yields an overall decrease in the fraction of shared content that is false. Our findings have implications for designing social media and news sharing platforms that draw from richer signals of content credibility contributed by users. In addition, our validated taxonomy can be used by platforms and researchers as a way to gather rationales in an easier fashion than free-response.
Preprint
Across two studies with over 1,600 U.S. adults recruited online, we present evidence that people share false claims about COVID-19 partly because they simply fail to think sufficiently about whether or not content is accurate when deciding what to share. In Study 1, participants were far worse at discerning between true and false content when deciding what they would share on social media relative to when they are asked directly about accuracy. Furthermore, cognitive reflection and science knowledge were associated with stronger discernment. In Study 2, we found that a simple accuracy reminder at the beginning of the study – i.e., judging the accuracy of a non-COVID-19-related headline – more than doubled the level of truth discernment in participants’ sharing intentions. Our results, which mirror those found previously for political fake news, suggest that nudging people to think about accuracy is a simple way to improve choices about what to share on social media.