ArticlePDF Available

A Response to a Comment on Hall et al. (2024)

Authors:

Abstract

In their commentary on Hall et al., Buckwalter and Friedman (2024) claimed that the replication should have been interpreted as successful, argued that the researchers’ conclusions were incorrect, and implied that the replication effort was misguided. As a subset of contributors to the Hall et al. replication, we appreciate the opportunity to respond to their comment. Although we recognize the potential for disagreement in the interpretation of research results, Buckwalter and Friedman’s critique ignored several key features of the research, and many of their arguments and proposed interpretations were already addressed by Hall et al. (2024). In response to their comment, we (a) explain why Hall et al. did not replicate all of the original findings, (b) emphasize how Hall et al. were accurate and nuanced in the description and interpretation of their results, and (c) caution against focusing on the mechanisms underlying a psychological phenomenon before it is clearly established
https://doi.org/10.1177/25152459241268249
Advances in Methods and
Practices in Psychological Science
October-December 2024, Vol. 7, No. 4,
pp. 1 –4
© The Author(s) 2024
Article reuse guidelines:
sagepub.com/journals-permissions
DOI: 10.1177/25152459241268249
www.psychologicalscience.org/AMPPS
ASSOCIATION FOR
PSYCHOLOGICAL SCIENCE
Creative Commons CC BY: This article is distributed under the terms of the Creative Commons Attribution 4.0 License
(https://creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission
provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).
Commentary
The conditions under which laypeople attribute knowl-
edge to protagonists have long been debated by experi-
mental philosophers (Colaço etal., 2014; Nagel etal.,
2013; Weinberg etal., 2001). Consider the case of “Dar-
rel,” who accurately recognizes the species of an animal
in the woods even though it was the only one of its kind
among many animals of an identical-looking species.
Responses to this so-called “Gettier-type” case have been
studied to examine whether laypeople consider luckily
true beliefs as constituting actual knowledge. In their
Experiment 1, Turri et al. (2015) compared knowledge
attributions in this version of the “Darrel” case to a ver-
sion in which his belief is a clear case of knowledge;
they found no difference between these conditions and
concluded that “a salient but failed threat to the truth of
a judgment does not significantly affect whether it is
viewed as knowledge” (p. 381). Hall etal. (2024) repli-
cated and extended Experiment 1 of Turri etal., testing
condition differences by using the Darrel case alongside
two other counterfeit-object Gettier-type cases in a large
multinational study. Hall etal. found that participants
were less likely to attribute knowledge to the protago-
nists when beliefs were only luckily true (i.e., the Gettier
conditions) than when the truth of the beliefs was not
under threat (i.e., the knowledge conditions). This sig-
nificant condition difference reported by Hall etal. stands
in contrast to the null result reported by Turri etal.
In their commentary on Hall etal., Buckwalter and
Friedman (2024) claimed that the replication should have
been interpreted as successful, argued that the research-
ers’ conclusions were incorrect, and implied that the
replication effort was misguided. As a subset of contribu-
tors to the Hall et al. replication, we appreciate the
opportunity to respond to their comment. Although we
recognize the potential for disagreement in the interpre-
tation of research results, Buckwalter and Friedman’s
critique ignored several key features of the research, and
many of their arguments and proposed interpretations
were already addressed by Hall etal. (2024). In response
to their comment, we (a) explain why Hall etal. did not
replicate all of the original findings, (b) emphasize how
1268249AMPXXX10.1177/25152459241268249Schmidt et al.Advances in Methods and Practices in Psychological Science
research-article2024
Corresponding Author:
Jordan Wagge, School of Psychology, Avila University, Kansas City,
Missouri
Email: jordan.wagge@avila.edu
A Response to a Comment
on Hall etal. (2024)
Kathleen Schmidt1, Gerald J. Haeffel2, Neil Levy3,4 ,
David Moreau5,6 , Sean T. H. Lee7, Erin M. Buchanan8,
Anthony J. Krafnick9, Martin Voracek10 , Gerit Pfuhl11,
Krystian Barzykowski12, Marta Kowal13 , and Jordan Wagge14
1Department of Psychology, Ashland University, Ashland, Ohio; 2Department of Psychology, University
of Notre Dame, Notre Dame, Indiana; 3Department of Philosophy, Macquarie University, Sydney,
Australia; 4Uehiro Centre for Practical Ethics, University of Oxford, Oxford, England; 5School of
Psychology, University of Auckland, Auckland, New Zealand; 6Centre for Brain Research, University
of Auckland, Auckland, New Zealand; 7School of Social and Health Sciences, James Cook University,
Singapore; 8Analytics, Harrisburg University of Science and Technology, Harrisburg, Pennsylvania;
9Department of Psychology, Dominican University, River Forest, Illinois; 10Department of Cognition,
Emotion, and Methods in Psychology, Faculty of Psychology, University of Vienna, Vienna, Austria;
11Department of Psychology, Norwegian University of Science and Technology, Trondheim, Norway;
12Institute of Psychology, Faculty of Philosophy, Jagiellonian University, Krakow, Poland; 13IDN
Being Human Lab - Institute of Psychology, University of Wrocław, Wrocław, Poland; and 14School of
Psychology, Avila University, Kansas City, Missouri
Keywords
epistemic intuitions, Gettier cases, knowledge, replication
Received 5/31/24; Revision accepted 7/1/24
2 Schmidt et al.
Hall etal. were accurate and nuanced in the description
and interpretation of their results, and (c) caution against
focusing on the mechanisms underlying a psychological
phenomenon before it is clearly established.
Different Statistical and Interpretive
Conclusions
Although Hall etal. (2024) found evidence for several
of the original claims of Turri etal. (2015), they did not
replicate one notable null result: Unlike Turri etal., they
observed a difference in knowledge attribution between
the Gettier condition and the knowledge control condi-
tion. Although small in magnitude, this difference was
identified robustly across three tested vignettes, different
measures of knowledge attribution, and multiple analy-
ses. Hall etal. further examined possible moderators,
none of which eliminated the condition difference that
Turri etal. failed to find. Given two studies that differ
on the statistical significance of their results, we contend
that the study with the larger, more diverse sample and
higher analytic power to detect differences is more likely
to yield valid, replicable, and generalizable results.
Because the difference between the aforementioned
conditions is absolutely central to the philosophical
debate over Gettier cases, and the experimental philo-
sophical literature has accordingly focused on it, we are
confident that Hall etal.’s failure to replicate the original
null result is an important contribution.
Nonetheless, Buckwalter and Friedman (2024) con-
cluded that the two discrepant findings were “basically
the same.” This assertion is based on their interpretation
of descriptive differences in the frequencies of knowl-
edge attributions between conditions and enables them
to generate a narrative (or, reframing) that would align
with that of Turri etal. (2015). However, Hall etal.’s
(2024) preregistered analysis plan did not include this
method of assessing the results. The purpose of prereg-
istration is to reduce researcher degrees of freedom and
to discourage exploratory analyses disguised as confir-
matory, especially those that support preferred narratives
(e.g., Nosek etal., 2018; Wicherts et al., 2016). In this
case, Hall etal. anticipated a null result aligned with the
original finding, conducted a prespecified analysis with
any data-driven changes transparently noted, detected a
nominally significant result, and reported those findings
accordingly. Based on these prespecified conditions and
consistent with the framework of null hypothesis signifi-
cance testing (NHST), we must state that Hall etal.’s
findings were different from those of the original study.
Although one can, and Hall etal. (2024) did, calculate
an effect size from the null result reported in Experiment
1 of Turri etal. (2015; i.e., odds ratio [OR] = 2.00, 95%
confidence interval [CI] = [0.77, 5.21]), interpreting the
results of the statistical test beyond a failure to reject the
null hypothesis would violate the logic of NHST (e.g.,
Frick, 1996). The null result merely indicates that Turri
etal. had insufficient evidence to support a claim of a
difference between conditions. Nonetheless, as Hall
etal. acknowledged, the effect they observed (i.e., OR =
1.86, 95% CI = [1.78, 1.94]) was similar in size to the
original. Accordingly, as both Hall etal. and Buckwalter
and Friedman (2024) pointed out, the large sample in
the replication study may be one explanation for the
difference in statistical conclusions. Lacking this insight,
Turri etal. concluded that knowledge attributions were
insensitive to the differences in Gettier and knowledge
conditions in their Experiment 1—a claim that was
not supported by the results of Hall etal. Nonetheless,
Buckwalter and Friedman suggested that the percentage
differences in the original, likely underpowered, Turri
etal. study should have been the sole criteria for replica-
tion rather than the statistical conclusions or interpreta-
tions therefrom. However, the CI of the original effect
is quite wide and includes effect sizes that would imply
a greater likelihood of knowledge attribution in the
knowledge condition than the Gettier condition (OR > 1),
a null result (OR = 1), and a greater likelihood of knowl-
edge attribution in the Gettier condition than the knowl-
edge condition (OR < 1); therefore, their results were
inconclusive. In contrast, the results from Hall etal. rep-
resent a clear, albeit small, Gettier-intuition effect. The
results of the original study simply cannot be classified
as “basically the same” as those presented by Hall etal.
Accurate and Nuanced
Contrary to Buckwalter and Friedman’s (2024) reading,
Hall etal. (2024) did not “claim to . . . show that people
deny knowledge to lucky agents” or “infer from this
finding that there is a common psychological tendency
to deny that true-by-luck beliefs are knowledge.” Rather,
Hall etal. accurately reported a nominally significant
condition difference, which they then interpreted accord-
ingly. For example, they stated the following:
This result did not correspond to that found by
Turri etal., who failed to detect a significant dif-
ference in knowledge attribution between these
two conditions . . . we did find effect sizes in the
same range as Turri etal. when directly comparing
like conditions; however, the null result did not
replicate.
Hall etal. (2024) further noted their sample’s baseline
skepticism (i.e., that a “notable number of participants
[43.41%; see Table 7] denied knowledge to protagonists
even in clear cases of justified true belief”) and for this
Advances in Methods and Practices in Psychological Science 7(4) 3
reason, were careful in their interpretation of the condi-
tion difference. Specifically, they acknowledged that
“Gettier intuitions were by no means common,” “the
small size of the effect suggests that Gettier intuitions
were not prevalent in our research,” and that “given the
small size of the observed effect, the theoretical signifi-
cance of this result is debatable.” They also provided a
nuanced discussion of how differences in methods,
design, analytic approach, and geographical origin and
composition of the samples could have affected results.
Constructs and Mechanisms
We agree with Buckwalter and Friedman (2024) that
researchers who claim to answer vague questions about
poorly defined constructs may very well lead readers to
inappropriate conclusions. However, Hall etal.’s (2024)
rhetorical use of “Gettier-type case” and “Gettier intu-
ition” did not preclude the specificity of their claims or
the recognition of their constraints on generality. Their
limitations were readily acknowledged. Contrary to what
Buckwalter and Friedman implied, Hall etal. did not
make strong statements about the theoretical implica-
tions of their results.
Buckwalter and Friedman (2024) argued for the devel-
opment of theories and identification of mechanisms in
research on luckily true beliefs. We agree that the study
of mechanisms is important; in fact, Hall etal. (2024)
explored the moderating role of luck attributions directly
in their analyses.1 However, the examination of mecha-
nisms should, in our view, follow the confirmation of an
effect’s replicability. The approach of Hall etal. was to
first establish the presence of the phenomenon before
(and alongside) identifying the factors that influence it.
Turri etal. (2015) employed a particular strategy and
methodology for studying knowledge attributions in their
Experiment 1 and made conclusions based on the results.
The purpose of Hall etal.’s replication was to use similar
strategies to determine whether Turri etal.’s findings
were robust. The replication results demonstrate that
prior to studying mechanisms, establishing agreement on
what researchers are trying to understand is necessary.
Examining “specific forms of luck” and theorizing about
their importance, as Buckwalter and Friedman recom-
mended, would have been premature. When results do
not conform to hypotheses, researchers tend to avoid
accepting the conclusion that their theory may not be
correct (i.e., refuted; Popper, 1959). Instead, not infre-
quently, they embark on subtle searches for subgroups
or moderators for which their theory may hold. Hall etal.
found differences and highlighted methodological issues,
which, had they not been identified, could have hindered
progress on theoretical development.
Balancing Replication Goals
Buckwalter and Friedman (2024) suggested that Hall
etal. (2024) did not appreciate the tension between two
replication goals: that of assessing reliability and that
of understanding phenomena. Hall etal. set out to do
a simple, direct replication study with both pedagogical
and scientific goals. Because of the partnership between
the Psychological Science Accelerator (Moshontz etal.,
2018) and the Collaborative Replications and Education
Project (Grahe et al., 2024; Wagge etal., 2019), the
research project increased in complexity and scope over
the course of the Stage 1 Registered Report review
process. As a consequence, additional piloted vignettes
were incorporated, measures were added and changed,
alternative response options were included, extension
hypotheses were proposed by coauthors, and plans to
assess moderating variables were formulated. The con-
ceptual replication Hall etal. conducted emerged out
of precisely those tensions noted by Buckwalter and
Friedman and the additional consideration of pedagogi-
cal value. We hope that the findings of Hall etal. will
contribute to the incremental understanding of epis-
temic intuitions and, in turn, inspire future inquiry along
these lines, that is, examining subtypes, mechanisms,
and moderators of so-called Gettier intuitions.
Transparency
Action Editor: David A. Sbarra
Editor: David A. Sbarra
Author Contributions
Kathleen Schmidt: Conceptualization; Project adminis-
tration; Writing – original draft; Writing – review &
editing.
Gerald J. Haeffel: Conceptualization; Writing – original
draft; Writing – review & editing.
Neil Levy: Conceptualization; Writing – review & editing.
David Moreau: Conceptualization; Writing – review &
editing.
Sean T. H. Lee: Conceptualization; Writing – review &
editing.
Erin M. Buchanan: Conceptualization; Writing – review
& editing.
Anthony J. Krafnick: Conceptualization; Writing – review
& editing.
Martin Voracek: Conceptualization; Writing – review &
editing.
Gerit Pfuhl: Conceptualization; Writing – review &
editing.
Krystian Barzykowski: Conceptualization; Writing –
review & editing.
Marta Kowal: Conceptualization; Writing – review &
editing.
Jordan Wagge: Conceptualization; Project administration;
Writing – original draft; Writing – review & editing.
4 Schmidt et al.
Declaration of Conflicting Interests
The author(s) declared that there were no conflicts of inter-
est with respect to the authorship or the publication of this
article.
Funding
K. Schmidt was supported by a grant from the John Temple-
ton Foundation (61825). N. Levy was supported by
a grant from the John Templeton Foundation (62631).
K. Barzykowski was supported by a grant from the National
Science Centre, Poland (2019/35/B/HS6/00528). J. Wagge
was supported by a grant from the National Science Foun-
dation (2141930).
ORCID iDs
Kathleen Schmidt https://orcid.org/0000-0002-9946-5953
Gerald J. Haeffel https://orcid.org/0000-0002-4029-1493
Neil Levy https://orcid.org/0000-0002-5679-1986
David Moreau https://orcid.org/0000-0002-1957-1941
Sean T. H. Lee https://orcid.org/0000-0002-4059-9738
Erin M. Buchanan https://orcid.org/0000-0002-9689-4189
Anthony J. Krafnick https://orcid.org/0000-0002-1692-
0413
Martin Voracek https://orcid.org/0000-0001-6109-6155
Marta Kowal https://orcid.org/0000-0001-9050-1471
Jordan Wagge https://orcid.org/0000-0002-5105-2084
Note
1. In this analysis, luck attributions did not moderate the Gettier-
intuition effect.
References
Buckwalter, W., & Friedman, O. (2024). Robust evidence for
knowledge attribution and luck: A comment on Hall etal.
(2024). Advances in Methods and Practices in Psychological
Science, 7(4). https://doi.org/10.1177/25152459241268220
Colaço, D., Buckwalter, W., Stich, S., & Machery, E. (2014).
Epistemic intuitions in fake-barn thought experiments.
Episteme, 11(2), 199–212. https://doi.org/10.1017/epi.2014.7
Frick, R. W. (1996). The appropriate use of null hypothesis
testing. Psychological Methods, 1(4), 379–390. https://doi
.org/10.1037/1082-989X.1.4.379
Grahe, J. E., Brandt, M. J., Wagge, J. R., Legate, N., Wiggins, B. J.,
Christopherson, C. D., Weisberg, Y., Corker, K. S., Chartier,
C. R., Fallon, M., Hildebrandt, L., Hurst, M. A., Lazarevic, L.,
Levitan, C., McFall, J., McLaughlin, H., Pazda, A.,
IJzerman, H., Nosek, B. A., . . . France, H. (2024).
Collaborative Replications and Education Project (CREP).
OSF. https://doi.org/10.17605/OSF.IO/WFC6U
Hall, B. F., Schmidt, K., Wagge, J. R., Lewis, S. C., Weissgerber,
S. C., Kiunke, F., Pfuhl, G., Stieger, S., Tran, U. S.,
Barzykowski, K., Bogatyreva, N., Kowal, M., Massar, K.,
Pernerstorfer, F., Sorokowski, P., Voracek, M., Chartier, C. R.,
Brandt, M. J., Grahe, J. E., . . . Buchanan, E. M. (2024).
Registered Replication Report: A large multilab cross-
cultural conceptual replication of Turri et al. (2015).
Advances in Methods and Practices in Psychological
Science, 7(4). https://doi.org/10.1177/25152459241267902
Moshontz, H., Campbell, L., Ebersole, C. R., IJzerman, H., Urry,
H. L., Forscher, P. S., Grahe, J. E., McCarthy, R. J., Musser,
E. D., Antfolk, J., Castille, C. M., Evans, T. R., Fiedler, S.,
Flake, J. K., Forero, D. A., Janssen, S. M. J., Keene, J. R.,
Protzko, J., Aczel, B., . . . Chartier, C. R. (2018). The
Psychological Science Accelerator: Advancing psychology
through a distributed collaborative network. Advances
in Methods and Practices in Psychological Science, 1(4),
501–515. https://doi.org/10.1177/2515245918797607
Nagel, J., San Juan, V., & Mar, R. A. (2013). Lay denial of
knowledge for justified true beliefs. Cognition, 12(3),
652–661. https://doi.org/10.1016/j.cognition.2013.02.008
Nosek, B. A., Ebersole, C. R., DeHaven, A. C., & Mellor, D. T.
(2018). The preregistration revolution. Proceedings of the
National Academy of Sciences, USA, 115(11), 2600–2606.
https://doi.org/10.1073/pnas.1708274114
Popper, K. R. (1959). The logic of scientific discovery. Basic Books.
Turri, J., Buckwalter, W., & Blouw, P. (2015). Knowledge
and luck. Psychonomic Bulletin & Review, 22(2), 378–390.
https://doi.org/10.3758/s13423-014-0683-5
Wagge, J. R., Brandt, M. J., Lazarevic, L. B., Legate, N.,
Christopherson, C., Wiggins, B., & Grahe, J. E. (2019).
Publishing research with undergraduate students via repli-
cation work: The Collaborative Replications and Education
Project. Frontiers in Psychology, 10, Article 439232. https://
doi.org/10.3389/fpsyg.2019.00247
Weinberg, J. M., Nichols, S., & Stich, S. (2001). Normativity and
epistemic intuitions. Philosophical Topics, 29(1), 429–460.
https://doi.org/10.5840/philtopics2001291/217
Wicherts, J. M., Veldkamp, C. L. S., Augusteijn, H. E. M.,
Bakker, M., van Aert, R. C. M., & van Assen, M. A. L. M.
(2016). Degrees of freedom in planning, running, analyz-
ing, and reporting psychological studies: A checklist to
avoid p-hacking. Frontiers in Psychology, 7, Article 1832.
https://doi.org/10.3389/fpsyg.2016.01832
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
According to the justified true belief (JTB) account of knowledge, people can truly know something only if they have a belief that is both justified and true (i.e., knowledge is JTB). This account was challenged by Gettier, who argued that JTB does not explain knowledge attributions in certain situations, later called “Gettier-type cases,” wherein protagonists are justified in believing something to be true, but their belief was correct only because of luck. Laypeople may not attribute knowledge to protagonists with justified but only luckily true beliefs. Although some research has found evidence for these so-called Gettier intuitions, Turri et al. found no evidence that participants attributed knowledge in a counterfeit-object Gettier-type case differently than in a matched case of JTB. In a large-scale, cross-cultural conceptual replication of Turri and colleagues’ Experiment 1 (N = 4,724) using a within-participants design and three vignettes across 19 geopolitical regions, we did find evidence for Gettier intuitions; participants were 1.86 times more likely to attribute knowledge to protagonists in standard cases of JTB than to protagonists in Gettier-type cases. These results suggest that Gettier intuitions may be detectable across different scenarios and cultural contexts. However, the size of the Gettier intuition effect did vary by vignette, and the Turri et al. vignette produced the smallest effect, which was similar in size to that observed in the original study. Differences across vignettes suggest that epistemic intuitions may also depend on contextual factors unrelated to the criteria of knowledge, such as the characteristics of the protagonist being evaluated.
Article
Full-text available
Concerns about the veracity of psychological research have been growing. Many findings in psychological science are based on studies with insufficient statistical power and nonrepresentative samples, or may otherwise be limited to specific, ungeneralizable settings or populations. Crowdsourced research, a type of large-scale collaboration in which one or more research projects are conducted across multiple lab sites, offers a pragmatic solution to these and other current methodological challenges. The Psychological Science Accelerator (PSA) is a distributed network of laboratories designed to enable and support crowdsourced research projects. These projects can focus on novel research questions or replicate prior research in large, diverse samples. The PSA’s mission is to accelerate the accumulation of reliable and generalizable evidence in psychological science. Here, we describe the background, structure, principles, procedures, benefits, and challenges of the PSA. In contrast to other crowdsourced research networks, the PSA is ongoing (as opposed to time limited), efficient (in that structures and principles are reused for different projects), decentralized, diverse (in both subjects and researchers), and inclusive (of proposals, contributions, and other relevant input from anyone inside or outside the network). The PSA and other approaches to crowdsourced psychological science will advance understanding of mental processes and behaviors by enabling rigorous research and systematic examination of its generalizability.
Article
Full-text available
The designing, collecting, analyzing, and reporting of psychological studies entail many choices that are often arbitrary. The opportunistic use of these so-called researcher degrees of freedom aimed at obtaining statistically significant results is problematic because it enhances the chances of false positive results and may inflate effect size estimates. In this review article, we present an extensive list of 34 degrees of freedom that researchers have in formulating hypotheses, and in designing, running, analyzing, and reporting of psychological research. The list can be used in research methods education, and as a checklist to assess the quality of preregistrations and to determine the potential for bias due to (arbitrary) choices in unregistered studies.
Article
Full-text available
Nearly all success is due to some mix of ability and luck. But some successes we attribute to the agent's ability, whereas others we attribute to luck. To better understand the criteria distinguishing credit from luck, we conducted a series of four studies on knowledge attributions. Knowledge is an achievement that involves reaching the truth. But many factors affecting the truth are beyond our control, and reaching the truth is often partly due to luck. Which sorts of luck are compatible with knowledge? We found that knowledge attributions are highly sensitive to lucky events that change the explanation for why a belief is true. By contrast, knowledge attributions are surprisingly insensitive to lucky events that threaten, but ultimately fail to change the explanation for why a belief is true. These results shed light on our concept of knowledge, help explain apparent inconsistencies in prior work on knowledge attributions, and constitute progress toward a general understanding of the relation between success and luck.
Article
Full-text available
The many criticisms of null hypothesis testing suggest when it is not useful and what is should not be used for. This article explores when and why its use is appropriate. Null hypothesis testing is insufficient when size of effect is important, but it is ideal for testing ordinal claims relating the order of conditions, which are common in psychology. Null hypothesis testing also is insufficient for determining beliefs, but it is ideal for demonstrating sufficient evidential strength to support an ordinal claim, with sufficient evidence being 1 criterion for a finding entering the corpus of legitimate findings in psychology. The line between sufficient and insufficient evidence is currently set at p < .05; there is little reason for allowing experimenters to select their own value of alpha. Thus null hypothesis testing is an optimal method for demonstrating sufficient evidence for an ordinal claim.
Article
Progress in science relies in part on generating hypotheses with existing observations and testing hypotheses with new observations. This distinction between postdiction and prediction is appreciated conceptually but is not respected in practice. Mistaking generation of postdictions with testing of predictions reduces the credibility of research findings. However, ordinary biases in human reasoning, such as hindsight bias, make it hard to avoid this mistake. An effective solution is to define the research questions and analysis plan before observing the research outcomes-a process called preregistration. Preregistration distinguishes analyses and outcomes that result from predictions from those that result from postdictions. A variety of practical strategies are available to make the best possible use of preregistration in circumstances that fall short of the ideal application, such as when the data are preexisting. Services are now available for preregistration across all disciplines, facilitating a rapid increase in the practice. Widespread adoption of preregistration will increase distinctiveness between hypothesis generation and hypothesis testing and will improve the credibility of research findings.
Article
Intuitively, there is a difference between knowledge and mere belief. Contemporary philosophical work on the nature of this difference has focused on scenarios known as "Gettier cases." Designed as counterexamples to the classical theory that knowledge is justified true belief, these cases feature agents who arrive at true beliefs in ways which seem reasonable or justified, while nevertheless seeming to lack knowledge. Prior empirical investigation of these cases has raised questions about whether lay people generally share philosophers' intuitions about these cases, or whether lay intuitions vary depending on individual factors (e.g. ethnicity) or factors related to specific types of Gettier cases (e.g. cases that include apparent evidence). We report an experiment on lay attributions of knowledge and justification for a wide range of Gettier Cases and for a related class of controversial cases known as Skeptical Pressure cases, which are also thought by philosophers to elicit intuitive denials of knowledge. Although participants rated true beliefs in Gettier and Skeptical Pressure cases as being justified, they were significantly less likely to attribute knowledge for these cases than for matched True Belief cases. This pattern of response was consistent across different variations of Gettier cases and did not vary by ethnicity or gender, although attributions of justification were found to be positively related to measures of empathy. These findings therefore suggest that across demographic groups, laypeople share similar epistemic concepts with philosophers, recognizing a difference between knowledge and justified true belief.
Article
In epistemology, fake-barn thought experiments are often taken to be intuitively clear cases in which a justified true belief does not qualify as knowledge. We report a study designed to determine whether non-philosophers share this intuition. The data suggest that while participants are less inclined to attribute knowledge in fake-barn cases than in unproblematic cases of knowledge, they nonetheless do attribute knowledge to protagonists in fake-barn cases. Moreover, the intuition that fake-barn cases do count as knowledge is negatively correlated with age; older participants are less likely than younger participants to attribute knowledge in fake-barn cases. We also found that increasing the number of defeaters (fakes) does not decrease the inclination to attribute knowledge.