ArticlePDF Available

Test Anxiety and Metacognitive Performance in the Classroom

Authors:

Abstract and Figures

Test anxiety is a context-specific academic anxiety which can result in poorer academic and metacognitive performance. We assessed how the quantity and relative weight of assessments contribute to the effects of test anxiety on performance and metacognitive accuracy in a smaller seminar-style class on human memory (study 1) and a larger lecture-style class on cognitive psychology (study 2). Students took six low-stakes quizzes each worth 10% of their final grade in study 1 and two high-stakes exams each worth 40% of their final grade in study 2. All students provided their state anxiety and predicted their scores before and after each assessment. Students in both classes also provided their trait (overall) anxiety after the final assessment. In both studies, students' higher post-state anxiety appeared to be associated with worse assessment performance; however, pre-and post-state anxiety decreased across the quarter in study 1 but remained constant in study 2. Additionally, we found that metacognitive accuracy moderated the effect of post-state anxiety on performance in study 1. Students with higher trait anxiety in study 1 were underconfident in their scoring predictions, while in study 2 students with higher trait anxiety performed worse on their assessments. Thus, students' metacognitive accuracy appears to be influenced by trait anxiety when taking low-stakes quizzes, while performance is related to trait anxiety when taking high-stakes exams.
This content is subject to copyright. Terms and conditions apply.
INTERVENTION STUDY
Test Anxiety and Metacognitive
Performance in the Classroom
Katie M. Silaj
1
&Shawn T. Schwartz
1,2
&Alexander L. M. Siegel
1,3
&
Alan D. Castel
1
Accepted: 26 January 2021/
#Springer Science+Business Media, LLC, part of Springer Nature 2021
Abstract
Test anxiety is a context-specific academic anxiety which can result in poorer academic
and metacognitive performance. We assessed how the quantity and relative weight of
assessments contribute to the effects of test anxiety on performance and metacognitive
accuracy in a smaller seminar-style class on human memory (study 1) and a larger
lecture-style class on cognitive psychology (study 2). Students took six low-stakes
quizzes each worth 10% of their final grade in study 1 and two high-stakes exams each
worth 40% of their final grade in study 2. All students provided their state anxiety and
predicted their scores before and after each assessment. Students in both classes also
provided their trait (overall) anxiety after the final assessment. In both studies, students
higher post-state anxiety appeared to be associated with worse assessment performance;
however, pre- and post-state anxiety decreased across the quarter in study 1 but remained
constant in study 2. Additionally, we found that metacognitive accuracy moderated the
effect of post-state anxiety on performance in study 1. Students with higher trait anxiety
in study 1 were underconfident in their scoring predictions, while in study 2 students with
higher trait anxiety performed worse on their assessments. Thus, studentsmetacognitive
accuracy appears to be influenced by trait anxiety when taking low-stakes quizzes, while
performance is related to trait anxiety when taking high-stakes exams.
Keywords Test anxiety .Metacognition .Trait anxiety .Low-stakes testing .Classroom study
Educational Psychology Review
https://doi.org/10.1007/s10648-021-09598-6
*Katie M. Silaj
kmsilaj@ucla.edu
1
Department of Psychology, University of California, Los Angeles, 502 Portola Plaza, Los Angeles,
CA 90095, USA
2
Present address: Department of Ecology and Evolutionary Biology, University of California, Los
Angeles, Los Angeles, CA, USA
3
Present address: Leonard Davis School of Gerontology, University of Southern California, Los
Angeles, CA, USA
Test anxiety is a context-specific academic anxiety in which an individual reacts to an
evaluative event like a quiz, midterm test, or final exam in an anxious, fearful, or nervous
manner (Cassady, 2010). It has been estimated to impact 2540% of students (Cassady, 2010)
and can result in a combination of negative outcomes inducing poor test performance (Cassady
& Johnson, 2002). Test anxiety can be viewed as both a state anxiety that is triggered by the
testing event, but also as a trait anxiety in which the effects of test anxiety are more severe for
people with high levels of baseline anxiety (hereafter trait anxiety; Goetz et al., 2013;Harris
et al., 2019; Spielberger, 1972). According to Zohar (1998), the impact of test anxiety depends
on both the individuals trait anxiety and the situational factors surrounding the test, such as the
individuals level of confidence in their content knowledge or the level of stakes associated
with the test.
Interventions for test anxiety can increase studentsperformance while having no effect on
trait anxiety (Harris et al., 2019). Thus, it is important to consider individual differences in trait
anxiety when investigating the impact of test anxiety on performance.
Furthermore, external factors may also play a role in the effects of test anxiety, such as the
stakes (Wood et al., 2016) and format (Hembree, 1988) of the assessment. For example, higher
test anxiety is associated with poor performance on high-stakes standardized tests (Cassady &
Johnson, 2002; Putwain, 2008), and the multiple-choice testing format is associated with
higher test anxiety when compared to a matching format (Hembree, 1988)oranopen-ended
format (Birenbaum & Feldman, 1998).
Lastly, metacognition, or the ability to assess and control our cognitive processes (Everson
et al., 1994), is related to test anxiety such that students with high test anxiety often show a
lower level of metacognitive skillfulness compared to students with low test anxiety. Some
work has suggested that metacognition may be a mediator between test anxiety and perfor-
mance (Veenman et al., 2000) and between test anxiety and study strategies (Spada et al.,
2006). Hence, the focus of the current study was to examine how self-reported state anxiety
affects student performance and metacognitive accuracy, and how those measures vary with
levels of self-reported trait anxiety along with the number and weight of assessments students
take in a given course.
Test Anxiety and Academic Performance
Test anxiety is associated with poorer academic performance in elementary, high school, and
college students (Cassady & Johnson, 2002;Putwain&Best,2011; Williams, 1991) and is
considered to be one of the most disruptive factors in exam performance (Cizek & Burg, 2006;
von der Embse & Hasson, 2012). High test anxiety is related to lower test scores in general
and, specifically, to poor performance on college entrance exams and standardized tests, which
provides evidence that the stakes of assessments may induce higher levels of test anxiety (von
der Embse et al., 2018). In one study, von der Embse and Hasson (2012) examined the effects
of test anxiety on student performance on a high-stakes state graduation exam in an urban and
a suburban high school. They found that test anxiety accounted for 415% of the variance in
test scores when controlling for school. This result illustrates the wide-ranging effects of test
anxiety and raises the concern that the effects of test anxiety and poor performance on high-
stakes assessments may be compounded for students in school districts that fail to meet
standards for funding. Additionally, poor performance on high-stakes exams could lead to
other negative consequences for students such as low course grades (Segool et al., 2013), less
Educational Psychology Review
attentional control during exams (Fernández-Castillo & Caurcel, 2015), and increased anxiety
and depression (Leadbeater et al., 2012; von der Embse et al., 2018). Thus, it is important to
better understand which factors associated with real classroom environments are related to the
relationship between test anxiety and performance.
Interventions for Test Anxiety
Despite decades of research on test anxiety, there remains a lack of consensus amongst experts
in the field regarding the most effective interventions to combat its detrimental effects on
student performance. One intervention that an instructor might implement to encourage the
development of better study habits and a more positive outlook on testing events is to frame
assessments as retrieval practice.There is a considerable amount of research that supports
the idea that testing oneself is a much more powerful learning event than simply restudying
material (Roediger & Karpicke, 2006), with retrieval practice being described as a desirable
difficulty (Bjork, 1994). The positive implications of testing oneself are often not apparent
until after the learner has taken the assessment and may seem like a challenge to the learner
initially. This knowledge has led researchers in applied educational settings to implement
testing as an intervention to promote student retention of learned content (Karpicke, 2017).
The results of these classroom experiments have led to recommendations that teachers
should use testing as a tool to further student learning (Dunlosky et al., 2013;Dunnetal.,
2013). Perhaps regarding test anxiety, more frequent testing events provided by the instructor
can help the learner to engage in retrieval practice, which may negate the feelings of
anxiousness in response to tests over time. A recent study investigated the effects of retrieval
practice on test anxiety in middle and high school students (Argarwal et al., 2014) and reported
that 72% of students credited retrieval practice with making them feel less nervous for tests and
exams. As such, this may be a promising avenue for reducing test anxiety in a classroom
setting, although it has yet to have been explored within a real-world college setting.
An intervention that may be more easily implemented in an undergraduate education
context is varying the frequency of testing events within a given course. In a recent study,
participants studied five word lists, and then after a distractor task, they either took an interim
test or restudied the words (Yang et al., 2020). Performance on a cumulative test at the end of
the study was significantly higher for participants who took interim tests after each list
compared with the group that restudied the words. Additionally, participants took a Test
Anxiety Inventory (TAI; Chinese version; Yue, 1996) which revealed that the low-stakes
assessments can improve performance regardless of levels of test anxiety. Although this was
demonstrated in a controlled, experimental context, we are currently interested in examining
the effects of format and relative weight of assessments on performance and metacognitive
accuracy in a classroom setting.
Test Anxiety and Metacognitive Accuracy
Test anxiety and metacognition are intimately linked (Spada et al., 2006). Thus, the process of
monitoring oneself throughout the testing cycle may be an important factor in determining
performance outcomes. Heightened anxiety brought upon by metacognitive awareness of a
lack of preparation or ability causes the learner to become more metacognitively aware of their
Educational Psychology Review
failure to succeed on the test, thus inducing more anxiety (Covington, 1985;Naveh-Benjamin,
1991). This can lead to additional cognitive interference because these resources that could
have been devoted to retrieving test-relevant knowledge are essentially divided due to
heightened levels of anxiety (Cassady & Johnson, 2002; Kurosawa & Harackiewicz, 1995;
Veenman et al., 2000). Furthermore, individual differences exist such that students with high
test anxiety may display lower levels of metacognitive abilities when compared to students
with low levels of test anxiety (Veenman et al., 2000).
Veenman et al. (2000) assessed this relationship between metacognition and performance
using the metacognitive word knowledge task. In this task, participants were asked if they
knew the definitions of 31 health science words, and then were asked to provide definitions.
The authors assessed metacognition by computing hits (e.g., the number of words participants
said they knew and correctly identified their meaning on the vocabulary task) and false alarms
(e.g., the number of words participants said they knew but did not correctly identify their
meaning on the vocabulary task) rates. Participants with low test anxiety exhibited a superior
metacognitive skillfulness during math performance relative to participants with high test
anxiety (Veenman et al., 2000). In another study, test anxiety exerted a negative influence
on studentsperformance on a metacognitive word knowledge task, independent of overall
reading ability (Everson et al., 1994).
Metacognitive accuracy can interact with math anxiety such that higher math anxiety
predicts lower test performance except in students that are high in metacognitive skillfulness
(Legg & Locker, 2009). Additionally, when students are more metacognitively aware, they
report higher confidence in their ability to answer test problems correctly. Some work has
shown that with feedback, students can improve their metacognitive accuracy from one exam
to another (Callender et al., 2016). Thus, more retrieval practice opportunities within an
academic term may lead to increased metacognition, which may especially benefit students
that are high in test anxiety.
Overview of the Present Study
In the current study, we examined the effects of self-reported state anxiety on performance and
metacognitive accuracy in a true classroom setting, and how these measures vary with
studentslevels of self-reported trait anxiety and (i) the number and (ii) relative weight of
assessments on a students overall grade within a given course. In a smaller, seminar-style
class in human memory, students took six quizzes that were each worth 10% of their final
grade (study 1), while in a larger, lecture-style class in cognitive psychology, students took two
exams, each worth 40% of their grade (study 2). Thus, study 1 provided a context in which we
could examine the impact of state and trait anxiety on student performance and metacognition
in a smaller, low-stakes setting where students had more opportunities for retrieval practice,
while study 2 provided a context in which we could examine the impact of state and trait
anxiety on student performance and metacognition in a larger, high-stakes environment with
less retrieval practice opportunities.
Students in both classes completed a questionnaire both before and after each assessment
within the span of the academic quarter (11 weeks: 10 instructional weeks + 1 final exam
week). The questionnaires prompted students to predict and postdict their scores on the
assessment and rate their level of state anxiety pre- and post-assessment. Students also gave
a rating of trait anxiety either after the final assessment of the academic quarter (study 1) or
Educational Psychology Review
after each assessment (study 2). We used single-report measures as opposed to a formal
questionnaire to avoid introducing additional anxiety into the testing environment and as a
practical means to address time constraints imposed by the classroom setting. We were
interested in measuring self-reported state anxiety and its relationship with performance and
metacognitive accuracy while controlling for individual differences in self-reported trait
anxiety.
We expected that student scores on assessments and their metacognitive accuracy in
predicting their scores would significantly increase across the quarter for both classes due to
experience with the content and expectations of the course. We also expected state anxiety to
significantly decrease from pre- to post-assessment for students in both classes due to a sense
of relief that the test is over, in addition to across the quarter for students who engaged in more
frequent, relatively lower-stakes assessments as they will gain retrieval practice (Argarwal
et al., 2014). We expected students who report high ratings of trait anxiety to also report
significantly higher ratings of state anxiety and have significantly lower scores on the
assessments compared to students who report low ratings of trait anxiety (Zohar, 1998).
Lastly, we expected metacognition to significantly interact with test anxiety, such that students
with high test anxiety would be less accurate in their assessment score predictions and
postdictions compared to students with low test anxiety (Veenman et al., 2000).
Study 1
Method
Participants
Participants were 49 University of California, Los Angeles (UCLA), undergraduate students
enrolled in Psychology 124C: Human Memory during Winter Quarter 2018. One outlier was
removed due to providing ratings outside of the requested range, so there were 48 students
included in the final sample.
Materials and Procedure
There were six, short-answer format, low-stakes quizzes given during weeks 39(noquiz
week 6). The quizzes covered materials from the lecture and assigned readings from the
previous week. Possible quiz scores ranged from 0 to 10 and the lowest quiz score was
dropped at the end of the quarter. Quizzes accounted for 50% of the final grade (each quiz was
worth 10%). There was also a written assignment, research proposal, and in-class presentation
that in combination contributed the remaining 50% of the final grade (not included in the data
analyzed here). Quiz topics included (in this order) short-term and working memory, encoding
and studying, forgetting and retrieval/testing, autobiographical and eyewitness memory, self-
regulated learning, metamemory, memory and aging, and expertise and applications. An
example of the type of question a student may have answered on a quiz would be of the
following form and level of conceptual understanding: Define and distinguish between
proactive and retroactive interference and give unique examples of each. How can you show
a release of proactive interference, and what does this suggest about the nature of
forgetting?
Educational Psychology Review
Before each quiz, students were asked How nervous/anxious do you feel at the
current moment?and would respond on a Likert scale from 1 (not at all anxious)to
10 (very much anxious). We used this measure to assess student levels of state
anxiety in an effort to minimize any effects of anxiety a full questionnaire might
induce. Single-scale measures have been found to correlate well with more extensive
scales (Davey et al., 2007; Núñez-Peña et al., 2013)andaremorepracticalandless
invasive for our purposes when considering inherent student anxiety and time con-
straints in a more naturalistic, classroom setting.
Students were then asked to predict their scores out of 10 points. They had 45 min
to answer the three short answer questions. After the quiz, students reported their state
anxiety on the same scale a second time. Lastly, they were asked to postdict their
scores out of 10 points. The students were told that their responses to the pre- and
post-quiz surveys were voluntary and would not influence their grades. This procedure
was followed for all six quizzes. At the end of quiz 6 (week 9), students were asked
In general, how anxious of a person do you think you are?and responded on a
Likert scale ranging from 1 (not at all anxious in general)to10(very much anxious
in general). This measure was requested following the final post-quiz state anxiety
rating of the academic quarter and was used to assess individual levels of trait
anxiety.
Additionally, an assigned course readergraded all assignments for the course and was
blind to the hypotheses and predictions of the current study as well as unaware of participants
self-reported ratings when grading the quizzes. All materials and procedures were performed
ethically and approved by the UCLA Institutional Review Board.
Results and Discussion
For the following analyses, we first conducted a repeated-measures analysis of variance
(ANOVA) on each dependent variable while treating quiz number throughout the quarter as
a within-subjects independent measure (for ANOVAs where Mauchlystestofsphericitywas
violated, Greenhouse-Geisser sphericity corrections were used). Then, we conducted both
multiple regression analyses and interrupted regressions (Simonsohn, 2018)totestifquiz
number significantly predicted participantsoverall quiz performance, estimation accuracy,
and anxiety throughout the quarter. We did not assume monotonicity of each relationship, and
therefore tested the plausibility of both the presence of linear and non-linear relationships
between quiz number and each dependent variable (Bowman et al., 1998). Interrupted
regressions were subsequently used to reveal the potential existence of U-shaped (or
inverted U-shaped) relationships and were conducted using the Robin Hoodalgorithm to
identify a break point along measures of the dependent variable (two-lines test, Simonsohn,
2018). Upon examining dependent variables throughout the academic quarter, a U-shaped
relationship might have occurred if, for example, a dependent measure was higher or lower at
both the beginning and end of the academic quarter relative to the middle. Such a pattern may
result from external factors potentially elevating levels of anxiety or performance due to the
nature of uncertainty when becoming acquainted with a new course at the beginning of a term
or knowing that stakes are high for students with borderline grades at the end of the term.
Additionally, the interrupted regression model using the Robin Hoodalgorithm generally
yields higher statistical power for detecting U-shaped relationships than other statistical
alternatives, and also is less likely to result in higher rates of false-positive and false-
Educational Psychology Review
negative results that can occur when assuming and forcing a standard quadratic functional
form on the data (Simonsohn, 2018). In accordance with Simonsohn (2018), we deemed a U-
shaped relationship to exist, and therefore rejected the null hypothesis of an absence of a U-
shaped relationship, if both slopes of the two lines were significant and of opposite sign. For
each dependent variable, if both the results from the multiple regression model and the
interrupted regression model were significant, we opted for the most parsimonious explanation
by choosing the multiple regression (linear) model to explain the apparent relationship
between the data. Additionally, we reported adjusted R2values as percent of variance
explained for all linear regressions.
Quiz Performance Throughout the Quarter
To examine overall quiz performance with respect to quiz number, we conducted a 1 (Exam:
quiz) × 6 (Quiz number: 1, 2, ,6) repeated-measures ANOVA on overall quiz scores, which
revealed a significant main effect of quiz number, F(2.78, 108.50) = 10.03, p<.001,η2=.20
(Table 1). Post hoc paired-samples t-tests with a Bonferroni correction indicated that scores
were higher on quizzes 3 (M=8.58, SD = 1.31), 4 (M=9.35, SD = .62), 5 (M=8.85, SD =
1.03), and 6 (M=8.98, SD = 1.07) relative to quiz 1 (M=7.88, SD = 1.91), higher on quiz 4
(M=9.35, SD = .62) relative to quiz 2 (M=8.45, SD = .95), and higher on quiz 4 (M=9.35,
SD = .62) relative to quiz 3 (M=8.58, SD = 1.31), adjusted ps < .03. No other comparisons
were significant, adjusted ps > .17.
Next, we conducted regressions using both pre-quiz and post-quiz state anxiety as predic-
tors. The multiple regression model took the following form: quiz score = β0+β1(quiz
number) + β2(pre-quiz state anxiety) + β3(post-quiz state anxiety) + β4(trait anxiety). Four
predictors significantly explained 14.8% of the variance, R2=.16, F(4, 269) = 12.84, p<.0001,
and we found that quiz number (β1=.20,p< .0001) and post-quiz state anxiety (β3=.19, p<
.0001) significantly predicted quiz score, but pre-quiz state anxiety (β2=.05,p= .24) and trait
anxiety (β4=.04,p= .27) did not (Table 2). The interrupted regression model took the same
form as the multiple regression model and revealed no significant U-shaped relationship β14=
.38, p<.0001;β56=.12, p= .24). Overall, these findings indicate a significant positive linear
relationship between quiz number and average quiz score, such that when controlling for both
pre-quiz and post-quiz state anxiety along with trait anxiety, the average quiz score for each
student increased with each subsequent quiz throughout the quarter (see Fig. 1).
Table 1 Repeated-measures analysis of variance coefficients in study 1
Outcome Fdf p η2
Quiz performance 10.03 (2.78, 108.50) < .001*** .20
Predictions 4.71 (3.70, 136.95) .002** .11
Postdictions 7.62 (3.17, 104.74) < .001*** .19
Prediction calibration 2.01 (5, 195) .08 .05
Postdiction calibration 1.93 (3.40, 132.55) .12 .05
Overall calibration 2.66 (5, 195) .02* .06
Calibration difference score 4.12 (3.62, 170.26) .004** .08
Pre-quiz state anxiety 4.66 (2.21, 10.26) < .001*** .11
Post-quiz state anxiety 3.16 (5, 170) .009** .09
State anxiety difference sc ore 3.81 (5, 235) .002** .08
*p<.05, **p<.01,***p<.001
Educational Psychology Review
Table 2 Multiple regression analyses coefficients using pre-quiz and post-quiz state anxiety as a predictor in study 1
Outcome % variance explained
Explained
R2Fdf β0β1β2β3β4
Quiz score 14.8% .16**** 12.84 (4, 269) 8.32**** .20**** .05 .19**** .04
Prediction 5.7% .07*** 5.13 (4, 267) 8.33**** .05 .11* .02 .09*
Postdiction 18.5% .20**** 16.39 (4, 267) 9.54**** .02 .09 .29**** .11**
Prediction calibration 8.0% .09**** 6.90 (4, 269) .22 .17* .16* .16* .15**
Postdiction calibration 11.2% .12**** 9.57 (4, 269) 1.20** .19*** .03 .13* .15***
Overall calibration 9.1% .10**** 7.81 (4, 269) .71 .18** .06 .02 .15***
Calibration difference score 7.1% .08**** 6.25 (4, 269) 1.16*** .13** .13** .12** .04
*p<.05, **p<.01,***p< .001, ****p< .0001
Model: Outcome = β0+β1(quiz number) + β2(pre-quiz state anxiety) + β3(post-quiz state anxiety) + β4(trait anxiety)
Educational Psychology Review
Estimation Accuracy Throughout the Quarter
To examine quiz score metacognitive prediction (prior to taking each quiz) with respect to quiz
number, we conducted a 1 (Exam: quiz)×6(Quiznumber:1, 2, ,6) repeated-measures
ANOVA on quiz score prediction, which revealed a significant main effect of quiz number,
F(3.70, 136.95) = 4.71, p=.002, η2=.11(Table1). Post hoc paired-samples t-tests with a
Bonferroni correction indicated that metacognitive score prediction was higher on quiz 4 (M=
7.87, SD = 1.25) than on quiz 1 (M=6.96, SD = 1.94), and lower on quiz 5 (M=7.07, SD =
1.53) than on quiz 4 (M=7.87, SD = 1.25), adjusted ps < .006. No other comparisons were
significant, adjusted ps > .055.
Next, we conducted separate regressions examining quiz score prediction and postdiction,
each using both pre-quiz state anxiety and post-quiz state anxiety as predictors. For quiz score
prediction, the multiple regression model took the following form: quiz score prediction = β0+
β1(quiz number) + β2(pre-quiz state anxiety) + β3(post-quiz state anxiety) + β4(trait
anxiety). The four predictors significantly explained 5.7% of the variance, R2=.07, F(4, 267)
=5.13,p=.0005, and we found that pre-quiz state anxiety significantly predicted quiz score
prediction (β2=.11, p= .04) as did trait anxiety (β4=.09, p=.04),butquiznumber(β1=
.05, p= .36) and post-quiz state anxiety (β3=.02, p= .76) did not. The interrupted regression
model took the same form as the multiple regression model and revealed no significant U-
shaped relationship (β13=.31,p=.05;β46=.10, p= .29) (Table 2). Thus, while there
appeared to be a main effect of quiz number, when controlling for participantsstate anxiety
and trait anxiety, quiz number was no longer a significant predictor of quiz score prediction
(see Fig. 2a).
Secondly, to examine quiz score metacognitive postdiction (directly following taking each
quiz) with respect to quiz number, we conducted a 1 (Exam: quiz) × 6 (Quiz number: 1, 2,,
6) repeated-measures ANOVA on quiz score postdiction, which revealed a significant main
effect of quiz number, F(3.17, 104.74) = 7.62, p<.001, η2=.19(Table1). Post hoc paired-
Fig. 1 Mean quiz score performance in study 1 across the academic quarter. All error bars represent ± 1 standard
error
Educational Psychology Review
samples t-tests with a Bonferroni correction indicated that metacognitive score postdiction was
higher on quizzes 2 (M=8.71, SD = 1.02), 3 (M=8.25, SD = 1.10), and 4 (M=8.32, SD =
1.79) relative to quiz 1 (M=7.37, SD = 1.63), and lower on quiz 5 (M=7.65, SD = 1.70)
compared to quiz 2 (M=8.71, SD = 1.02), adjusted ps < .008. No other comparisons were
significant, adjusted ps > .10.
For quiz score postdiction, the multiple regression model took the following form: quiz
score postdiction = β0+β1(quiz number) + β2(pre-quiz state anxiety) + β3(post-quiz state
anxiety) + β4(trait anxiety). The four predictors significantly explained 18.5% of the variance,
R2=.20, F(4, 267) = 16.39, p<.0001, and we found that post-quiz state anxiety significantly
predicted quiz score postdiction (β3=.29, p< .0001) as did trait anxiety (β4=.11, p=
.005), but quiz number β1=.02, p= .63) and pre-quiz state anxiety (β2=.09,p=.08)did
not. The interrupted regression model took the same form as the multiple regression model and
revealed a significant inverted U-shaped relationship (β12=1.34,p<.0001;β36=.23, p<
.0001) (Table 2). Thus, while there appeared to be a main effect of quiz number, controlling for
participantsstate anxiety and trait anxiety revealed that quiz number was no longer a
significant linear predictor of quiz score postdiction; however, quiz score postdiction was
seemingly lower at both the beginning and end of the academic quarter relative to the middle
as revealed by the significant inverted U-shaped relationship (see Fig. 2b).
We then examined another form of estimation accuracy using studentspre- and postdiction
calibration (actual score predicted/postdicted score; such that a positive value would indicate
underconfidence and a negative value would indicate overconfidence). To examine quiz score
prediction estimation accuracy (for metacognitive predictions provided prior to taking each
quiz) with respect to quiz number, we conducted a 1 (Exam: quiz) × 6 (Quiz number: 1, 2, ,
6) repeated-measures ANOVA on quiz score prediction calibrations, which did not reveal a
significant main effect of quiz number, F(5, 195) = 2.01, p=.08,η2=.05(Table1).
For quiz score prediction calibration, the multiple regression model took the following
form: quiz score prediction calibration = β0+β1(quiz number) + β2(pre-quiz state anxiety) +
β3(post-quiz state anxiety) + β4(trait anxiety). The four predictors significantly explained
8.0% of the variance, R2=.09, F(4, 269) = 6.90, p<.0001, and we found that quiz number
significantly predicted quiz score prediction calibration (β1=.17,p= .01), as did pre-quiz state
anxiety (β2=.16,p= .02), post-quiz state anxiety (β3=.16, p= .01), and trait anxiety (β4=
Fig. 2 Mean quiz score predictions (a) and postdictions (b) of performance in study 1 across the academic
quarter. All error bars represent ± 1 standard error
Educational Psychology Review
.15, p= .003). The interrupted regression model took the same form as the multiple regression
model and revealed no significant U-shaped relationship (β15=.20,p=.03;β6=.01, p=
.97) (Table 2). Overall, these findings indicate a significant positive linear relationship for quiz
score prediction calibration through time, as when controlling for both state and trait anxiety,
the average quiz score prediction calibration for each student increased with each subsequent
quiz throughout the quarter (see Fig. 3a).
Secondly, to examine quiz score postdiction estimation accuracy (for metacognitive
postdictions provided directly following taking each quiz) with respect to quiz number, we
conducted a 1 (Exam: quiz)×6(Quiznumber:1, 2, ,6) repeated-measures ANOVA on quiz
score postdiction calibration, which did not reveal a significant main effect of quiz number,
F(3.40, 132.55) = 1.93, p=.12,η2=.05(Table1).
For quiz score postdiction calibration, the multiple regression model took the following
form: quiz score postdiction calibration = β0+β1(quiz number) + β2(pre-quiz state anxiety)
+β3(post-quiz state anxiety) + β4(trait anxiety). The four predictors significantly explained
Fig. 3 Mean score calibration (actual pre-/postdicted score) before (pre-quiz; panel a) and after (post-quiz;
panel b) each quiz in study 1 across the academic quarter. Mean calibration difference scores (absolute value of
difference between pre-quiz and post-quiz score calibration; panel c) in study 1 across the academic quarter. All
error bars represent ± 1 standard error
Educational Psychology Review
11.2% of the variance, R2=.12, F(4, 269) = 9.57, p<.0001, and we found that quiz number
significantly predicted quiz score postdiction calibration (β1=.19,p= .0005), as did post-quiz
state anxiety (β3=.13,p= .01) and trait anxiety (β4=.15,p= .0005), but pre-quiz state
anxiety (β2=.03, p= .53) did not. The interrupted regression model took the same form as
the multiple regression model and revealed no significant U-shaped relationship (β13=.25,
p=.15;β46=.31,p=.001)(Table2). Similarly, these findings indicate a significant positive
linear relationship for quiz score postdiction calibration through time, as when controlling for
both state and trait anxiety, the average quiz score postdiction calibration for each student
increased with each subsequent quiz throughout the quarter (see Fig. 3b).
Additionally, to test whether or not studentsoverall calibration (the average of the
prediction and postdiction score calibration) was generally underconfident or overconfident,
we conducted a one-sample t-test with the alternative hypothesis that studentsaverage overall
calibration would be greater than 0 (i.e., underconfident), and found this to be the case, t(47) =
6.71, p< .001, Cohensd=.97. We then conducted a 1 (Exam: quiz) × 6 (Quiz number: 1, 2,
,6) repeated-measures ANOVA on studentsoverall calibration, which revealed a signifi-
cant main effect of quiz number, F(5, 195) = 2.66, p=.02,η2=.06(Table1). Post hoc paired-
samples t-tests with a Bonferroni correction indicated no significant differences for overall
quiz score calibration between any other quiz, adjusted ps > .25. Next, we used a multiple
regression model that took the following form: quiz score overall calibration = β0+β1(quiz
number) + β2(pre-quiz state anxiety) + β3(post-quiz state anxiety) + β4(trait anxiety). The
four predictors significantly explained 9.1% of the variance, R2=.10,F(4, 269) = 7.81, p<
.0001, and we found that quiz number significantly predicted quiz score overall calibration (β1
=.18,p= .001) as did trait anxiety (β4=.15,p= .0004), but pre-quiz state anxiety (β2=.06,p
= .28) and post-quiz state anxiety (β3=.02, p=.76)didnot(Table2). The interrupted
regression model took the same form as the multiple regression model and revealed no
significant U-shaped relationship (β15=.20,p=.007;β6=.08, p= .80). These findings
indicate a significant positive linear relationship for quiz score overall calibration through time,
as when controlling for both state and trait anxiety, the overall quiz score calibration for each
student increased with each subsequent quiz throughout the quarter (i.e., students became more
underconfident over time); this finding is consistent with prior work showing that students may
become more underconfident as their performance improves with practice (Koriat et al., 2002).
Finally, to examine how the difference in calibration from pre-quiz to post-quiz changed
throughout time, we computed calibration difference scores (the absolute value of the differ-
ence between pre-quiz and post-quiz calibration) each week, for each student in the course. We
then conducted a 1 (Exam: quiz) × 6 (Quiz number: 1, 2, ,6) repeated-measures ANOVA on
studentscalibration difference scores, which revealed a significant main effect of quiz
number, F(3.62, 170.26) = 4.12, p=.004, η2=.08(Table1). Post hoc paired-samples t-tests
with a Bonferroni correction indicated significantly lower calibration differences scores for
weeks 3 (M=1.05, SD = 1.37), 4 (M=1.01, SD = 1.57), 5 (M=1.01, SD = 1.22), and 6 (M
=.91, SD = 1.32) compared to week 2 (M=1.95, SD = 2.02), adjusted ps < .02. No other
comparisons were significant, adjusted ps > .74. Next, we used a multiple regression model
that took the following form: calibration difference score = β0+β1(quiz number) + β2(pre-
quiz state anxiety) + β3(post-quiz state anxiety) + β4(trait anxiety). The four predictors
significantly explained 7.1% of the variance, R2=.08, F(4, 269) = 6.25, p< .0001, and we
found that quiz number significantly predicted quiz calibration difference scores (β1=.13, p
= .004), as did pre-quiz state anxiety (β2=.13,p= .003) and post-quiz state anxiety (β3=
.12, p= .002), but trait anxiety (β4=.04,p= .17) did not (Table 2). The interrupted
Educational Psychology Review
regression model took the same form as the multiple regression model and revealed no
significant U-shaped relationship (β15=.15, p=.01;β6=.24, p= .38). Thus, we revealed
that metacognitive calibration scores appear to be more discrepant from pre- to post-
quiz earlier on in the quarter, becoming more stable from quiz 3 onward (see Fig. 3c).
Anxiety Throughout the Quarter
To examine pre-quiz state anxiety with respect to quiz number, we conducted a 1 (Exam: quiz)
× 6 (Quiz number: 1, 2, ,6) repeated-measures ANOVA on pre-quiz state anxiety, which
revealed a significant main effect of quiz number, F(2.21, 10.26) = 4.66, p<.001,η2=.11
(Table 1). Post hoc paired-samples t-tests with a Bonferroni correction indicated significantly
lower pre-quiz state anxiety for quizzes 3 (M=5.73, SD = 2.33)and4(M=5.49, SD = 2.20)
relative to quiz 1 (M=6.93, SD = 2.06), adjusted ps < .006. No other comparisons were
significant, adjusted ps > .31. Next, we conducted separate regressions examining pre-quiz and
post-quiz state anxiety, each using both quiz number and trait anxiety as predictors. For pre-
quiz state anxiety, the multiple regression model took the following form: pre-quiz state
anxiety = β0+β1 (quiz number) + β2(trait anxiety). Two predictors significantly explained
17.0% of the variance, R2=.18, F(2, 277) = 29.60, p< .0001, and we found that trait anxiety
significantly predicted pre-quiz state anxiety (β2=.37,p< .0001), but quiz number (β1=.13,
p= .06) did not (Table 3). The interrupted regression model took the same form as the multiple
regression model and revealed no significant U-shaped relationship (β13=.56, p=.009;β4
6=.18,p= .20). Thus, although there was a significant main effect of quiz number, controlling
for trait anxiety revealed there to be no significant linear relationship between quiz number and
pre-quiz state anxiety (see Fig. 4a).
Secondly, to examine post-quiz state anxiety with respect to quiz number, we conducted a 1
(Exam: quiz)×6(Quiznumber:1, 2, ,6) repeated-measures ANOVA on post-quiz state
anxiety, which revealed a significant main effect of quiz number, F(5, 170) = 3.16, p=.009,
η2=.09(Table1). Post hoc paired-samples t-tests with a Bonferroni correction indicated
significantly lower post-quiz state anxiety for quizzes 2 (M=4.76, SD = 2.44) and 4 (M=
4.74, SD = 2.28) relative to quiz 1 (M=5.99, SD = 2.21), adjusted ps<.019.Noother
comparisons were significant, adjusted ps > .056. For post-quiz state anxiety, the multiple
regression model took the following form: post-quiz state anxiety = β0+β1(quiz number) +
β2(trait anxiety). Two predictors significantly explained 8.8% of the variance, R2=.09, F(2,
271) = 14.14, p< .0001, and we found that trait anxiety significantly predicted pre-quiz state
anxiety (β2=.27,p< .0001), but quiz number (β1=.12, p= .12) did not (Table 3). The
interrupted regression model took the same form as the multiple regression model and revealed
no significant U-shaped relationship (β13=.65, p=.004;β46=.09,p= .50). Therefore,
Table 3 Multiple regression analyses coefficients for state anxiety in study 1
Outcome % variance explained R2Fdf Intercept β1β2
Pre-quiz state anxiety 17.0% .18**** 29.60 (2, 277) 4.37**** .13 .37****
Post-quiz state anxiety 8.8% .09**** 14.14 (2, 271) 3.80**** .12 .27****
Overall state anxiety 15.8% .16**** 27.09 (2, 277) 4.12**** .14* .32****
State anxiety difference score 1.0% .02 2.31 (2, 271) 1.78*** .09 .05
*p<.05, **p<.01,***p< .001, ****p<.0001
Model: Outcome = β0+β1(quiz number) + β2(trait anxiety)
Educational Psychology Review
even though there was a main effect of quiz number, controlling for trait anxiety revealed no
significant relationship between quiz number and post-quiz state anxiety (see Fig. 4b).
Thirdly, to examine overall state anxiety with respect to quiz number controlling for trait
anxiety, we used a multiple regression model of the form: overall state anxiety = β0+β1(quiz
number) + β2(trait anxiety). Two predictors significantly explained 15.8% of the variance, R2
=.16, F(2, 277) = 27.09, p< .0001, and we found that quiz number significantly predicted
overall state anxiety (β1=.14, p= .04) as did trait anxiety (β2=.32,p< .0001) (Table 3).
The interrupted regression model took the same form as the multiple regression model and
revealed no significant U-shaped relationship (β15=.91,p<.0001;β610 =0,p= .99). Thus,
we revealed that overall state anxiety appears to decrease throughout the quarter even when
controlling for trait anxiety.
Lastly, to examine how the difference in state anxiety changed from pre-quiz to post-quiz
throughout time, we computed state anxiety difference scores (the absolute value of the
Fig. 4 Mean state anxiety measures before (pre-quiz; panel a) and after (post-quiz; panel b) each quiz in study 1
across the academic quarter. State anxiety difference score (absolute value of difference between pre-quiz and
post-quiz anxiety;panelc). All error bars represent ± 1 standard error
Educational Psychology Review
difference between pre-quiz and post-quiz state anxiety) each week, for each student in the
course. We then conducted a 1 (Exam: quiz) × 6 (Quiz number: 1, 2, ,6) repeated-measures
ANOVA on studentsstate anxiety difference scores, which revealed a significant main effect
of quiz number, F(5, 235) = 3.81, p=.002, η2=.08(Table1). Post hoc paired-samples t-tests
with a Bonferroni correction indicated significantly lower state anxiety differences scores for
weeks 4 (M=1.53,SD = 1.49) and 6 (M=1.44, SD = 1.30) compared to week 2 (M=2.55,
SD = 2.28), adjusted ps < .006. No other comparisons were significant, adjusted ps > .057.
Next, we used a multiple regression model that took the following form: anxiety difference
score = β0+β1(quiz number) + β2(trait anxiety). The two predictors did not significantly
predict state anxiety difference scores, R2=.02,F(2, 271) = 2.31, p=.10; β1=.09, p=.12;
β2=.05,p= .16 (Table 3). The interrupted regression model took the same form as the
multiple regression model and revealed no significant U-shaped relationship (β12=.47,p=
.18; β36=.15, p= .047). Thus, we revealed that state anxiety difference scores tend to vary
with quiz number as indicated by the significant main effect (see Fig. 4c).
Anxiety from Pre-quiz to Post-quiz
To test our hypothesis that state anxiety would decrease from pre- to post-quiz, we conducted a
paired samples t-test comparing pre-quiz anxiety and post-quiz anxiety averaged across all six
quizzes. The t-test was significant, confirming that pre-quiz anxiety (M=5.97,SD = 1.82) was
significantly higher than post-quiz anxiety (M=4.92,SD =1.78);t(47) = 5.39, p<.001,
Cohensd=.78.
Anxiety and Quiz Performance
We first computed Pearsons correlations between overall quiz performance and measures of
trait anxiety, overall state anxiety, pre-quiz anxiety, and post-quiz anxiety. The correlations
between quiz performance and trait anxiety, r(46) =.008, p=.96, quiz performance and
overall state anxiety, r(46) =.28, p=.05, and quiz performance and pre-quiz anxiety, r(46) =
.10, p=.49, were not significant. The correlation between quiz performance and post-quiz
anxiety was significant, r(46) =.42, p=.003, thus indicating that studentsquiz performance
varied inversely with their anxiety after taking each quiz.
Due to the possibility that Pearsons correlations between overall quiz performance and
measures of state anxiety could be inflated due to the overlap of their variances, we computed
partial Pearsons correlations that held trait anxiety constant. The partial correlations between
quiz performance and overall state anxiety, r(45) = .32, p= .03, and post-quiz anxiety, r(45)
=.46, p= .001, were significant, while the partial correlation between quiz performance and
pre-quiz anxiety, r(45) = .11, p= .45, was not significant. These analyses suggest, rather, that
when holding trait anxiety constant, studentsquiz performance varied inversely with both
their overall state anxiety and post-quiz anxiety.
Anxiety and Estimation Accuracy
We first computed Pearsons correlations between overall score calibrations and measures of trait
anxiety, overall state anxiety, pre-quiz anxiety, and post-quiz anxiety. The correlation between
overall score calibrations and trait anxiety was significant, r(46) =.36, p=.01, indicating that
metacognitive accuracy varied with levels of trait anxiety; however, the correlations between
Educational Psychology Review
overall score calibrations and overall state anxiety, r(46) =.16, p=.28, pre-quiz anxiety, r(46) =
.19, p=.19, and post-quiz anxiety, r(46) =.10, p=.51, were not significant.
Due to the possibility that Pearsons correlations between overall score calibrations and
measures of state anxiety could be inflated due to the overlap of their variances, we computed
partial Pearsons correlations that held trait anxiety constant. None of the partial correlations
between overall score calibrations and overall state anxiety, r(45) = .02, p= .87, pre-quiz
anxiety, r(45) = .01, p= .93, and post-quiz anxiety, r(45) = .06, p= .71, were significant.
Thus, when holding trait anxiety constant, we no longer find significant relationships between
any measure of anxiety and estimation accuracy.
Lastly, we examined the relationship between state anxiety and metacognition and how
they might interact to influence overall quiz performance. Table 3illustrates the different
models that were tested. We found there to be a significant moderation effect such that the
effect of post-quiz anxiety on quiz performance seems to significantly depend on participants
metacognitive prediction calibration, R2=.39,F(4, 269) = 43.40, p<.0001;βint =.03,p=.01.
We also found there to be a significant moderation effect such that the effect of post-quiz
anxiety on quiz performance seems to significantly depend on participantsmetacognitive
postdiction calibration, R2= .32, F(4, 269) = 31.71, p< .0001; βint =.05, p= .01.
Consequently, there was also a significant moderation effect of post-quiz anxiety on quiz
performance such that it significantly depends on the metacognitive calibration difference
score, R2=.18,F(4, 269) = 14.75, p< .0001; βint =.05, p= .02. No other interactions were
significant, all ps > .06 (see Table 3).
Study 1 Results Summary
In study 1, quiz performance increased with each subsequent quiz and post-quiz state anxiety
was a predictor of quiz performance, while pre-quiz state anxiety and trait anxiety were not
related to performance. Metacognitive pre- and postdictions did not vary with quiz when
controlling for state and trait anxiety. Metacognitive pre- and postdiction calibration increased
across the quarter when controlling for state and trait anxiety, and post-quiz anxiety was also a
predictor of postdiction calibration. Thus, while studentsperformance increased throughout
the quarter, they became increasingly more underconfident with each subsequent quiz indi-
cating that they were underestimating their performance. Overall state anxiety decreased with
quiz number, but this varied with trait anxiety, such that students with higher trait anxiety had
higher overall state anxiety. Finally, the effect of post-state anxiety on performance was
moderated by pre- and postdiction calibrations and pre-postdiction difference scores (Table 4).
Study 2
Method
Participants
Participants consisted of 299 UCLA undergraduate students enrolled in Psychology 120A:
Cognitive Psychology during Spring Quarter 2018 (92 students were excluded for missing
data and 4 students were excluded as outliers: 203 students included in the final sample.
Excluded outliers ±2.5 SDs from Z-score for overall score calibration).
Educational Psychology Review
Materials and Procedure
There were two, non-cumulative exams administered during weeks 5 and 10. Exams
were 60 multiple-choice questions based on the previous four weeksworth of
information. Scores ranged from 0 to 60, and exams accounted for 80% of the final
grade (each exam worth 40%). An example of the type of multiple-choice question a
student may have answered on an exam would be of the following form and level of
conceptual understanding: Which of the following tasks is least appropriate as a
means of testing implicit memory?There was also a written assignment, research
proposal, and in-class presentation that factored into the final grade. The procedure for
each exam was the same as the procedure for each quiz administered in study 1 and
the same single-report scales were used (pre-quiz anxiety measure, prediction, test,
post-anxiety measure, postdiction). The only difference was that trait anxiety was
measured after each of the two exams. Students were informed that their participation
in the surveys would not affect their test grades.
Similar to study 1, assigned course teaching assistants graded all assignments for
the course and were blind to the hypotheses and predictions of the current study, as
well as studentsmetacognitive ratings and reported anxiety measures. Additionally,
all materials and procedures were performed ethically and approved by the UCLA
Institutional Review Board.
Table 4 How metacognition moderates effects of state anxiety on performance in study 1
Model % variance explained R2Fdf Intercept β1β2β3βint
A 33.3% .34**** 35.80 (4, 275) 8.69**** .14*** .26* .18*** .02
B 29.1% .30**** 29.66 (4, 275) 8.32**** .15*** .23* .11** .02
C 37.7% .39**** 42.36 (4, 269) 8.63**** .13*** .35** .18**** .01
D 36.4% .37**** 40.98 (4, 275) 8.60**** .15*** .16 .17**** .03
E 33.3% .34**** 35.75 (4, 275) 8.27**** .15**** .24** .11*** .02
F 38.3% .39**** 43.40 (4, 269) 8.57**** .14*** .14 .18**** .03*
G 21.1% .22**** 19.64 (4, 275) 8.69**** .17**** .29** .15**** .01
H 16.2% .17**** 14.43 (4, 275) 8.19**** .18**** .22* .06 .004
I 31.0% .32**** 31.71 (4, 269) 8.86**** .12** .60**** .19**** .05*
J 12.1% .13**** 10.59 (4, 275) 8.44**** .19**** .12 .08 .03
K 8.48% .10**** 7.46 (4, 275) 8.10**** .20**** .04 .02 .02
L 16.8% .18**** 14.75 (4, 269) 8.49**** .19**** .14 .09* .05*
*p<.05, **p<.01,***p< .001, ****p<.0001
Significant interactions are bolded. Model A: Quiz score = β0+β1(quiz number) + β2(overall calibration) + β3
(overall state anxiety) + βint (β0*β3). Model B: Quiz score = β0+β1(quiz number) + β2(overall calibration) +
β3(pre-quiz state anxiety) + βint (β2*β3). Model C: Quiz score = β0+β1(quiz number) + β2(overall
calibration) + β3(post-quiz state anxiety) + βint (β2*β3). Model D: Quiz score = β0+β1(quiz number) + β2
(pre-quiz calibration) + β3(overall state anxiety) + βint (β2*β3). Model E: Quiz score = β0+β1(quiz number) +
β2(pre-quiz calibration) + β3(pre-quiz state anxiety) + βint β2*β3). Model F: Quiz score = β0+β1(quiz
number) + β2(pre-quiz calibration) + β3(post-quiz state anxiety) + βint (β2*β3). Model G: Quiz score = β0+β1
(quiz number) + β2(post-quiz calibration) + β3(overall state anxiety) + βint (β2*β3). Model H: Quiz score = β0
+β1(quiz number) + β2(post-quiz calibration) + β3(pre-quiz state anxiety) + βint (β2*β3).ModelI:Quizscore
=β0+β1(quiz number) + β2(post-quiz calibration) + β3(post-quiz state anxiety) + βint (β2*β3). Model J: Quiz
score = β0+β1(quiz number) + β2(calibration difference score) + β3(overall state anxiety) + βint (β2*β3).
Model K: Quiz score = β0+β1(quiz number) + β2(calibration difference score) + β3(pre-quiz state anxiety) +
βint β2*β3). Model L: Quiz score = β0+β1(quiz number) + β2(calibration difference score) + β3(post-quiz
state anxiety) + βint (β2*β3)
Educational Psychology Review
Results and Discussion
Exam Performance
In order to determine if students performed differently between exam 1 and exam 2, we
conducted a paired samples t-test between studentsscores for each exam. Scores improved
from exam 1 (M=49.67, SD = 5.91) to exam 2 (M=51.90, SD = 4.92), t(206) = 6.67, p<
.001, Cohensd=.46.
Estimation Accuracy
We conducted a 2 (Exam: midterm, final)×2(Calibration:pre-exam, post-exam)repeated-
measures ANOVA on overall score calibration to examine how well students metacognitively
estimated their scores before and after each exam (see Fig. 5). There was a significant main
effect of exam such that overall score calibration was significantly more positive (i.e., more
underconfident) for exam 2 (M=2.13, SD = 5.61) compared to exam 1 (M=.33, SD = 5.97),
regardless of whether these metacognitive estimations of accuracy were provided before or
after each exam, F(1, 202) = 13.56, p<.001,η2= .04. These analyses also revealed a
Fig. 5 Mean overall metacognitive score calibration (actual score predicted/postdicted score) measures for
both exams in study 2. More positive scores indicate higher underconfidence. All error bars represent ± 1
standard error
Educational Psychology Review
significant main effect of calibration such that postdiction score calibration (M=1.83, SD =
4.59) was significantly more underconfident than was prediction score calibration (M=.63,
SD = 5.32), F(1, 202) = 22.42, p<.001, η2= .02. Finally, there was a significant interaction
between exam and calibration, F(1, 202) = 9.60, p=.002, η2= .005. Follow-up post hoc
independent samples t-tests with a Bonferroni correction revealed there to be significant
differences between prediction calibration for exam 1 (M=.18, SD = 7.64) and exam 2 (M
=2.20, SD = 7.58), between pre- (M=.18, SD = 7.64) and postdiction (M=1.69, SD =
7.32) calibration for exam 1, and between the prediction calibration for exam 1 (M=.18, SD =
7.64) and the postdiction calibration for exam 2 (M=3.00, SD = 6.84) (adjusted ps<.001).
Anxiety Across Exams
In order to determine whether or not measures of anxiety remained relatively constant from
exam 1 to exam 2, we conducted multiple paired samples t-tests between trait anxiety, pre-
exam anxiety, and post-exam anxiety for exam 1 and exam 2. We found no significant
differences for any measure of anxiety between exam 1 and exam 2, indicating that students
were no more anxious for the final exam compared to the midterm exam: trait anxiety at exam
1(M=6.22, SD = 2.36) vs. exam 2 (M=6.37, SD = 2.34), t(206) = 1.66, p=.10,Cohensd
=.12; pre-exam anxiety for exam 1 (M=6.28, SD = 2.18) vs. exam 2 (M=6.50, SD = 2.39),
t(206) = 1.44, p= .15, Cohensd=.10; post-exam anxiety for exam 1 (M=5.85, SD = 2.39)
vs. exam 2 (M=5.83, SD = 2.44), t(206) = .12, p=.91,Cohensd=.01.
Anxiety from Pre- to Post-exam
To test our hypothesis that anxiety would decrease from pre- to post-exam, we conducted a
paired samples t-test comparing pre-exam anxiety and post-exam anxiety averaged across both
exams. The t-test was significant confirming that pre-exam anxiety (M= 6.35, SD = 2.00) was
significantly higher than post-exam anxiety (M=5.80,SD = 2.18); t(202) = 4.50, p<.001,
Cohensd=.32.
Anxiety and Exam Performance
Similar to study 1, we first computed Pearsons correlations between overall exam perfor-
mance and measures of trait anxiety, overall state anxiety, pre-exam anxiety, and post-exam
anxiety. The correlations between exam performance and trait anxiety, r(201) =.20, p=
.004, overall state anxiety, r(201) =.32, p<.001, pre-exam anxiety, r(201) =.25, p<.001,
and post-exam anxiety, r(201) =.33, p< .001, were all significant, thus indicating that
studentsexam performance varied inversely with all measures of their anxiety before and after
taking each exam.
Then, due to the possibility that Pearsons correlations between overall exam
performance and measures of state anxiety could be inflated due to the overlap of
their variances, we computed partial Pearsons correlations holding trait anxiety
constant. All partial correlations between overall exam performance and overall state
anxiety, r(200) = .26, p< .001, pre-exam anxiety, r(200) = .17, p= .01, and post-
exam anxiety, r(200) = .27, p< .0001, were significant, thus indicating that
studentsexam performance varied inversely with all measures of their
state anxiety before and after taking each exam when holding trait anxiety constant.
Educational Psychology Review
Anxiety and Estimation Accuracy
We also computed Pearsons correlations between overall score calibrations and measures of
trait anxiety, overall state anxiety, pre-exam anxiety, and post-exam anxiety. None of the
correlations between overall score calibrations and trait anxiety, r(201) = .02, p= .73, overall
state anxiety, r(201) = .07, p= .32, pre-exam anxiety, r(201) = .09, p= .19, and post- exam
anxiety, r(201) = .04, p= .59, were significant.
Then, due to the possibility that Pearsons correlations between overall score
calibrations and measures of state anxiety could be impacted by trait anxiety, we
computed partial Pearsons correlations holding trait anxiety constant, and found that
none of the partial correlations between overall score calibrations and overall state
anxiety, r(200) = .07, p= .33, pre-exam anxiety, r(200) = .09, p= .18, and post-
exam anxiety, r(200) = .03, p= .67, were significant. Thus, when holding trait
anxiety constant, we still do not find significant relationships between any measure
of anxiety and estimation accuracy.
Study 2 Results Summary
In study 2, higher state anxiety (pre- and post-quiz and overall) was associated with lower
exam performance when controlling for trait anxiety. Neither state nor trait anxiety impacted
the estimation accuracy of performance for participants. Students also became more
underconfident from exam 1 to exam 2 while their scores increased, indicating that students
were metacognitively unaware of their performance gains.
General Discussion
The goal of the current research was to explore the effects of test anxiety on undergraduate
studentsoverall performance and metacognitive accuracy and how the frequency and stakes
of the assessment influence those effects. In both studies, we found that state anxiety decreased
from before to after assessments, and overall score calibration increased throughout the
academic quarter (see Table 5for a summary of the convergence and divergence of results
from study 1 to study 2). Additionally, higher post-exam state anxiety was associated with
worse assessment performance. Although students were less anxious once they completed the
quiz or exam, they did not recognize that their performance increased. Students were more
anxious after the exam when they knew they had performed poorly (or were not anxious
knowing that they had done well).
In study 1, we found there to be a main effect of quiz number on pre- and post-quiz state
anxiety; however, regression analyses controlling for trait anxiety revealed there to be no
significant linear relationship between quiz number and pre/post-quiz state anxiety. Overall,
students appeared to be more anxious at the beginning of the quarter relative to the middle of
the quarter, with state anxiety generally stabilizing after acclimation to the course. Accord-
ingly, since pre- and post-state anxiety did not decrease with quiz in study 1, this further
suggests that the impact of frequent, low-stakes quizzes on state anxiety may vary with
individual differences in trait anxiety. Similarly, pre-exam state anxiety was constant across
the two exams (middle and end of the term) in study 2.
Educational Psychology Review
In study 1, trait anxiety varied positively with overall score calibration, but was not
associated with quiz performance; however, in study 2, higher trait anxiety was associated
with poorer exam performance, but not with estimation accuracy. Specifically, we found that
studentsmetacognitive predictions increased throughout the term in both studies, in addition
to overall performance increasing throughout the term for both classes. Additionally, in study
1, overall score calibrations increased with each subsequent quiz when controlling for both
state and trait anxiety, which illustrates that students became more underconfident with
practice even though their performance increased across the quarter. Similarly, students
became more underconfident from exam 1 to exam 2 in study 2, despite their increase in
performance.
Metacognitive postdictions increased throughout the term in study 1 but remained un-
changed in study 2. Though, this increase in pre-/postdictions in study 1 can be accounted for
by holding trait anxiety constant in analyses. When considering individual differences in trait
anxiety, pre-postdictions stayed constant across the quarter suggesting that metacognitive
predictions are related to trait anxiety. Additionally, prediction calibration increased from
exam 1 to exam 2 in study 2, and both prediction calibration and postdiction calibration
increased throughout the term in study 1; trait anxiety was a predictor of this increase, but state
anxiety was not related to this increase. This is consistent with our findings that trait anxiety
seems to be related to metacognition when taking frequent, low-stakes quizzes. Though,
postdiction calibration did not significantly change throughout the term in study 2 while
overall score calibration did increase throughout the term for both courses. As such, this
observed change in metacognitive performance throughout the term in both studies reveals a
potential underconfidence with practice effect(Koriat et al., 2002), such that students
became more underconfident throughout the quarter upon completing more assessments.
Students in study 1 with higher trait anxiety exhibited higher metacognitive
underconfidence when taking the more frequent, lower-stakes quizzes, though this trait anxiety
was not related to the studentsoverall performance. Yet, when controlling for trait anxiety,
there does not appear to be a significant relationship between state anxiety and metacognitive
accuracy. Conversely, trait anxiety was not related to studentsmetacognitive accuracy in
study 2, despite its relationship with overall performance. These results reveal a critical
implication for determining testing practices that are more inclusive to individuals with higher
levels of baseline anxiety. The current findings are compelling in that they reveal the
importance of considering studentslevels of test anxiety to motivate the implementation of
Table 5 Comparison of the main findings across studies
Results summary Study 1 Study 2
Assessment performance1++
Overall calibration1++
Pre-assessment state anxiety1No change
Post-assessment state anxiety1No change
Changes in state anxiety2−−
Overall state anxiety and performance3−−
+indicates an increase across time points or a positive correlation while indicates a decrease across time
points or a negative correlation
1Throughout the quarter
2From pre-assessment to post-assessment
3Partial Pearsons correlation controlling for trait anxiety
Educational Psychology Review
more frequent, lower-stakes quizzes in courses. A primary goal for educators and those
designing more modular and adaptable course curricula at all levels of education should be
to consider the various intrinsic (Deci et al., 1999) and extrinsic (Ryan & Weinstein, 2009)
factors that may perpetually enhance (Putwain et al., 2012; Vansteenkiste et al., 2004)or
decline (Ratelle et al., 2007) the studentseducational outcomes in the classroom.
To further examine how metacognitive accuracy influences the relationship between state
anxiety and performance, we ran a series of regression models, which revealed that
metacognitive accuracy was a moderator of the effects of post-quiz state anxiety on perfor-
mance in study 1. Specifically, pre- and postdiction calibrations and pre-postdiction calibration
difference scores moderated the relationship between post-quiz state anxiety on performance
(see Table 4).
Thus, it seems that although we found that higher post-quiz state anxiety is related to lower
performance outcomes, this varies with prediction accuracy, postdiction accuracy, and differ-
ences between pre- and postdiction measures. Individual differences in metacognitive abilities
seem to be related to the effects of post-state anxiety on performance when taking more
frequent, lower-stakes quizzes.
We should also note that although many of our regression models were significant in study
1, R2values were relatively small, explaining between 3 and 17% of the variance amongst the
data. The implications of these results indicate an inherent amount of unexplainable variability
within our dataset, most likely due to the practical considerations that our sample was
constrained to a real, classroom setting comprised of enrolled students in the course, as
opposed to a controlled, randomly sampled in-lab group of recruited participants.
More broadly, integrating optimized tools and techniques to develop more equitable and
customized approaches for different students to achieve learning outcomes, regardless of their
background and preparation, is a critical consideration that must be accounted for in future
course planning. For instance, students from lower socioeconomic backgrounds may already
be prone to higher levels of stress and anxiety (Lantz et al., 2005), which may further instigate
higher levels of test anxiety in these individuals. Additionally, there are intergenerational
effects of certain academic anxieties (Maloney et al., 2015) such that students of parents with
math anxiety experience lower math achievement when their parents have been highly
involved in their math development. These considerations are essential when designing
assessments that provide optimal and equitable conditions for learning achievement. In fact,
it is vital to consider studentsdiverse backgrounds at all stages of the test cycle as learners
with high test anxiety have been shown to have deficits in encoding, organization, and storage
of learned information (Mueller, 1980). These deficits during the test preparation phase often
lead to poor test performance for students with high test anxiety such that anxiety prompts
interference, distractibility, or inefficient cue utilization strategies (Cassady, 2004; Schwarzer
& Jerusalem, 1992). During the test reflection phase, learners typically form attributional
biases, which can lead to negative future attitudes and behavior patterns towards testing. If a
student attributes their failure on the assessment to a lack of ability, they are more likely to
avoid proper test preparation activities for future tests (Elliot & McGregor, 1999).
Some limitations of the current study include only investigating student anxiety both
immediately before and after the assessment without considering other factors that could
potentially explain why these effects were observed. More specifically, we did not consider
how the differences in low-stakes quizzes may prompt students to adjust their study practices
throughout the term (study 1) compared to a different class with a completely different
assessment structure (study 2). Therefore, future research should examine how students
Educational Psychology Review
individual levels of preparation for various types of assessments mediate their levels of post-
quiz anxiety and performance. Additionally, future classroom-based studies examining test
anxiety should seek to collect a variety of demographic factors, along with reports and
measures of studentsefforts and strategies to prepare for quizzes and exams, to subsequently
compare differences in test anxiety both within and across groups of students at all stages of
the testing cycle. These findings, when considered with studentsindividual circumstances,
may provide a framework for designing more equitable assessments that take into account
factors potentially interacting with or provoking studentstest anxiety. It is also worth noting
that although single-report measures were most practical for this study, future research should
consider these relationships between characteristics of assessments (i.e., format, frequency,
and stakes) and state and trait anxiety while using validated scales to measure individual
differences in different types of test anxiety.
In conclusion, the findings of the current study suggest that less-frequent, higher-stakes
assessments may trigger studentsbaseline levels of anxiety in that students may inherently
feel more pressure to perform better on these exams compared to testing environments with
more-frequent, lower-stakes quizzes, as each individual assessment would not necessarily have
as substantial of an impact on a students overall course grade comparatively. This finding
complements prior work showing that students are overwhelmed with stress, anxiety, and
worry due to testing in high-stakes contexts (e.g., Segool et al., 2013; Triplett & Barksdale,
2005).
Though, studentsfeelings regarding testing seem to be inconsistent across studies exam-
ining the relationship between test anxiety and the relative weight of an assessment (Mulvenon
et al., 2001;Mulvenonetal.,2005; Putwain, 2008). Our findings, therefore, contribute to these
prior equivocal results in the literature that prompted us to further investigate how the
relationship between test anxiety and the stakes of an assessment vary with one another, albeit
including an examination of how test anxiety interacts with studentsmetacognitive accuracy.
Acknowledgements We thank Gerardo Ramirez, Barbara Knowlton, Robert Bjork, Elizabeth Bjork, and
Courtney Clark for helpful comments regarding this work. We also thank Mary Hargis, Dillon Murphy, Julia
Schorn, Mary Whatley, and the rest of the members of the Memory and Lifespan Cognition Laboratory at UCLA
for their guidance and support.
Code Availability No custom code or software were written for this project. Code to run analyses in R
are provided in the OSF link below.
Author Contribution A. D. Castel and A. L. M. Siegel conceived of the study. A. L. M. Siegel collected the
data, and S. T. Schwartz analyzed the data with A. L. M. Siegel and K. M. Silaj. K. M. Silaj and S. T. Schwartz
wrote the manuscript. A. L. M. Siegel and A. D. Castel contributed to the editing of the manuscript.
Funding This research was supported in part by the National Institutes of Health (National Institute on Aging;
Award Number R01 AG044335 to Dr. Alan Castel).
Data Availability Data and survey materials are available via OSF (https://osf.io/pv89n/).
Declarations
Conflict of Interest The authors declare no competing interests.
Educational Psychology Review
References
Argarwal, P. K., DAntonio, L., Roediger, H. L. I., McDermott, K. B., & McDaniel, M. A. (2014). Classroom-
based programs of retrieval practice reduce middle school and high school studentstest anxiety. Journal of
Applied Research in Memory and Cognition, 3(3), 131139.
Birenbaum, M., & Feldman, R. A. (1998). Relationships between learning patterns and attitudes towards two
assessment formats. Educational Research, 40(1), 9098.
Bjork, R. A. (1994). Memory and metamemory considerations in the training of human beings. In J. Metcalfe &
A. P. Shimamura (Eds.), Metacognition: knowing about knowing (pp. 185205). MIT Press.
Bowman, A. W., Jones, M. C., & Gijbels, I. (1998). Testing monotonicity of regression. Journal of
Computational and Graphical Statistics, 7(4), 489500.
Callender, A. A., Franco-Watkins, A. M., & Roberts, A. S. (2016). Improving metacognition in the classroom
through instruction, training, and feedback. Metacognition and Learning, 11(2), 215235.
Cassady, J. C. (2004). The impact of test anxiety on text comprehension and recall in the absence of external
evaluative pressure. Applied Cognitive Psychology, 18(3), 311325.
Cassady, J. C. (2010). Test anxiety: contemporary theories and implications for learning. In J. C. Cassady (Ed.),
Anxiety in schools: the causes, consequences, and solutions for academic anxieties (pp. 726). Peter Lang.
Cassady, J. C., & Johnson, R. E. (2002). Cognitive test anxiety and academic performance. Contemporary
educational anxiety, 27,270295.
Cizek, G. J., & Burg, S. S. (2006). Addressing test anxiety in a high-stakes environment: strategies for
classrooms and schools.CorwinPress.
Covington, M. V. (1985). Test anxiety: causes and effects over time. In H. M. van der Ploeg, R.
Schwarzer, & C. D. Spielberger (Eds.), Advances in test anxiety research (Vol. 4, pp. 5568).
Swets & Zeitlinger.
Davey, H. M., Barratt, A. L., Butow, P. N., & Deeks, J. J. (2007). A one-item question with a Likert or Visual
Analog Scale adequately measured current anxiety. Journal of Clinical Epidemiology, 60(4), 356360.
Deci, E. L., Koestner, R., & Ryan, R. M. (1999). A meta-analytic review of experiments examining the effects of
extrinsic rewards on intrinsic motivation. Psychological Bulletin, 125(6), 627668.
Dunlosky, J., Rawson, K. A., Marsh, E. J., Nathan, M. J., & Willingham, D. T. (2013). Improving students
learning with effective learning techniques: promising directions from cognitive and educational psycholo-
gy. Psychological Science in the Public Interest, 14(1), 458.
Dunn, D. S., Saville, B. K., Baker, S. C., & Marek, P. (2013). Evidence-based teaching: tools and techniques that
promote learning in the psychology classroom. Australian. Journal of Psychology, 65(1), 513.
Elliot, A. J., & McGregor, H. A. (1999). Test anxiety and the hierarchical model of approach and avoidance
achievement motivation. Journal of Personality and Social Psychology, 76(4), 628644.
Everson, H. T., Smodlaka, I., & Tobias, S. (1994). Exploring the relationship of test anxiety and metacognition
on reading test performance: a cognitive analysis. Anxiety, stress, and coping, 7(1), 8596.
Fernández-Castillo, A., & Caurcel, M. J. (2015). State test-anxiety, selective attention and concentration in
university students. International Journal of Psychology, 50(4), 265271.
Goetz, T., Bieg, M., Lüdtke, O., Pekrun, R., & Hall, N. C. (2013). Do girls really experience more anxiety in
mathematics? Psychological Science, 24(10), 20792087.
Harris, R. B., Grunspan, D. Z., Pelch, M. A., Fernandes, G., Ramirez, G., & Freeman, S. (2019). Can test anxiety
interventions alleviate a gender gap in an undergraduate STEM course? CBELife Sciences Education, 18,
19.
Hembree, R. (1988). Correlates, causes, effect, and treatment of test anxiety. Review of Educational Research,
58(1), 4777.
Karpicke, J. D. (2017). Retrieval-based learning: a decade of progress. In J. T. Wixted (Ed.), Cognitive
psychology of memory, of learning and memory: a comprehensive reference (Vol. 2, pp. 487514).
Oxford: Academic Press.
Koriat, A., Sheffer, L., & Maayan, H. (2002). Comparing objective and subjective learning curves: judgments of
learning exhibit increased underconfidence with practice. Journal of Experimental Psychology: General,
131(2), 147162.
Kurosawa, K., & Harackiewicz, J. M. (1995). Test anxiety, self-awareness, and cognitive interference: a process
analysis. Journal of Personality, 63(4), 931951.
Lantz, P. M., House, J. S., Mero, R. P., & Williams, D. R. (2005). Stress, life events, and socioeconomic
disparities in health: results from the AmericansChanging Lives Study. Journal of Health and Social
Behavior, 46(3), 274288.
Educational Psychology Review
Leadbeater, B., Thompson, K., & Gruppuso, V. (2012). Co-occurring trajectories of symptoms of anxiety,
depression, and oppositional defiance from adolescence to young adulthood. Journal of Clinical Child &
Adolescent Psychology, 41(6), 719730.
Legg, A. M., & Locker, L. (2009). Math performance and its relationship to math anxiety and metacognition.
North American Journal of Psychology, 11,471486.
Maloney, E. A., Ramirez, G., Gunderson, E. A., Levine, S. C., & Beilock, S. L. (2015). Intergenerational effects
of parentsmath anxiety on childrens math achievement and anxiety. Psychological Science, 26(9), 1480
1488.
Mueller, J. H. (1980). Test anxiety and the encoding and retrieval of information. In I. G. Sarason (Ed.), Test
anxiety: theory, research, and applications (pp. 6386). Erlbaum.
Mulvenon, S. W., Connors, J. V., & Lenares, D. (2001). Impact of accountability and school testing on students:
is there evidence of anxiety?.
Mulvenon, S. W., Stegman, C. E., & Ritter, G. (2005). Test anxiety: a multifaceted study on the perceptions of
teachers, principals, counselors, students, and parents. International Journal of Testing, 5(1), 3761.
Naveh-Benjamin, M. (1991). A comparison of training programs intended for different types of test anxious
students: further support for an information-processing model. Journal of Educational Psychology, 83(1),
134139.
Núñez-Peña, M. I., Suárez-Pellicioni, M., Guilera, G., & Mercadé-Carranza, C. (2013). A Spanish version of the
short Mathematics Anxiety Rating Scale (sMARS). Learning and Individual Differences, 24, 204210.
Putwain, D. W. (2008). Deconstructing test anxiety. Emotional and Behavioural Difficulties, 13(2), 141155.
Putwain, D. W., & Best, N. (2011). Fear appeals in the primary classroom: effects on test anxiety and test grade.
Learning and Individual Differences, 21(5), 580584.
Putwain, D. W., Kearsley, R., & Symes, W. (2012). Do creativity self-beliefs predict literacy achievement and
motivation? Learning and Individual Differences, 22(3), 370374.
Ratelle, C. F., Guay, F., Vallerand, R. J., Larose, S., & Senécal, C. (2007). Autonomous, controlled, and
amotivated types of academic motivation: a person-oriented analysis. Journal of Educational Psychology,
99(4), 734746.
Roediger, H. L., & Karpicke, J. D. (2006). Test-enhanced learning: taking memory tests improves long-term
retention. Psychological Science, 17(3), 249255.
Ryan, R. M., & Weinstein, N. (2009). Undermining quality teaching and learning: a self- determination theory
perspective on high-stakes testing. Theory and Research in Education, 7(2), 224233.
Schwarzer, R., & Jerusalem, M. (1992). Advances in anxiety theory: a cognitive process approach. In K. A.
Hagtvet & T. B. Johnsen (Eds.), Advances in test anxiety research (Vol. 7, pp. 231). Swets & Zeitlinger.
Segool, N. K., Carlson, J. S., Goforth, A. N., von der Embse, N., & Barterian, J. A. (2013). Heightened test
anxiety among young children: elementary school studentsanxious responses to high-stakes testing.
Psychology in the Schools, 50(5), 489499.
Simonsohn, U. (2018). Two lines: a valid alternative to the invalid testing of U-shaped relationships with
quadratic regressions. Advances in Methods and Practices in Psychological Science, 1(4), 538555.
Spada, M. M., Nikcevic, A. V., Moneta, G. B., & Ireson, J. (2006). Metacognition as a mediator of the effect of
test anxiety on a surface approach to studying. Educational Psychology, 26(5), 615624.
Spielberger, C. (1972). Anxiety as an emotional state. In C. D. Spielberger (Ed.), Anxiety: current trends in
theory and research (Vol. 1, pp. 2349). Academic Press.
Triplett, C. F., & Barksdale, M. A. (2005). Third through sixth gradersperceptions of high- stakes testing.
Journal of Literacy Research, 37(2), 237260.
Vansteenkiste, M., Simons, J., Lens, W., Sheldon, K. M., & Deci, E. L. (2004). Motivating learning, perfor-
mance, and persistence: the synergistic effects of intrinsic goal contents and autonomy-supportive contexts.
Journal of Personality and Social Psychology, 87(2), 246260.
Veenman, M. V. J., Kerseboom, L., & Imthorn, C. (2000). Test anxiety and metacognitive skillfulness:
availability versus production deficiencies. Anxiety, Stress, and Coping, 13(4), 391412.
von der Embse, N., & Hasson, R. (2012). Test anxiety and high-stakes test performance between school settings:
implications for educators. Preventing school failure: alternative education for children and youth, 56(3),
180187.
von der Embse, N., Jester, D., Roy, D., & Post, J. (2018). Test anxiety effects, predictors, and correlates: a 30-
year meta-analytic review. Journal of Affective Disorders, 227,483493.
Williams, J. E. (1991). Modeling test anxiety, self-concept and high school studentsacademic achievement.
Journal of Research & Development in Education, 25(1), 5157.
Wood, S. G., Hart, S., Little, S., & Phillips, S. A. (2016). Test anxiety and a high-stakes standardized reading
comprehension test: a behavioral genetics perspective. Merrill-Palmer Quarterly, 62(3), 233251.
Educational Psychology Review
Yang, C., Sun, B., Potts, R., Yu, R., Luo, L., & Shanks, D. R. (2020). Do working memory capacity and test
anxiety modulate the beneficial effects of testing on new learning? Journal of Experimental Psychology:
Applied Advance online publication.
Yue, X. (1996). Test anxiety and self-efficacy: Levels and relationship among secondary school students in Hong
Kong. Psychologia: An International Journal of Psychology in the Orient, 39(3), 193202.
Zohar, D. (1998). An additive model of test anxiety: role of exam-specific expectations. Journal of Educational
Psychology, 90(2), 330340.
PublishersNote Springer Nature remains neutral with regard to jurisdictional claims in published maps
and institutional affiliations.
Educational Psychology Review
... vida (Gonzaga et al., 2016), gerando pensamentos intrusivos, preocupações e medo de fracassar (Klug et al., 2019). A ansiedade frente a provas é uma condição psicológica experimentada quando os indivíduos se deparam com situações avaliativas, aumentando à medida que a prova se aproxima (Lotz & Sparfeldt, 2017;Silaj et al., 2021). ...
... Compreender essa ansiedade é importante para potencializar o desenvolvimento escolar do estudante e prevenir problemas emocionais, sintomatologias e quadros psicopatológicos, como ansiedade e depressão (Cizek & Burg, 2006;Putwain et al., 2021;Raymo et al., 2019). No contexto educacional, há interesse contínuo em compreender os fatores cognitivos e comportamentais que influenciam o desempenho acadêmico dos alunos (Brown et al., 2008;Silaj et al., 2021). ...
Article
Full-text available
Evaluative contexts tend to be anxiogenic for most people. Thus, understanding levels of anxiety in the face of tests is a possibility to help individual students, as well as to understand broader educational processes. This study aimed to evaluate evidence of internal structure validity, accuracy and invariance in the short version of the Test Anxiety Inventory (IAP). The study was carried out with 1137 students aged between 18 and 63 years (M = 25.94; SD = 7.37) in the state of Rio de Janeiro. The results indicated a structure formed by four factors: concern, lack of confidence, distraction and emotion. Good precision estimates were also evidenced, with alphas ranging from 0.86 to 0.93. Evidences of gender invariance were presented for the IAP, with women presenting higher levels of anxiety, and focal actions can be considered, mainly in the female public. The short version of the inventory is a good resource for assessments in the educational context and helps to understand anxiety in these situations. New studies are suggested to look for evidence of criterion validity and control of response biases in the IAP.
... We defined three categories based on calibration scores (Silaj et al., 2021) representing high accuracy (reference category), underconfidence, and overconfidence. ...
... Fifth, regarding metacognitive accuracy, we only assessed prospective but not retrospective confidence judgments. Additionally, we categorized the determined calibration score (see Silaj et al., 2021) to be able to interpret the results from the conducted GLMMs. However, this treatment might have resulted in information loss. ...
Article
Full-text available
With the ongoing digitalization of home automation and appliances, users often face situations in which they lack the information required to reach an intended goal when using an unfamiliar device. To solve such technical problems, users need to acquire and apply operational knowledge in a self-regulated way. Our study assesses skills in technical problem solving (TPS) and provides the first empirical validation of a score interpretation of a developed TPS test. The examined hypotheses concern the internal structure of the TPS test and the association of the item solution rate with device characteristics (complexity and intransparency) and person-by-device characteristics (the extent of system exploration, use of operating manuals, and device-specific prior knowledge). An ad hoc adult sample (n = 213) processed nine computer-simulated TPS units, each including a knowledge acquisition and knowledge application phase per device. As expected, the internal structure of the TPS test was one-dimensional. Intransparency, but not complexity, affected the item solution rate. Completeness of system exploration, use of operating manuals, and device-specific prior knowledge independently showed positive effects on item solution rates. Along with an evaluation of the developed TPS test, we discuss the role of the salience of device features for knowledge acquisition.
... Although frequent testing can strain students and potentially cause continuous stress, it may not induce more stress than high-stakes exams (Silaj et al., 2021;Yang et al., 2023). Additionally, frequent testing can burden teachers with increased marking time, leading to higher departmental costs (Bekki et al., 2012;Collins et al., 2019). ...
... Despite the negative consequence that the PSTs felt stressed by the weekly tests, at least in the beginning of the semester, the outcome of the new assessment model is predominantly positive. Even though each test caused some stress, it is possible that the overall perceived stress over the semester was not greater than if a final written high-stakes exam had a greater impact on the final grade (Silaj et al., 2021;Yang et al., 2023). Furthermore, the weekly tests made the students start studying from the course start (Roediger & Karpicke, 2006), and they also had the potential to make the PSTs less nervous for the two exams at the end of the course. ...
Article
Full-text available
This paper reports on a revision of the assessment model from the first mathematics course for pre-service teachers (PSTs) aiming for grades 5–10, at a Norwegian university. The weight of the final written exam was reduced and a new, mastery-based testing model, with weekly small tests, was introduced. Results from this study show that the PSTs were positive about the new assessment form, which contributed to a feeling of mastering the subject, although they also expressed some inconveniences, such as the weekly testing could be perceived as stressful. Further, the continuous testing made the PSTs start studying from the start of the semester and not postponing it to the last weeks before the final exam. It is likely that this contributed to the improved result on the national exam at the end of the semester, which was significantly better than the results from the previous years.
... Our results lead us to hypothesise that students who more precisely monitor their own internal states tend towards individualism as a coping mechanism when facing uncertainty in an anxiogenic context. While the literature mainly describes the effects that anxiety can have on certain aspects of learning, such as reduced confidence (Silaj et al., 2021) or impaired problem-solving skills (Llera & Newman, 2020), some studies show that students with high metacognition or who engage in self-regulatory cognitive strategies are less susceptible to these emotions (Brady et al., 2018;Legg & Locker, 2009). Whether these differences depend on changes in the use of certain strategies (as the results of this article might suggest) warrants further study. ...
... The trainability of metacognition immediately brings to mind the external factors such as stress and anxiety, which can unintentionally influence metacognition. Although the possibility of training the accuracy of metacognitive judgments is not yet confirmed, several studies show that metacognition, particularly ME, can be influenced positively and negatively by stress Massoni et al., 2014;, and are also related to stress reactivity , and students' feelings of anxiety (Barrientos et al., 2022;Silaj et al., 2021). However, the effect of stress and anxiety on students does not seem to be the same for everyone. ...
Thesis
Metacognition is the human capacity to monitor and control our own psychological activity. This capacity is fundamental for several processes that involve the individual being aware of this activity, e.g., learning. Building on a substantial body of theory, metacognition is recognized as one of the main components of learning and a key predictor of academic performance. Despite its acknowledged significance, the literature in the area has yet to explore the specific aspects of metacognition that explain all these relationships. Metacognition is not a unidimensional construct but rather a construct with three facets: knowledge about our own cognitive processes (metacognitive knowledge, MK), real-time monitoring and reflection on these processes (metacognitive experiences, ME), and the regulatory strategies we use to control them (metacognitive skills, MS). In light of empirical evidence highlighting the differences between these facets and their differential role over other processes, the relevance of investigating the individual and collective roles of the facets of metacognition in the learning process becomes apparent. To address this gap, we conducted two original studies. Our first study explores how MK, ME, and MS influence associative learning and academic performance in a sample of undergraduate students. The results show that, while academic performance is related to all facets, associative learning only depends on the quality of the individual's ME. A mediated effect of MK on learning through ME and a moderating effect of the use of MS on the relationship between ME and learning are also observed. Our second study investigates the impact of MK and ME on problem-solving skills within a collaborative context. While the results did not corroborate the main objective of the study, valuable insights were gained regarding the influence of ME on the use of collaborative strategies in the classroom. In summary, our studies affirm the idea that the facets of metacognition play different roles in the learning process and offer valuable information into their relationships within educational contexts.
... An understanding of the complex interplay of these elements can facilitate the development of targeted interventions to alleviate communication apprehension and foster improved performance outcomes. Prior research has demonstrated that communication anxiety has a considerable impact on students' academic performance and engagement, particularly in high-stakes contexts such as presentations, thesis defenses, and oral examinations (Archbell & Coplan, 2021;Silaj et al., 2021). As evidenced by studies conducted by Ayres and Hopf (1993) and Lucas and Stob (2019), students frequently encounter elevated levels of anxiety when compelled to present in the presence of evaluative audiences. ...
Article
Full-text available
Objective. The objective of this study was to examine the strategies employed by students to mitigate communication anxiety during thesis seminar examinations, a high-stress situation commonly encountered by students in communication science programs in Indonesia. Methodology. A qualitative study was conducted involving in-depth interviews with 130 communication science students from two universities in Lampung Province, Indonesia, who were about to undertake thesis seminar examinations. To ensure the integrity and reliability of the findings, triangulation analysis was applied. Results. The analysis revealed a number of strategies that students employ to mitigate anxiety during thesis seminar examinations. These strategies encompassed a range of techniques, including self-soothing practices, controlled breathing, reciting prayers, formulating preparatory notes, simulated presentations (mirror rehearsals), simulated interactions with peers, cognitive restructuring, and experiential learning. Conclusion. This study illuminated the array of strategies that students employ to mitigate anxiety during thesis seminar examinations, offering insights into coping methods well-suited to high-stress academic settings. The findings underscored the importance of integrating cognitive, emotional, and culturally relevant approaches into interventions designed to help students manage anxiety in public speaking contexts. Future research could further explore the effectiveness of these strategies across different academic and cultural settings, informing support programs that enhance students’ resilience and performance.
... Anxiety about a particular subject often stems from a lack of confidence, which can arise if students miss classes, neglect assignments, or refrain from asking questions. This reluctance to admit gaps in their knowledge can exacerbate their anxiety (Silaj et al., 2021). To alleviate this anxiety, it is beneficial to foster an inclusive environment where students are encouraged, provided with numerous opportunities to collaborate, ask questions, and actively participate in the learning process. ...
Article
Full-text available
Culturo-Techno-Contextual Approach (CTCA) is a culturally responsive teaching strategy that empowers African students to learn science meaningfully in the digital age. CTCA integrates culture, technology, and locational context to enhance students' comprehension of scientific concepts. This study examined the effectiveness of CTCA in reducing students learning anxiety towards science concept. A pretest-posttest, non-equivalent group design was adopted. The sample comprised 121 senior secondary II students (equivalent to grade 11) from two purposively selected schools in Lagos State. The Nutrition Anxiety Scale, which had a reliability coefficient of 0.70, was used data collection. The treatment period lasted four weeks following the pretest. The experimental group was taught using the CTCA, while the control group received conventional lecture-based instruction. The results revealed that those students who were taught using the culturo-techno-contextual approach had a significantly lower level of anxiety than the comparison group [F (1, 198) = 16.54; p < 0.05]. A statistically significant difference was not found based on gender [F(1,49)=.49; P >0.05]. Therefore, within the limitations of this study, it was concluded that CTCA is an effective teaching approach for lowering anxiety towards science. It is recommended that CTCA be adopted in teaching scientific concepts.
... Several studies have documented negative relations between students' test anxiety and academic performance (see Seipp, 1991). Later studies have confirmed these results in student and scholastic contexts (Breuer et al., 2023;Cassady, 2004;Cassady & Johnson, 2002;Chapell et al., 2005;Kassim et al., 2008;Silaj et al., 2021). Feelings of worry accompanied by irrelevant thoughts, potentially disrupting task-relevant thinking (Kassim et al., 2008;Sellers, 2000), have been revealed to be negatively related with task performance and positively with cognitive impairment (Sarason, 1984;Seipp, 1991). ...
Article
Assessment methods impact student learning and performance. Various recommendations address challenges of assessment in education, emphasizing test validity and reliability, aligning with ongoing efforts in psychological assessment to prevent test bias, a concern also relevant in evaluating student learning outcomes. Examinations in education commonly use either free-response (FR) or multiple-choice (MC) response formats, each with its advantages and disadvantages. Despite frequent reports of high construct equivalence between them, certain group differences based on differing person characteristics still need to be explained. In this study, we aimed to investigate how test takers’ characteristics and behavior—particularly test anxiety, risk propensity, conscientiousness, lecture attendance, and study time—impact test scores in exams with FR and MC format. Data was collected from 376 students enrolled in one of two Psychology lectures at a large Austrian University at the beginning of the semester and post-exam in a real-life setting. Multilevel analyses revealed that, overall, students achieved higher scores on FR items compared to MC items. Less test anxiety, higher conscientiousness, and more study time significantly increased student examination performance. Lecture attendance impacted performance differently according to the exam items’ response format: Students who attended more lectures scored higher on the MC items compared to the FR items. Risk propensity exhibited no significant effect on exam scores. The results offer deeper insights into the nuanced interplay between academic performance, personality, and other influencing factors with the aim of establishing more reliable and valid performance tests in the future. Limitations and implications of the results are discussed.
Article
Full-text available
Introduction: The present study aimed to compare the effectiveness of brain exercise training and Captain's Log computerized cognitive rehabilitation training on cognitive avoidance and executive functioning of students with learning disabilities. Methods: The method of this study was semi-experimental with a pre-test-post-test design with a control group and a three-month follow-up period. The research population consisted of all 7-9-year-old students (first and third grades) with learning disabilities who were referred to the learning disability centers of Tonekabon city, Iran, in the academic year 2022-2023, of which 45 people were selected by questionnaire and in There were two experimental groups and one control group. The first group underwent eight sessions of brain exercise training, and the second group underwent ten sessions of filial therapy. The research tools were the Executive Functions Questionnaire (2000) and the Cognitive Avoidance Questionnaire (2008). Results: The findings obtained from the mixed variance analysis revealed a significant difference between the groups in the post-test and follow-up stages (P<0.001). Both intervention methods led to a reduction in cognitive avoidance and an increase in executive function, and the effectiveness over time had a lasting effect. Furthermore, the amount of cognitive rehabilitation therapy was more effective than brain exercises on cognitive avoidance and executive function (P<0.001). Conclusion: In general, the research results showed that both intervention methods are effective in reducing cognitive avoidance and improving executive functioning in students with learning disabilities.
Article
Full-text available
Although testing has repeatedly been shown to be one of the most effective strategies for consolidating retention of studied information (the backward testing effect) and facilitating mastery of new information (the forward testing effect), few studies have explored individual differences in the beneficial effects of testing. The current study recruited a large sample (1,032 participants) to explore the potential roles of working memory capacity and test anxiety in the enhancing effects of testing on new learning, and the converse influence of testing on test anxiety. The results demonstrated that administering interim tests during learning appears to be an effective technique to potentiate new learning, regardless of working memory capacity and test anxiety. At a final test on all studied materials, individuals with low working memory capacity benefited more from interim testing than those with high working memory capacity. These testing effects are minimally modulated by levels of trait/state test anxiety, and low-stake interim testing neither reduced nor increased test anxiety. Overall, the results imply that low-stake interim tests can be administered to boost new learning irrespective of learners' level of WMC, test anxiety, and of possible reactive effects of testing on test anxiety. (PsycInfo Database Record (c) 2020 APA, all rights reserved).
Article
Full-text available
Gender gaps in exam scores or final grades are common in introductory college science and engineering classrooms, with women underperforming relative to men with the same admission test scores or college grade point averages. After failing to close a historically documented gender gap in a large introductory biology course using interventions targeted at training a growth mindset, we implemented interventions designed to reduce student test anxiety. We combined evidence-based exercises based on expressive writing and on reappraising physiological arousal. We also used a valid measure to quantify test anxiety at the start and end of the course. This instrument measures an individual’s self-declared or perceived test anxiety—also called trait anxiety—but not the immediate or “state” anxiety experienced during an actual exam. Consistent with previous reports in the literature, we found that women in this population declared much higher test anxiety than men and that students who declared higher test anxiety had lower exam scores than students who declared lower test anxiety. Although the test anxiety interventions had no impact on the level of self-declared trait anxiety, they did significantly increase student exam performance. The treatment benefits occurred in both men and women. These data suggest that 1) a combination of interventions based on expressive writing and reappraising physiological arousal can be a relatively easy manner to boost exam performance in a large-enrollment science, technology, engineering, and mathematics (STEM) course and encourage emotion regulation; 2) women are more willing than men to declare that they are anxious about exams, but men and women may actually experience the same level of anxiety during the exam itself; and 3) women are underperforming in STEM courses for reasons other than gender-based differences in mindset or test anxiety.
Article
Full-text available
When participants studied a list of paired associates for several study-test cycles, their judgments of learning (JOLs) exhibited relatively good calibration on the 1st cycle, with a slight overconfidence. However, a shift toward marked underconfidence occurred from the 2nd cycle on. This underconfidence-with-practice (UWP) effect was very robust across several experimental manipulations, such as feedback or no feedback regarding the correctness of the answer, self-paced versus fixed-rate presentation, different incentives for correct performance, magnitude and direction of associative relationships, and conditions producing different degrees of knowing. It was also observed both in item-by-item JOLs and in aggregate JOLs. The UWP effect also occurred for list learning and for the memory of action events. Several theoretical explanations for this counterintuitive effect are discussed.
Article
Many psychological theories predict U-shaped relationships: The effect of x is positive for low values of x, but negative for high values, or vice versa. Despite implying merely a change of sign, hypotheses about U-shaped functions are tested almost exclusively via quadratic regressions, an approach that imposes an arbitrary functional-form assumption that in some scenarios can lead to a 100% rate of false positives (e.g., the incorrect conclusion that y = log(x) is U shaped). Estimating two regression lines, one for low and one for high values of x, allows testing for a sign change without a functional-form assumption. I introduce the Robin Hood algorithm as a way to set the break point between the lines. This algorithm delivers higher power to detect U shapes than all the other break-point-setting alternatives I compared with it. The article includes simulations demonstrating the performance of the two-lines test and reanalyses of published results using this test. An app for running the two-lines test is available at http://webstimate.org/twolines.
Article
Background and objectives: Since the 1988 publication of Hembree's seminal meta-analysis on test anxiety, much has changed in the theoretical conceptualization of test anxiety, as well as the influences on test performance. The present study examined the influence of test anxiety on a variety of educational outcomes, in addition to demographic and intrapersonal correlates. Design and methods: Results of 238 studies from 1988 to the present, were synthesized via a meta-analytic framework to explicate predictors, correlates, and relationships with the test anxiety construct. Pooled effect sizes were calculated. Results: Results indicated that test anxiety was significantly and negatively related to a wide range of educational performance outcomes, including standardized tests, university entrance exams, and grade point average. Results were most pronounced at the middle grades level. Self-esteem was a significant and strong predictor of test anxiety. Perceived difficulty of the test and the high-stakes nature or consequences of the test was also related to higher test anxiety. Conclusions: The magnitude of the relationship of test anxiety across a number of variables was in the small to moderate range. Implications for researchers and educators are discussed including the use of evidence-based assessment and interventions when warranted.
Article
Past research suggests that reading comprehension test performance does not rely solely on targeted cognitive processes such as word reading, but also on other nontarget aspects such as test anxiety. Using a genetically sensitive design, we sought to understand the genetic and environmental etiology of the association between test anxiety and reading comprehension as measured by a high-stakes test. Mirroring the behavioral literature of test anxiety, three different dimensions of test anxiety were examined in relation to reading comprehension—namely, intrusive thoughts, autonomic reactions, and off-task behaviors. Participants included 425 sets of twins from the Florida Twin Project on Reading. The results indicated test anxiety was negatively associated with reading comprehension test performance, specifically through common shared environmental influences. The significant contribution of test anxiety to reading comprehension on a high-stakes test supports the notion that nontargeted factors might interfere with accurately assessing students' reading abilities.
Chapter
This chapter provides a comprehensive review of the past decade of research on retrieval-based learning. It describes common approaches used to study retrieval practice and outlines theoretical accounts of retrieval-based learning. This chapter reviews research that has manipulated initial retrieval practice activities in a wide variety of ways, extended the benefits of retrieval practice to final assessments that measure educationally meaningful learning outcomes, and generalized retrieval-based learning across learner populations, to different types of materials, and to authentic educational contexts.