ArticlePDF Available

Challenges and opportunities for classroom-based formative assessment and AI: a perspective article

Frontiers
Frontiers in Education
Authors:

Abstract

The integration of artificial intelligence (AI) into educational contexts may give rise to both positive and negative ramifications for teachers’ uses of formative assessment within their classrooms. Drawing on our diverse experiences as academics, researchers, psychometricians, teachers, and teacher educators specializing in formative assessment, we examine the pedagogical practices in which teachers provide feedback, facilitate peer- and self-assessments, and support students’ learning, and discuss how existing challenges to each of these may be affected by applications of AI. Firstly, we overview the challenges in the practice of formative assessment independently of the influence of AI. Moreover, based on the authors’ varied experience in formative assessment, we discuss the opportunities that AI brings to address the challenges in formative assessment as well as the new challenges introduced by the application of AI in formative assessment. Finally, we argue for the ongoing importance of self-regulated learning and a renewed emphasis on critical thinking for more effective implementation of formative assessment in this new AI-driven digital age.
Frontiers in Education 01 frontiersin.org
Challenges and opportunities for
classroom-based formative
assessment and AI: a perspective
article
ThereseN.Hopfenbeck
1,2, 3*, ZhonghuaZhang
1, Sundance
ZhihongSun
1, PamRobertson
1 and JoshuaA.McGrane
1,2
1 Assessment and Evaluation Research Centre, Graduate School of Education, The University of
Melbourne, Parkville, VIC, Australia, 2 Kellogg College, University of Oxford, Oxford, England, 3 The
University of Science and Technology, Trondheim, Norway
The integration of artificial intelligence (AI) into educational contexts may give rise to
both positive and negative ramifications for teachers’ uses of formative assessment
within their classrooms. Drawing on our diverse experiences as academics, researchers,
psychometricians, teachers, and teacher educators specializing in formative
assessment, weexamine the pedagogical practices in which teachers provide feedback,
facilitate peer- and self-assessments, and support students’ learning, and discuss how
existing challenges to each of these may be aected by applications of AI. Firstly,
we overview the challenges in the practice of formative assessment independently
of the influence of AI. Moreover, based on the authors’ varied experience in formative
assessment, wediscuss the opportunities that AI brings to address the challenges in
formative assessment as well as the new challenges introduced by the application
of AI in formative assessment. Finally, weargue for the ongoing importance of self-
regulated learning and a renewed emphasis on critical thinking for more eective
implementation of formative assessment in this new AI-driven digital age.
KEYWORDS
artificial intelligence, formative assessment, self-regulation, critical thinking, classroom
based assessment
Introduction
In an era marked by rapid technological advancements, articial intelligence (AI) is now
increasingly used in diverse sectors of our society, fundamentally transforming the way welive,
work, and learn. Within the eld of educational assessment, the introduction of AI has raised
both concerns and optimisms, particularly with respect to the dynamics around AI and
formative assessment classroom practices. In the current paper, weexplore the opportunities
and challenges AI oers and underscore the continued signicance of self-regulated learning
and critical thinking as essential skills in this AI -driven digital age.
A brief background of classroom-based formative
assessment
Classroom-based assessment has been internationally researched for decades, both
with respect to summative assessments that typically occur at the end of a learning
process (e.g., McMillan, 2013; Brookhart, 2016), as well as formative assessments that
OPEN ACCESS
EDITED BY
Gavin T. L. Brown,
The University of Auckland, NewZealand
REVIEWED BY
Syamsul Nor Azlan Mohamad,
MARA University of Technology, Malaysia
Jason M. Lodge,
The University of Queensland, Australia
Kim Schildkamp,
University of Twente, Netherlands
*CORRESPONDENCE
Therese N. Hopfenbeck
Therese.hopfenbeck@unimelb.edu.au
RECEIVED 01 August 2023
ACCEPTED 02 November 2023
PUBLISHED 23 November 2023
CITATION
Hopfenbeck TN, Zhang Z, Sun SZ,
Robertson P and McGrane JA (2023)
Challenges and opportunities for classroom-
based formative assessment and AI: a
perspective article.
Front. Educ. 8:1270700.
doi: 10.3389/feduc.2023.1270700
COPYRIGHT
© 2023 Hopfenbeck, Zhang, Sun, Robertson
and McGrane. This is an open-access article
distributed under the terms of the Creative
Commons Attribution License (CC BY). The
use, distribution or reproduction in other
forums is permitted, provided the original
author(s) and the copyright owner(s) are
credited and that the original publication in this
journal is cited, in accordance with accepted
academic practice. No use, distribution or
reproduction is permitted which does not
comply with these terms.
TYPE Perspective
PUBLISHED 23 November 2023
DOI 10.3389/feduc.2023.1270700
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 02 frontiersin.org
involve feedback processes promoting students’ learning as it
happens (e.g., Brown, 2018; Lipnevich and Smith, 2018). Black
and Wiliam (1998) emphasized the pivotal role of formative
assessment in providing valuable information not only to
teachers but also to students, guiding improvements in teaching
and learning to optimize student outcomes. Since the publication
of their classic work, they have continued to refine their model
through subsequent theoretical papers (e.g., Black and Wiliam,
1998). Additionally, they have supported their theoretical
insights with empirical studies, documenting the tangible impact
of formative assessment practices on students’ learning within
classroom settings (e.g., Wiliam etal., 2004).
While there is a consensus among researchers regarding the
positive effects of formative assessment on students’ learning
(Hattie 2009; Lipnevich and Smith, 2018), the term “formative
assessment” itself has faced critique for lacking a cohesive
definition. Instead, it has been argued to bea collection of
varied definitions and practices, making it challenging to
conduct rigorous evaluations of its effects (Bennett 2011;
Stobart and Hopfenbeck, 2014). We aim to navigate these
complexities by adopting Black and Wiliam’s (2009) definition
of formative assessment: “Practice in a classroom is formative
to the extent that the evidence about student achievement is
elicited, interpreted, and used by teachers, learners, and their
peers to make decisions about their next steps in instruction
that are likely to bebetter, or better founded than the decisions
they would have taken in the absence of the elicited
evidence” (p.9).
In other words, for a teacher to conduct formative assessment,
they will need to know each student, their learning progress,
and how to support them to achieve their learning goals. In
traditional classrooms, formative assessment challenges teachers,
as it requires them to find ways of following up a whole class or
classes of students and provide individualized feedback to
everyone, either through teacher-assessment, peer assessment,
self-assessment, group-assessment, or by other means (Double
etal., 2020). As wewill discuss in the next section, research has
shown that these practices are difficult to implement at scale and
in ways that are sustainable over time (Hopfenbeck and
Stobart, 2015).
Challenges to implementing formative
assessment
Several challenges to the implementation of formative assessment
have been documented by researchers as presented in the January
2015 Special Issue of the Assessment in Education journal. Wylie and
Lyon (2015) found substantial variation in the quality of implementing
formative assessments among 202 Mathematics and Science teachers
in the US context. ey suggest that more targeted professional
development is needed to secure high-quality implementations of
formative assessment practices. Further, since formative assessment
requires teachers to have high competency across a range of knowledge
and skills (e.g., domain content knowledge, pedagogical content
knowledge, assessment and data literacy, and knowledge of
measurement fundamentals), such professional development needs to
bewide in scope.
Challenges have also been found when stakeholders involved
in the assessment do not share a mutual understanding of its
purpose(s). For example, in the same Special Issue, Hopfenbeck
et al. (2015) conducted an evaluation of a large-scale
implementation of an assessment for learning program in Norway
and found that implementation was weaker in schools where the
assessment was perceived as part of an accountability system, while
in schools with a high degree of trust between teachers, head
teachers and the school owners at the municipality level, the
quality of the implementation was better. Similarly, a study of
school-based assessment in Singapore found that their high-stakes
examination-focused education system created tensions when
trying to implement formative assessment processes, thus
demonstrating how context matters in terms of the challenges that
arise between dierent stakeholders in the interaction between
formative assessment and accountability systems (Ratnam-Lim
and Tan, 2015). ese ndings indicate how teachers’ formative
assessment implementations are inuenced by accountability
structures, educational leadership, resources, workload and social
pressures within their context.
irdly, formative assessment practices have primarily been
researched and developed in contexts where students and teachers
have access to a wealth of resources, and, thus, do not necessarily
generalize to more challenging contexts. Halai etal. (2022) evaluated
an implementation of assessment for learning practices in six schools
in Dar es Salaam, Tanzania, and documented the challenges created
by very large class sizes, with one teacher responsible for up to 180
students at a time, as well as under-resourced classrooms.
Furthermore, it was found that the cultural assumptions of the
student role assumed in most of the Western, English-speaking
literature did not t what is seen as a good student in Tanzania.
Formative assessment practices expect students to beself-regulated
and proactive so they can participate in peer-discussions and
assessment, and as part of this, they are supposed to engage in dialog
in groups and with the teacher and beable to ask critical questions.
In contrast, a good student in Tanzania is expected to listen to the
teacher, not ask questions or be too critical, and overall follow
instructions and do what the teacher tells them to do. is is enforced
by the parents’ expectations of how schools and teachers need to help
raise the child. us, the interactive dialog between teachers, students
and peers that are at the heart of formative assessment can
beculturally and contextually sensitive, which poses challenges for
implementing a ‘one size ts all’ formative assessment practice across
dierent contexts.
Finally, studies have reported that teachers nd it challenging to
provide enough feedback to students, particularly at crucial times in
the learning process, as well as with the quality of feedback required
to further each student’s learning, due to time and other resource
constraints (Brooks etal., 2019; Gamlem and Vattøy, 2023). As a
result, formative assessment theory suggests that teachers need to
design classrooms where students can provide feedback to each other
to help reduce this workload (Wiliam, 2011). However, teachers still
report that they struggle to manage classrooms where these peer
assessment practices are established (Dignath etal., 2008; Halai etal.,
2022). So, even in well-resourced contexts where teachers endeavor to
engage best-practices in the implementation of formative assessment
in their classrooms, the high workload such practices engender
continues to bea barrier.
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 03 frontiersin.org
Evolving and revolutionizing formative
assessment: what can AI bring?
Given the challenges to implementing formative assessment in the
classroom that are outlined in the previous section, the advancement
of AI and relevant products (e.g., ChatGPT) may provide opportunities
to overcome some of these challenges, such as high number of
students, and only one teacher to provide feedback. Indeed,
researchers have identied assessment as one of the most signicant
areas of opportunity that AI and related technologies oer in
education (Cope etal., 2021; Swiecki etal., 2022; Zhai and Nehm,
2023). However, new challenges may also beintroduced through the
widespread use of AI, including practical and ethical challenges
(Milano etal., 2023). is section focuses on the changing landscape
of the practice of formative assessment, especially under the inuence
of AI, and discusses how AI can help support teachers to provide
formative assessment for students on a large and sustainable scale.
What do wemean by AI?
In our discussion, AI refers to the application of sophisticated
algorithms that allow computers and machines to simulate human
intelligence for successfully completing tasks (Murphy, 2019).
Although dierent technical approaches and methods, e.g., supervised
learning versus deep learning, have been used to develop AI systems,
the essence of AI is to use data to teach machines to make
classications, diagnoses, decisions, predictions, and/or
recommendations (Gardner et al., 2021). More specically, the
application of AI typically involves collecting large multivariate data
relevant to the task of interest, applying statistical methods and
sophisticated algorithms to process sets of input data to build a
model(s) that identies and weights features and/or patterns of the
input variables relevant to the task, then using a dierent pre-collected
dataset to validate the model(s) for the task where the correct output
is known in advance, and then applying the model(s) to generate a
task output(s) (e.g., classication, prediction, decision, or
recommendation) in a context where the correct output is unknown
(Murphy, 2019; Gardner etal., 2021).
Opportunities for implementing formative
assessment with AI
As mentioned above, one of the challenges in formative
assessment is to provide individualized, high-quality feedback to
students. It is highly resource intensive for teachers to personally give
or nd other ways (e.g., peer assessment, self-assessment, group
assessment) to provide individualized feedback to each student on a
large scale. However, AI can make some of the assessment procedures
fully or partly automated, making the assessment practices more
feasible to maintain, which can then reduce the time burden on
teachers (Swiecki etal., 2022). A typical example that has been widely
discussed is automated essay scoring systems (Ke and Ng, 2019;
Gardner etal., 2021). e application of AI in automated essay scoring
frees teachers from the labor-intensive grading process and allows
them to assign more extended writing tasks to students, automate the
grading process, and, more importantly, with the integration of
natural language processing-based AI, provide timely formative
feedback to help students revise and improve their writing
(Murphy, 2019).
Liu et al. (2016) showed that a machine learning enabled
automated scoring tool, c-rater-ML, could produce scores that were
comparable to human raters in scoring students’ responses to
constructed response questions about science inquiry, oering a
promising solution to improving the eciency of not only obtaining
the summative scores but also generating instant formative feedback
(Linn etal., 2014). Another example of how AI can help is by using
computers to support the management and delivery of formative
assessments (e.g., Webb et al., 2013; Tomasik etal., 2018). ese
systems have the capacity to discern distinct learning pathways in
students’ progress, enabling the identication of the most suitable
tasks or questions for each student at dierent points in time. In
addition, computerized formative assessment systems can optimize
the administration of formative assessments by determining their
frequencies and schedules customized for every individual student
(Shin etal., 2022). ese ndings demonstrate how AI can improve
the eciency and exibility of formative assessment practices at the
individual student level.
Another signicant opportunity that AI oers for formative
assessment is the improvement of feedback both in quantity and
quality (Gardner etal., 2021). e main goals of formative assessment
are to provide constructive feedback based on students’ responses and
to help teachers design dierentiated instructional strategies and
sustain students in self-regulating their learning. AI can delve into the
data to identify the patterns on which dynamic, customized,
individualized, and visualized feedback can beautomatically generated
(Verma, 2018; Tashu and Horvath, 2019; Lee, 2021, 2023). For
example, the adaptive nature of some computerized formative
assessment systems and intelligent tutoring systems enables every
student’s attainment to beindividually and more precisely assessed,
which facilitates more appropriate and targeted feedback based on
their individual learning stage and trajectory (Ma etal., 2014; Tomasik
et al., 2018; Mousavinasab et al., 2021). Adaptive multi-strategy
feedback models, based on AI methods, have been applied in the
context of such systems to automatically adapt the feedback generating
strategy to individual students, which have, in turn, been found to
generate more eective feedback than the traditional feedback
generation methods (Gutierrez and Atkinson, 2011).
In addition, AI can improve the quality and eectiveness of peer
assessment in classrooms with large class sizes. Peer assessment can
besupported with prompts from language models (Er etal., 2021).
is approach to peer assessment supports students in not only
providing feedback to peers but also reecting on and justifying their
judgments, providing further opportunities for them to develop their
self-regulated learning skills (Liu and Carless, 2006), and has been
found to provide useful peer feedback to students (Luaces etal., 2018).
In addition, the peer assessment reviews can both help the teachers to
better understand the performance of the students in their classroom
and also provide additional data (e.g., the review text) that can
beanalyzed using AI-based techniques (e.g., semantic, lexical, and
psycho-linguistic analyses; Vincent-Lamarre and Larivière, 2021) to
further enhance teachers’ understanding of the performance of
their students.
AI can also aid teachers in collecting and analysing longitudinal
formative assessment data, and in generating learner proles to trace
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 04 frontiersin.org
students’ learning progression over time (Swiecki etal., 2022). is
application of AI enables the scalable implementation of formative
assessment in both cross-sectional and longitudinal contexts, which
makes it more sustainable, allowing teachers to eciently monitor the
growth of student learning and identify the knowledge and skill gaps
in their learning over time (Barthakur et al., 2023). Another
contribution of AI to facilitating longitudinal formative assessment
lies in its ability to analyse the large-scale longitudinal formative
assessment data to trace the learning trajectories of the students and
predict their future learning states. For example, some of the widely
applied statistical methods in AI (e.g., hidden-Markov models,
articial neural networks) have been combined with traditional
cognitive diagnostic models (CDM) to analyse longitudinal formative
assessment data to track the changes of students’ learning over time
(e.g., Chen etal., 2018; Wen etal., 2020).
e application of AI also allows teachers to get an in-depth
understanding of students’ learning processes based on the analysis of
large volumes of ‘process data’ rather than just the assessment artefacts
(e.g., responses to questions, items or tasks) produced by the students.
With modern technologies, the processes leading to assessment
responses can becaptured in time-stamped log stream data (Cope
etal., 2021). For example, students’ actions (e.g., keystrokes, editing,
chat history, video watching) during an assessment can becaptured.
Which potentially contain additional information for understanding
how students produce their responses. With the support of AI, the
process data can be analyzed to investigate their strategies (e.g.,
identifying misconceptions), which can then provide invaluable
information for individualized feedback. In addition, taking advantage
of AI’s increasing capacity to deal with complex, multi-media data,
more authentic assessment tasks (e.g., multimedia, game-based
problem solving, essay writing, performance-based tasks) can
beeectively used in undertaking formative assessments (Swiecki
etal., 2022).
AI also oers opportunities for assessing some hard-to-measure
constructs. Students’ non-scholastic attributes, including social–
emotional traits (e.g., classroom engagement, self-ecacy, motivation,
resilience) and social-cognitive skills (e.g., metacognition,
collaborative problem-solving, critical thinking, digital literacy, self-
regulated learning), have been attracting more attention and are
increasingly recognized as equally important as their scholastic
achievement (Durlak and Weissberg, 2011). Advances in AI and
related technologies allow for these constructs to be more validly
assessed, instead of purely relying on students’ self-reported beliefs
and behavior through questionnaires. Now, the data collected through
dierent channels (e.g., time-stamped process data, eye contact,
feedback, facial expression, eye movements, body posture and gesture)
can bemined to develop indicators for assessing these dierent aspects
of student learning. For example, MOOC data has been used to design
indicators through a thorough analysis of students learning behaviors
in online courses to measure students’ self-regulated learning (e.g.,
Milligan and Grin, 2016) and leadership development in workplace
learning in an online environment (e.g., Barthakur et al., 2022).
Another example is the measurement of collaborative problem-solving
skills through process data that captured the actions and chats of pairs
of team members collaboratively solving tasks (Grin and Care,
2014). AI-based Large Language Models provide further promise for
mining chat history data to support assessing how team members
explore, dene, plan, execute and solve tasks in a collaborative way.
Challenges arising from using AI in
formative assessment
AI introduces not only opportunities but also challenges to
formative assessment practices (Swiecki etal., 2022). A primary
challenge that needs to beaddressed before teachers can apply AI
in their formative assessment practices is their lack of knowledge
and skills relevant to AI techniques as well as their limited access
to big data. us, although AI can potentially ease the workload of
teachers by automating some aspects of formative assessment (e.g.,
automatising scoring and tracing students’ learning progress), it
adds further burden through the need for professional development
in its use (Engeness, 2021). Moreover, despite the promising future
for formative assessment brought by big data, with the possibility
of collecting the process data through students’ learning, a new
challenge arises in identifying which part of the collected data is
most helpful and relevant to improve student learning. In addition,
the unique features of current big data (e.g., time-stamped process
data, sparse data) are signicantly dierent from that of the
traditional assessment data and pose a variety of challenges to the
psychometric methods for analyzing the data. To deal with this,
scholars have been endeavoring to introduce new methods to
integrate data science and machine learning into psychometrics
(e.g., von Davier etal., 2022).
Another challenge arising from using AI in formative assessment
is to tackle relevant issues about investigating the best way to use AI
in formative assessment practices. One of the hotly debated issues is
whether AI will replace teachers. We argue that AI should not
replace but facilitate teachers’ formative assessment practices and
promote the role of formative assessments in supporting instruction
and learning. As stated by Murphy (2019), “the best use of AI in
education is to augment teacher capacity by helping teachers deliver
more eective classroom instruction” (p. 14). Teachers need to
understand the limitations of the AI techniques when they review
the assessment results. For example, automated scoring systems have
long been criticized for their inability to measure higher-order
aspects of writing (e.g., creativity, argumentation, reasoning)
(Gardner et al., 2021). One of the primary aims of formative
assessment is to diagnose gaps in students’ learning based on the
well-established interpretability of the measurement scales. However,
many approaches based on machine learning are designed for
prediction involving complicated models for improving accuracy but
sacricing the ease of interpretation. erefore, any inferences from
the results of formative assessments involving the integration of AI
techniques should only be made aer having examined the
assessments’ validity and interpretability (Bejar etal., 2016; Scalise
etal., 2021). Teachers need to critically review how the assessment
results are reached and identify any sources of bias introduced by the
application of AI techniques in assessment, which in turn adds to
their workloads (Murphy, 2019). Finally, but not at least, the
introduction of AI in the classroom cannot happen without ethical
considerations for the use and application of it. Scholars have
emphasized the importance of having conversations with students
on the productive, ethical and critical relationship around the use of
AI and future technologies (Bearman etal., 2023) and improving
knowledge on data privacy for children (Johnston, 2023). With these
considerations in mind, we will now turn to one example of
formative assessment practices and AI.
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 05 frontiersin.org
The formative uses of rubrics and the
opportunities and challenges of using AI
e formative use of rubrics, i.e., the scoring guides that are used
to make judgments about the quality of students’ constructed
responses, such as writing, performances or products (Popham, 1997),
has been shown to have a positive inuence on learning. Specically,
Panadero and Jönsson (2013) argued that the use of quality rubrics
plays a key role in enhancing the eectiveness of formative assessment
practices. ere are two main ways that rubrics are thought to improve
formative assessment. ey make assessment expectations explicit,
thereby assisting with understanding and feedback, and they support
self-regulated learning by supporting learners to monitor their own
learning and make decisions about areas for improvement (Jönsson
and Panadero, 2016). While the use of rubrics for formative assessment
purposes has a positive eect on learning, this eect is amplied when
it is accompanied by teacher-given rubric feedback that addresses the
three feedback questions (Wollenschläger etal., 2016). Consequently,
wenow examine the formative use of rubrics as a specic example of
how AI can address existing challenges, provide additional
opportunities, and present new challenges to this formative
assessment practice.
Firstly, AI can support the formative uses of rubrics by helping
teachers to overcome some of the time needed to construct rubrics
and to teach students how to use them, as these have been found to
bea constraint to rubric use (English etal., 2022). Generative AI can
speed up rubric design, as teachers can use it to provide samples of
rubrics to assess specied constructs, and a teacher can choose to
directly use these rubrics or to use them as a source of ideas for
designing their own. AI also has the potential to assist students as they
learn to use rubrics by providing work samples matching dierent
levels on a rubric, by assessing student-generated work samples
against a rubric so a student can check the accuracy of self-
assessments, and by providing written feedback to accompany a rubric
assessment. ese possible AI-augmented rubric uses by students help
build agency, as the students can have more control over the timing
and style of feedback they receive. erefore, AI has the potential to
help teachers overcome some of the common challenges to using
rubrics in the classroom.
Nonetheless, the challenges presented by potential biases in
training data are also applicable to rubrics generated with AI (Li etal.,
2023). Rubrics for constructs with a greater cultural inuence, such as
communication, are likely to be more aected than those for
constructs where the subject matter is more consistent irrespective of
culture, like chemistry. In addition, while there is an acknowledged
need for more research on rubric design (e.g., English etal., 2022), the
ndings of such research oen fail to be commonly adopted by
teachers. One example of this is that most rubrics have structurally
aligned categories, e.g., all parts of the rubric have ve levels of quality.
Unfortunately, there is empirical evidence that this common structure
is ineective because it increases construct-irrelevant variance by
facilitating scoring based on a halo eect where the assessor makes a
global judgment of quality and simply aligns the ratings on dierent
criteria of the rubric to match rather than making independent
decisions for each (Humphry and Heldsinger, 2014). Rubrics, thus,
support more accurate judgments when the number of levels of
quality are tailored to the specic criterion being evaluated where
some criterion (e.g., quality of argumentation) have more levels than
others (e.g., use of paragraphs). Consequently, without careful
curation of training data sets to ensure they meet best practice in
rubric design, AI generated rubrics will likely propagate common
design aws. Moreover, exploration by researchers of the dierent
ways AI is already augmenting rubric use in classrooms is warranted,
especially in cultures and contexts that are not well represented in
training data sets.
The role of self-regulated learning and
critical thinking in formative uses of AI
Based on the formative assessment cycle in Ruiz-Primo and
Brookhart (2018), there is a natural bridge between self-regulated
learning and formative assessment, as formative assessment can
beconsidered as a self- and co-regulated process of improving
learning, which starts with defining and sharing learning goals
and then through a process of gathering or eliciting information,
analysing and interpreting the collected information, and finally
using the collected information to make a reflective judgment on
whether the pre-defined learning goal has been achieved or not.
There has been a call for linking the research into self-regulation
with formative assessment, as it is recognized that self-regulation
will enhance students’ ability to act as peer-assessors, do self-
assessment and take on the proactive role needed for formative
assessment practices (Brandmo etal., 2020). Despite decades of
educational research into what improves students’ learning, few
researchers have tried to combine the two fields of formative
assessment and self-regulation, although exceptions include Allal
(2010), Andrade and Brookhart (2016), Brown (2018), Butler and
Winne (1995), Nicol and MacFarlane-Dick (2006), and Panadero
etal. (2018). Moreover, more recent research has demonstrated
how students can benefit more from formative assessment
practices if they are self-regulated learners (Allal, 2020; Andrade
and Brookhart, 2020; Perry etal., 2020). With the rise of AI,
students’ ability to self-regulate will beeven more important, as
it opens opportunities but also challenges in how weplan, use
strategies, and evaluate our learning processes.
Furthermore, weargue for the importance of critical thinking for
both teachers and learners to navigate the principled use of AI and
leverage the eectiveness of formative assessment as part of their
process of self-regulated learning, particularly when confronted with
the novel challenges of AI. Although highlighting the importance of
critical thinking may seem like an already labored point in educational
settings, as it has been acknowledged as a fundamental generic skill
necessary for individuals to live and thrive in the 21st century (e.g.,
Davies and Barnett, 2015), none of the existing research has yet built
a connection between critical thinking, formative assessment and self-
regulated learning under the impact of AI. Before getting into the
specic argument on the role of critical thinking in formative
assessment and self-regulated learning, it is worth clarifying that self-
regulated learning is used in a broader way in this section, extending
beyond learners to encompass teachers who also need to apply their
self-regulated learning skills to eectively acquire new knowledge and
skills to harness the potential of AI in their teaching practice
eectively. In the following part, we will briey explain our
understanding of critical thinking, depicting the role of critical
thinking when facing new challenges brought by AI, and then
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 06 frontiersin.org
describing the role of critical thinking in formative assessment and
self-regulated learning.
When facing the uncertain, complex issues brought by the
advancement of AI, critical thinking, dened as “reasonable and
reective thinking focused on deciding what to believe or do
(Ennis etal., 2005: 1), becomes increasingly pronounced. As a
complex competency, good critical thinking practitioners are
expected to have a sophisticated level of epistemic beliefs (i.e.,
attitudes to knowledge and knowing), which are essential to
recognizing the uncertainty and complexity of such controversial
issues as the ethical use of AI in education. ese beliefs lay a
foundation for the engagement of their thinking skills (e.g.,
understanding, applying, analysing, synthesizing, evaluating) to
bewell-informed of the issue and navigate through a vast amount
of potentially conicting information (King and Kitchener, 1994;
Kuhn and Weinstock, 2004; Wertz, 2019; Sun, 2021, 2023). As
theorized by Dewey (1910), suspending judgments may bethe
most eective course of action prior to acquiring a comprehensive
understanding of a relevant issue.
For dierent responses on the issue of whether AI should beused
in educational settings, it is not surprising to witness resistance toward
emerging technologies because there is a natural fear and unease that
oen accompanies the introduction of new technologies (Ball and
Holland, 2009). However, if critical thinking is engaged before
deciding what to believe or do, this natural tendency can bechallenged.
AI is far from a novel concept and has been an academic discipline
since the 1950s (Haenlein and Kaplan, 2019; Gillani et al., 2023), and
various AI technologies (e.g., image recognition, smart speakers, self-
driving cars) and models (e.g., AlphaGo, Deep Blue and ELIZA) have
already signicantly impacted our ways of living and working
(Haenlein and Kaplan, 2019). Yet, limited transformation has
happened in education, with 20th-century traditions and practices
still dominating our schools (Luckin and Holmes, 2016). Despite the
recognition of the enormous benets and potential of AI in
transforming education, scholars’ impatience is mounting because
many of these promising ideas remain conned to the lab or lecture
halls with few practical breakthroughs (Luckin and Holmes, 2016).
Specic to the context of formative assessment and self-
regulated learning, the role of critical thinking is also pivotal,
which equips both teachers and learners to eectively address
emerging challenges. For teachers, as the landscape evolves with a
growing array of AI products and assessment data, being ooded
by abundant online learning resources can be overwhelming
(Schwartz, 2020). It would beincreasingly important for teachers
to critically evaluate what, when, and how to utilize these resources
to enhance their teaching methodologies and bolster student
learning. When facing an increasing amount of data that has been
collected or needs to be collected, teachers need to critically
discern how assessment data can best inform their pedagogical
strategies rather than have data to dictate their teaching.
Additionally, teachers should exercise discernment in determining
the level of trust they can place in specic AI models when making
judgments about student learning outcomes. is becomes
especially crucial for teachers who should critically assess the
potential biases that AI models might carry due to the use of
training data (Li et al., 2022, 2023). Furthermore, as AI
advancements have the potential to liberate teachers from routine
and time-consuming tasks like assignment grading and rubric
development, they must engage in critical reection. ey need to
consider which skills they should prioritize for their professional
development, such as data literacy, and what skills should remain
at the core of their teaching, notably critical and creative thinking.
is critical assessment of their evolving role is essential in
navigating the transformative impact of AI in education.
Regarding individual learners, engaging critical thinking can
have positive contribution to the eectiveness of formative
assessment and self-regulated learning when facing the
opportunities and challenges introduced by the advancement of
AI. For instance, the advancement of AI, indeed, can certainly
beused to generate text to pass the assessment of a subject, but if
learners are practicing their critical thinking and self-regulated
learning skills, they may ask themselves some reective questions,
such as what is the purpose of learning? Will a certain way of using
AI contribute to achieving their learning goals? When specic
solutions have not been produced to address the new challenges
brought by AI, individual learner’s practice of their critical
thinking and self-regulated learning may contribute to the ethical
use of new technologies. Despite some instances of learners
exploiting AI to evade plagiarism detection systems, it is
encouraging to learn from recent empirical research that many
students genuinely benet from the timely feedback and
companionship provided by AI (Skeat and Ziebell, 2023).
Moreover, these students display ethical awareness, being cautious
and mindful of their AI usage even in the absence of well-
developed regulations governing AI in education.
While some scholars have suggested that assessment is holding us
back from transforming our education systems (Luckin and Holmes,
2016, p.35), the advancement of AI may catalyze a “Renaissance in
Assessment” (Hill and Barber, 2014). Although the acceptance of AI
may encounter some resistance, the power of new technologies, if
unleashed with principled and research-driven use, may signicantly
change and improve ways of teaching and learning. In this vein,
Australian educational policymakers made a signicant shi by
granting permission for the use of ChatGPT and generative AI in all
government schools. is change followed the release of the Australian
Framework for Generative Articial Intelligence in Schools. It is
encouraging to observe the transition from a policy that limited the
use of ChatGPT across every Australian state and territory, except
South Australia, to a more welcoming and adaptable stance.
Conclusion
As wehave outlined in this article, despite decades of research on
formative assessment practices, teachers still face several challenges in
implementing these practices on a large scale. e use of AI in
classrooms has the potential for supporting formative assessment
practices, although we will argue, it will require some careful
considerations. Based upon what wehave outlined in this article,
wewill conclude with the following suggestions on how to integrate
AI into formative assessment:
1. Utilize AI for feedback assistance, particularly in large classes
where teachers struggle to give timely feedback to all students.
2. Promote self-regulating skills as students will need to take even
more responsibility for their own learning, when using AI. is
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 07 frontiersin.org
includes goal-setting, monitoring progress, and adjusting their
study strategies based upon AI feedback.
3. Emphasize the ethical use of AI in formative assessment and
discuss the importance of integrity and responsible use of AI
tools, to avoid inappropriate uses.
4. Emphasize the role of teachers in guiding the students’ use of
AI. Teacher can help students interpret AI feedback, set
learning goals, and make informed decisions based upon
AI recommendations.
5. Encourage collaborative research between educators and
researchers to explore the eectiveness of AI in formative
assessment. Co-design studies with teachers and students to
assess how AI impacts learning outcomes and
student engagement.
6. Recognize the evolving role of teachers and facilitators of
AI-enhanced learning.
In the changing times of AI, students more than ever need
teachers to guide them using AI, and as researchers, we would
encourage colleagues to take part in co-designing studies with teachers
and students, where wetogether examine how to improve students
learning through formative assessment practices, critical thinking,
self-regulated learning and AI.
Data availability statement
e original contributions presented in the study are included in
the article/supplementary material, further inquiries can bedirected
to the corresponding author.
Author contributions
TH: Conceptualization, Funding acquisition, Project
administration, Supervision, Writing – original dra. ZZ:
Conceptualization, Writing – original dra. SS: Conceptualization,
Writing – original dra, Writing – review & editing. PR: Writing –
original dra, Writing – review & editing. JM: Conceptualization,
Writing – review & editing.
Funding
e author(s) declare that no nancial support was received for
the research, authorship, and/or publication of this article.
Conflict of interest
e authors declare that the research was conducted in the
absence of any commercial or nancial relationships that could
beconstrued as a potential conict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors
and do not necessarily represent those of their aliated organizations,
or those of the publisher, the editors and the reviewers. Any product
that may be evaluated in this article, or claim that may be made by its
manufacturer, is not guaranteed or endorsed by the publisher.
References
Allal, L. (2010). “Assessment and the regulation of learning” in International
encyclopedia of education. eds. P. Peterson, E. Baker and B. McGaw (Elsevier), 348–352.
Allal, L. (2020). Assessment and the co-regulation of learning in the classroom. Assess.
Educ. 27, 332–349. doi: 10.1080/0969594X.2019.1609411
Andrade, H. L., and Brookhart, S. M. (2020). Classroom assessment as the co-regulation of
learning. Assess. Educ. 27, 350–372. doi: 10.1080/0969594X.2019.1571992
Ball, W., and Holland, S. (2009). e fear of new technology: A naturally
occurring phenomenon. e American Journal of Bioethics, 9, 14–16. doi:
10.1080/15265160802617977
Barthakur, A., Kovanovic, V., Joksimovic, S., Zhang, Z., Richey, M., and Pardo, A.
(2022). Measuring leadership development in workplace learning using automated
assessments: learning analytics and measurement theory approach. Br. J. Educ. Technol.
53, 1842–1863. doi: 10.1111/bjet.13218
Barthakur, A., Dawson, S., and Kovanovic, V. (2023). Advancing leaner proles with
learning analytics: a scoping review of current trends and challenges. In LAK23: 13th
international learning analytics and knowledge conference, 606–612. doi:
10.1145/3576050.3576083
Bearman, M., Ajjawi, R., Boud, D., Tai, J., and Dawson, P. (2023) CRADLE suggest
assessment and gen AI. Centre for Digital Learning, Deakin University, Melbourne, Australia.
Bejar, I. I., Mislevy, R. J., and Zhang, M. (2016). “Automated scoring with validity in
mind” in e Wiley handbook of cognition and assessment: frameworks, methodologies,
and applications. eds. A. A. Rupp and J. P. Leighton (Wiley), 226–246.
Bennett, R. E. (2011). Formative assessment: a critical review. Assess. Educ. 18, 5–25.
doi: 10.1080/0969594X.2010.513678
Black, P., and Wiliam, D. (1998). Assessment and classroom learning. Assess. Educ. 5,
7–74. doi: 10.1080/0969595980050102
Black, P., and Wiliam, D. (2009). Developing the theory of formative assessment. Educ.
Assess. 21, 5–31. doi: 10.1007/s11092-008-9068-5
Black, P., and Wiliam, D. (2018). Classroom assessment and pedagogy. Assess. Educ.
25, 551–575. doi: 10.1080/0969594X.2018.1441807
Boud, D. (2005). Enhancing learning through self-assessment Routledge.
Brandmo, C., Panadero, E., and Hopfenbeck, T. N. (2020). Bridging classroom
assessment and self-regulated learning. Assess. Educ. 27, 319–331. doi:
10.1080/0969594X.2020.1803589
Brookhart, S. M. (2016). “e use of teacher judgement for summative assessment in
the USA” in International teacher judgement practices. ed. V. Klenowski (Oxon, New
York: Routledge), 69–90.
Brooks, C., Carroll, A., Gillies, R. M., and Hattie, J. (2019). A matrix of feedback for
learning. Australian J. Teach. Educ. 44, 14–32. doi: 10.14221/ajte.2018v44n4.2
Brown, G. T. L. (2018). Assessment of student achievement. Oxon, New York: Routledge.
Butler, D. L., and Winne, P. H. (1995). Feedback and self-regulated learning: a
theoretical synthesis. Rev. Educ. Res. 65, 245–281. doi: 10.3102/00346543065003245
Chen, Y., Culpepper, S. A., Wang, S., and Douglas, J. (2018). A hidden Markov model
for learning trajectories in cognitive diagnosis with application to spatial rotation skills.
Appl. Psychol. Meas. 42, 5–23. doi: 10.1177/0146621617721250
Cope, B., Kalantzis, M., and Searsmith, D. (2021). Articial intelligence for education:
knowledge and its assessment in AI-enabled learning ecologies. Educ. Philos. eory 53,
1229–1245. doi: 10.1080/00131857.2020.1728732
Davier, A.A.von, Mislevy, R.J., and Hao, J. eds. (2021). Introduction to computational
psychometrics: towards a principled integration of data science and machine learning
techniques into psychometrics. Computational psychometrics: new methodologies for a
new generation of digital learning and assessment. Boston: Springer.
Davies, M., and Barnett, R. (2015). e Palgrave handbook of critical thinking in higher
education. Springer.
Dewey, J. (1910). How wethink. Boston: Dover Publications.
Dignath, C., Büttner, G., and Langfeldt, H. (2008). How can primary school students learn
self-regulated learning strategies most eectively? A meta-analysis on self-regulation training
programmes. Educ. Res. Rev. 3, 101–129. doi: 10.1016/j.edurev.2008.02.003
Double, K. S., McGrane, J. A., and Hopfenbeck, T. N. (2020). e impact of peer
assessment on academic performance: a meta-analysis of control group studies. Educ.
Psychol. Rev. 32, 481–509. doi: 10.1007/s10648-019-09510-3
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 08 frontiersin.org
Durlak, J. A., and Weissberg, R. P. (2011). Promoting social and emotional
development is an essential part of students’ education. Hum. Dev., 54, 1–3. doi:
10.1159/000324337
Engeness, I. (2021). Developing teachers’ digital identity: towards the pedagogic
design principles of digital environments to enhance students’ learning in the 21st
century. Eur. J. Teach. Educ. 44, 96–114. doi: 10.1080/02619768.2020.1849129
English, N., Robertson, P., Gillis, S., and Graham, L. (2022). Rubrics and formative
assessment in K-12 education: a scoping review of literature. Int. J. Educ. Res. 113:101964.
doi: 10.1016/j.ijer.2022.101964
Ennis, R., Millman, J., and Tomko, T. (2005). Cornell critical thinking tests level X &
Level Z Manual. e Critical inking Co.
Er, E., Dimitriadis, Y., and Gašević, D. (2021). A collaborative learning approach to
dialogic peer feedback: a theoretical framework. Assess. Eval. High. Educ. 46, 586–600.
doi: 10.1080/02602938.2020.1786497
Gamlem, S. M., and Vattøy, K.-D. (2023). “Feedback and classroom practice” in
International encyclopedia of education. eds. R. J. Tierney, F. Rizvi and K. Ercikan, vol.
13. 4th ed (Elsevier), 89–95.
Gardner, J., O'Leary, M., and Yuan, L. (2021). Articial intelligence in educational
assessment: ‘breakthrough? Or buncombe and ballyhoo?’. J. Comput. Assist. Learn. 37,
1207–1216. doi: 10.1111/jcal.12577
Gillani, N., Eynon, R., Chiabaut, C., and Finkel, K. (2023). Unpacking the “Black Box”
of AI in Education. J Educ Techno Soc, 26, 99–111.
Grin, P., and Care, E. (2014). Assessment and teaching of 21st century skills: methods
and approach. Springer Dordrecht: Springer.
Gutierrez, F., and Atkinson, J. (2011). Adaptive feedback selection for intelligent
tutoring systems. Expert Syst. Appl. 38, 6146–6152. doi: 10.1016/j.eswa.2010.11.058
Haenlein, M., and Kaplan, A. (2019). A brief history of articial intelligence: on the
past, present, and future of articial intelligence. Calif. Manag. Rev. 61, 5–14. doi:
10.1177/0008125619864925
Halai, A., Sarungi, V., and Hopfenbeck, T. N. (2022). Teachers’ perspectives and
practice of assessment for learning in classrooms in Tanzania. Int. Encycl. Educ. 63-72.
doi: 10.1016/B978-0-12-818630-5.09039-4
Hattie, J. A. (2009). Visible learning: a synthesis of over 800 meta-analyses relating to
achievement. London and New York: Routledge.
Hill, P. W., and Barber, M. (2014). Preparing for a renaissance in assessment Pearson.
Hopfenbeck, T. N., and Stobart, G. (2015). Large-scale implementation of assessment
for learning. Assess. Educ. 22, 1–2. doi: 10.1080/0969594X.2014.1001566
Hopfenbeck, T. N., Flórez Petour, M. T., and Tolo, A. (2015). Balancing tensions in
educational policy reforms: large-scale implementation of assessment for learning in
Norway. Assess. Educ. 22, 44–60. doi: 10.1080/0969594X.2014.996524
Humphry, S. M., and Heldsinger, S. A. (2014). Common structural design features of
rubrics may represent a threat to validity. Educ. Res. 43, 253–263. doi:
10.3102/0013189X14542154
Johnston, S. -K. (2023). Privacy considerations of using social robots in education: policy
recommendations for learning environments. United Nations, Department of Economics
and Social Aairs, Sustainable Development.
Jönsson, A., and Panadero, E. (2016). “e use and Design of Rubrics to support assessment
for learning” in Scaling up assessment for learning in higher education. eds. D. Carless, S. M.
Bridges, C. K. Y. Chan and R. Golfcheski (York: Springer), 99–111. (https://www.pearson.com/
content/dam/one-dot-com/one-dot-com/global/Files/about-pearson/innovation/open-ideas/
PreparingforaRenaissanceinAssessment.pdf)
Ke, Z., and Ng, V. (2019). Automated Essay Scoring: A Survey of the State of the Art.
Paper presented at the 28th International Joint Conference on Articial Intelligence.
King, P. M., and Kitchener, K. S. (1994). Developing reective judgment: understanding
and promoting intellectual growth and critical thinking in adolescents and adults.
SanFrancisco: Jossey-Bass.
Kuhn, D., and Weinstock, M. (2004). “What is epistemological thinking and
why does it matter?” in Personal epistemology: the psychology of beliefs about
knowledge and knowing. eds. B. K. Hofer and P. R. Pintrich (New York: Routledge),
121–144.
Lee, A. V. Y. (2021). Determining quality and distribution of ideas in online
classroom talk using learning analytics and machine learning. Educ. Technol. Soc.
24, 236–249.
Lee, A. V. Y. (2023). Supporting students’ generation of feedback in large-scale online
course with articial intelligence-enabled evaluation. Stud. Educ. Eval. 77:101250. doi:
10.1016/j.stueduc.2023.101250
Li, C., Xing, W., and Leite, W. (2022). Using fair AI to predict students’ math learning
outcomes in an online platform. Interact. Learn. Environ. 1-20, 1–20. doi:
10.1080/10494820.2022.2115076
Li, T., Reigh, E., He, P., and Adah Miller, E. (2023). Can weand should weuse articial
intelligence for formative assessment in science? J. Res. Sci. Teach. 60, 1385–1389. doi:
10.1002/tea.21867
van der Linden, W. J., and Glas, C. A. (2010). Elements of adaptive testing. New York:
Springer.
Linn, M. C., Gerard, L., Ryoo, K., McElhaney, K., Liu, O. L., and Raerty, A. N. (2014).
Computer-guided inquiry to improve science learning. Science 344, 155–156. doi:
10.1126/science.1245980
Lipnevich, A. A., and Smith, J. K. (2018). e Cambridge handbook on instructional
feedback Cambridge University Press.
Liu, N. F., and Carless, D. (2006). Peer feedback: the learning element of peer
assessment. Teach. High. Educ. 11, 279–290. doi: 10.1080/13562510600680582
Liu, O. L., Rios, J. A., Heilman, M., Gerard, L., and Linn, M. C. (2016). Validation of
automated scoring of science assessments. J. Res. Sci. Teach. 53, 215–233. doi: 10.1002/
tea.21299
Luaces, O., Díez, J., and Bahamonde, A. (2018). A peer assessment method to prov ide
feedback, consistent grading and reduce students’ burden in massive teaching settings.
Comput. Educ. 126, 283–293. doi: 10.1016/j.compedu.2018.07.016
Luckin, R., and Holmes, W. (2016). Intelligence unleashed: an argument for AI in
education. Available at: https://discovery.ucl.ac.uk/id/eprint/1475756
Ma, W., Adesope, O. O., Nesbit, J. C., and Liu, Q. (2014). Intelligent tutoring systems
and learning outcomes: a meta-analysis. J. Educ. Psychol. 106, 901–918. doi: 10.1037/
A0037123
McMillan, J. H. (2013). SAGE handbook of research on classroom assessment.
LosAngeles: Sage.
Milano, S., McGrane, J. A., and Leonelli, S. (2023). Large language models challenge
the future of higher education. Nat. Mach. Intell. 5, 333–334. doi: 10.1038/
s42256-023-00644-2
Milligan, S. K., and Grin, P. (2016). Understanding learning and learning design in
MOOCs: a measurement-based interpretation. J. Learn. Analyt. 3, 88–115. doi:
10.18608/jla.2016.32.5
Mousavinasab, E., Zarifsanaiey, N., Niakan Kalhori, S. R., Rakhshan, M., Keikha, L.,
and Ghazi Saeedi, M. (2021). Intelligent tutoring systems: a systematic review of
characteristics, applications, and evaluation methods. Interact. Learn. Environ. 29,
142–163. doi: 10.1080/10494820.2018.1558257
Murphy, R . F. (2019). Articial intelligence applications to support K-12 teachers and
teaching. Rand Corp. 10, 1–20. doi: 10.7249/PE315
Nicol, D., and MacFarlane-Dick, D. (2006). Formative assessment and self-regulated
learning: a model and seven principles of good feedback practice. Stud. High. Educ. 31,
199–218. doi: 10.1080/03075070600572090
Panadero, E., and Jönsson, A. (2013). e use of scoring rubrics for formative
assessment purposes revisited: a review. Educ. Res. Rev. 9, 129–144. doi: 10.1016/j.
edurev.2013.01.002
Panadero, E., Andrade, H., and Brookhart, S. (2018). Fusing self-regulated learning and
formative assessment: a roadmap of where weare, how wegot here, and where weare going.
Aust. Educ. Res. 45, 13–31. doi: 10.1007/s13384-018-0258-y
Perry, N., Lisaingo, S., Ye e, N., Parent, N., Wan, X., and Muis, K. (2020). Collaborating
with teachres to design and implement assessments for self-regulated learning in the
context of authentic classroom writing tasks. Assess. Educ. 27, 416–443. doi:
10.1080/0969594X.2020.1801576
Popham, W. J. (1997). What’s wrong- and what’s right- with rubrics. Educ. Leadersh. 55,
72–75.
Ratnam-Lim, C. T. L., and Tan, K. H. K. (2015). Large-scale implementation of
formative assessment practices in an examination-oriented culture. Assess. Educ. 22,
61–78. doi: 10.1080/0969594X.2014.1001319
Ruiz-Primo, M., and Brookhart, S. (2018). Using feedback to improve l earning. Routledge.
Scalise, K., Wilson, M., and Gochyyev, P. (2021). A taxonomy of critical dimensions
at the intersection of learning analytics and educational measurement. Front. Educ.
6:656525. doi: 10.3389/feduc.2021.656525
Schwartz, S. (2020). Flood of online learning resources overwhelms teachers. Educ.
Wee k. Available at: March 25, 2020: https://www.edweek.org/teaching-learning/ood-
of-online-learning-resources-overwhelms-teachers/2020/03
Shin, J., Chen, F., Lu, C., and Bulut, O. (2022). Analyzing students’ performance in
computerized formative assessments to optimize teachers’ test administration decisions
using deep learning frameworks. Journal of Computers in Education 9, 71–91. doi:
10.1007/s40692-021-00196-7
Skeat, J., and Ziebell, N. (2023). University students are using AI, but not how
youthink. Available at: https://pursuit.unimelb.edu.au/articles/university-students-are-
using-ai-but-not-how-you-think
Stobart, G., and Hopfenbeck, T. (2014). “Assessment for learning and formative
assessment” in State of the eld review assessment and learning. eds. J.-A. Baird, T.
Hopfenbeck, P. Newton, G. Stobart and A. Steen-Utheim (Oxford: Norwegian
Knowledge Centre for Education).
Sun, S. Z. (2021). Epistemological beliefs: the key to enhance critical thinking for
higher education students in the east. Paper presented at the 2021 American Educational
Research Association (AERA) annual meeting.
Sun, Z. S. (2023). Developing and validating an operationalisable model for critical
thinking assessment in dierent cultures. e University of Melbourne.
Swiecki, Z., Khosravi, H., Chen, G., Martinez-Maldonado, R., Lodge, J. M.,
Milligan, S., et al. (2022). Assessment in the age of articial intelligence. Comp. Educ.
3:100075. doi: 10.1016/j.caeai.2022.100075
Hopfenbeck et al. 10.3389/feduc.2023.1270700
Frontiers in Education 09 frontiersin.org
Tashu, T. M., and Horvath, T. (2019). Semantic-based feedback recommendation for
automatic essay evaluation. Proceedings of SAI Intelligent Systems Conference 334–346.
London: Springer.
Tomasik, M. J., Berger, S., and Moser, U. (2018). On the development of a
computer-based tool for formative student assessment: epistemological,
methodological, and practical issues. Front. Psychol. 9:2245. doi: 10.3389/
fpsyg.2018.02245
Verma, M. (2018). Articial intelligence and its scope in dierent areas with special
reference to the eld of education. Int. J. Adv. Educ. Res. 3, 5–10.
Vincent-Lamarre, P., and Larivière, V. (2021). Textual analysis of articial intelligence
manuscripts reveals features associated with peer review outcome. Quant. Sci. Stud. 2,
662–677. doi: 10.1162/qss_a_00125
von Davier, A. A., Mislevy, R. J., and Hao, J. (2022). Computational psychometrics: New
methodologies for a new generation of digital learning and assessment: With examples in
R and Python. Springer.
Webb, M., Gibson, D., and Forkosh-Baruch, A. (2013). Challenges for information
technology supporting educational assessment. J. Comput. Assist. Learn, 29, 451–462.
doi: 10.1111/jcal.12033
Wen, H., Liu, Y., and Zhao, N. (2020). Longitudinal cognitive diagnostic
assessment based on the HMM/ANN model. Front. Psychol. 11:2145. doi: 10.3389/
fpsyg.2020.02145
Wertz, M. H. (2019). Epistemological developmental level and critical thinking skill level in
undergraduate university students. University of South Florida.
Wiliam, D. (2011). What is assessment for learning? Stud. Educ. Eval. 37, 3–14. doi:
10.1016/j.stueduc.2011.03.001
Wiliam, D., Lee, C., Harrison, C., and Black, P. (2004). Teachers developing assessment
for learning: impact on student achievement. Assess. Educ. 11, 49–65. doi:
10.1080/0969594042000208994
Wylie, C. E., and Lyon, C. J. (2015). The fidelity of formative assessment
implementation: issues of breadth and quality. Assess. Educ. 22, 140–160. doi:
10.1080/0969594X.2014.990416
Wollenschläger, M., Hattie, J., Machts, N., Möller, J., and Harms, U. (2016). Whatmakes
rubrics eective in teacher-feedback? Transparency of learning goalsis not enough. Contemp.
Educ. Psychol. 44-45, 1–11. doi: 10.1016/j.cedpsych.2015.11.003
Zhai, X., and Nehm, R. H. (2023). AI and formative assessment: the train has le the
station. J. Res. Sci. Teach. 60, 1390–1398. doi: 10.1002/tea.21885
... In the fast-changing environment of 21st-century education [1], the integration of digital technology into classroom assessment in secondary schools (typically encompassing grades 9-12 or ages [14][15][16][17][18] has become a game-changer, significantly altering the frameworks of teaching and learning [2], [3], [4], [5], [6]. ...
... The integration of digital technology in classroom assessment has garnered significant attention from researchers exploring its challenges, opportunities, and implications for teaching practices and student learning outcomes. In this regard, [18] provided a perspective article discussing the possible effects of artificial intelligence (AI) on formative assessment methods, emphasizing the advantages and difficulties AI introduces in this area. ...
Article
Full-text available
This study examines how digital technology affects classroom evaluation using bibliometrics. The research addresses a gap in the systematic analysis of academic publications on this topic, focusing on the evolution and impact of digital technologies on classroom assessment practices. The study uses data from leading databases such as Scopus from 2000 to 2023, identifying influential works, authors, and emerging trends. The findings show a significant increase in research activity in this area, with a 150% increase in related publications since 2010. In particular, the development and use of innovative digital assessment tools, such as game-based assessment and learning analytics, have emerged as key trends. These technologies are reshaping the way educators assess student performance, offering new possibilities for feedback and learning analysis. However, challenges related to equity, access, and teacher training persist. This study highlights the growing importance of digital technology in classroom assessment, signalling a shift towards more interactive, personalized, and data-driven assessment methods. It also identifies critical areas for future research, including equitable technology implementation and comprehensive training programs for educators. Based on the theoretical framework of Assessment for Learning, this study shows how digital technology can improve formative assessment practices. Future research should investigate the sustained effectiveness of digital assessment tools and explore strategies for bridging the digital divide in classroom assessment.
... As such, corrective feedback is thought to be beneficial for second language acquisition as it allows students to pick up grammatical features that may be lost due to the discontinued access to learning standards (Ellis, 2009). Research into the most effective form of WCF is important for teachers who spend an inordinate amount of time on providing such feedback and for learners who wish to improve their grammatical accuracy (Hopfenbeck et al., 2023;Storch, 2018). ...
... However, there are also some indications that automated feedback may position learners to be reli-ant on the provided feedback, and thus decreasing self-initiated corrections (Asiri & Khadawardi, 2024;Baron, 2023). Hopfenbeck et al. (2023) discuss both the potential benefits and challenges of using AI to support student learning. AI can significantly enhance formative assessment by providing personalised feedback, automating assessment tasks, and supporting large-scale, sustainable assessment practices. ...
Article
Full-text available
This study draws on the cultural-historical perspectives of Vygotsky and Galperin to examine the role of AI-generated feedback within the Assessment for Learning (AfL) process in fostering students’ development as learners. By leveraging Galperin’s concept of cultural tools and the developmental role of human activity, elaborated in his dissertation written almost a hundred years ago, this study elucidates how this theoretical framework can enhance our understanding of the pedagogical value of individually tailored feedback from AI, ultimately contributing to human development and inspire the Design Principles (DPs) of AI-based educational technologies. Essay Assessment Technology (EAT), designed according to the suggested DPs, is presented to illustrate the application of AfL strategies in schools, highlighting its potential to enhance students’ learning and their development as learners.
... These tools allow for greater flexibility but may lack the reliability and consistency of externally validated rubrics, primarily when used to assess subjective skills like communication or critical thinking. The one closest to this study is defined by Hopfenbeck et al. (2023), in which educators create rubrics to meet the unique demands of their instructional setting. They are seen as valuable for offering tailored feedback, though they may present challenges in consistency and scalability. ...
Article
Full-text available
Assessing English communicative abilities within an online environment poses challenges, particularly when aligning assessment instrument rubrics to enhance the evaluation process and learning outcomes. Home-grown rubrics (HGR) can address specific situations where standardised rubrics might only partially meet an assessment's goals or unique requirements. However, several challenges associated with HGR have been identified when assessing complex skills such as communication. These include inconsistencies in rubric design, limitations in capturing subjective aspects of performance, and difficulties in providing coherent, actionable feedback. This study investigates the relevance of HGR assessment rubrics for English communication skills to cater to online contexts, primarily during the pandemic. An explanatory sequential mixed-methods design focused on an Islamic institution’s English for Proficiency (Pre-Intermediate) course was used to gather data from students for quantitative analysis and from instructors for qualitative inquiries after using the HGR for a semester. A survey instrument with scales of agreement was adapted, piloted, and distributed to 87 students. An unstructured interview was chosen to uncover in-depth insights from four lecturers actively involved with the HGR. Descriptive and thematic analyses were employed to determine users’ perceived agreement and experiences in developing and utilising the HGR. Findings indicate that the students accepted the HGR as a tool for self-assessing their English communicative abilities in a virtual environment. The HGR has also enhanced both lecturers’ and students’ awareness of the assessment process, improved the efficiency of online assessments, and helped students focus on critical components of their assessments. The outcome substantiates the relevance of the HGR for the studied context, suggesting its broader applicability and tailor-made design for other online English language courses and skills.
... Any educational system's learning objectives are globally correlated with the effectiveness of the procedures used for student performance evaluation and assessment. The purpose of classroom assessments, which include formative evaluation, is to enhance students' learning in the classroom (Hopfenbeck et al., 2023). In order to monitor students' learning outcomes during the formation stage of learning, teachers provide feedback to the class, which is known as formative assessment. ...
Article
Full-text available
The relationship between learning and assessment has been re-evaluated over time, prompted by the increasing need for critical and reflective thinkers who are committed to lifelong learning. The purpose of this study is to investigate, using data from 10 lecturers and 18 postgraduate students at Nnamdi Azikiwe University in Awka, Nigeria, the impact of peer assessment (PA) and self-assessment (SA) procedures and their consequences on learning in higher education. Through the use of in-depth interviews and the distribution of questionnaires (Self-Assessment and Peer-Assessment Questionnaire-SAPAQ), a mixed method research strategy was used. In order to provide the students a practical understanding of the strategies, SA and PA procedures were conducted for them. The students were then given a questionnaire to fill out regarding how they felt about SA and PA. To learn more about the lecturers' opinions on the two approaches, interview sessions were also conducted with them. A thematic method was used to analyze the themes and sub-themes that emerged from the interview questions and responses, and descriptive statistics were used to analyze the quantitative data. A number of intriguing discoveries were found, including that: both PG students and their lecturers agreed that SA and PA were effective learning strategies that would increase students' motivation to learn if and when they were designed for formative assessment. Therefore, there will be more positive outcomes from SA and PA with theory-based (qualitative) courses than numerical-based courses because qualitative aspects of the courses would be easier for students to assess. Additionally, according to both sets of participants, SA and PA aid in the development of students' critical thinking abilities. They also recommend that the results of SA and PA be included in the final (summative) assessment so that students will take both strategies more seriously.
... The goal of many recently developed AI-powered systems was to support teachers with data-driven decisions to improve their practice, decrease their workload, and organize their classrooms more effectively. If implemented effectively, GenAI technologies could have the potential to streamline the grading process, and thus reduce educators' workload and provide more accurate and consistent formative assessment and feedback provision (Hopfenbeck et al., 2023) by providing instructional suggestions to teachers for them to adopt or ignore (Luckin et al., 2022). At the same time, proper integration of GenAI tools requires students and instructors to develop new skills, such as prompting (Misiejuk et al., 2024b) or GenAI literacy (Bozkurt, 2024). ...
Preprint
Full-text available
[Accepted for publications in Journal of Learning Analytics] Generative artificial intelligence (GenAI) has opened new possibilities for designing learning analytics (LA) tools, gaining new insights about student learning processes and their environment, and supporting teachers in assessing and monitoring students. This systematic literature review maps the empirical research of 41 papers utilizing GenAI and LA and interprets the results through the lens of the LA/EDM process cycle. Currently, GenAI is mostly implemented to automate discourse coding, scoring or classification tasks. Few papers used GenAI to generate data or to summarize text. Classroom integrations of GenAI and LA mostly explore facilitating human-GenAI collaboration, rather than implementing automated feedback generation or GenAI-powered learning analytics dashboards. The majority of papers use Generative Adversarial Network models to generate synthetic data, BERT models for classification or prediction tasks, BERT or GPT models for discourse coding, and GPT models for tool integration. Although most studies evaluate the GenAI output, we found examples of using GenAI without the output validation, especially, when its output is feeding into a LA pipeline aiming to, for example, develop a dashboard. This review offers a comprehensive overview of the field to aid LA researchers in the design of research studies and a contribution to establishing best practices to integrate GenAI and LA.
Chapter
The integration of Artificial Intelligence (AI), Virtual Reality (VR), and Augmented Reality (AR) in education is reshaping the learning landscape by providing immersive and personalized learning experiences. Immersive learning through VR and AR allows for practical simulations and experiential learning, bridging the gap between theoretical knowledge and real-world application. However, challenges such as ethical considerations, data privacy, and the digital divide need to be addressed to ensure inclusive and equitable access to these advanced learning tools. The chapter explores the benefits of AI-driven immersive learning, including improved student motivation, engagement, and retention, while also discussing scalability and the potential for widespread adoption across diverse educational settings. It emphasizes the importance of incorporating ethical frameworks and continuous research to align AI, VR, and AR applications with the evolving needs of modern education, ensuring that these technologies contribute positively to student outcomes and lifelong learning.
Article
Full-text available
Generative Artificial Intelligence (GenAI) is gaining momentum in schools as a means of support to the teaching and learning process. However, its use poses several controversial questions, especially in lower school grades, and teachers might often face ethical or intellectual obstacles preventing them from using AI in their classes. This study explores the perceptions of a sample of 1,223 teachers across subjects of instruction from 572 schools in the regional context (nursery, primary, lower and upper secondary), using a mixed-method approach. Results suggest that there is a widespread confusion on the possible applications of GenAI in education, possibly leading to reduced teachers' intention to integrate these tools in their practices. Results also point towards a general need for more CPD on the topic. Age, level of school and subject of instruction were found to moderate the effect of teachers' perceived readiness to use GenAI. Regarding possible negative implementations of GenAI, teachers showed to have mixed opinions, from open contrast to unreserved enthusiasm. Limitations of the study and future research lines are also addressed.
Preprint
Full-text available
In response to Li, Reigh, He, and Miller’s commentary on Can we and should we use artificial intelligence for formative assessment in science, we argue that AI is already being widely employed in formative assessment across various educational contexts. While agreeing with Li et al.'s call for further studies on equity issues related to AI, we emphasize the need for science educators to adapt to the AI revolution that has outpaced the research community. We challenge the somewhat restrictive view of formative assessment presented by Li et al., highlighting the significant contributions of AI in providing formative feedback to students, assisting teachers in assessment practices, and aiding in instructional decisions. We contend that AI-generated scores should not be equated with the entirety of formative assessment practice; no single assessment tool can capture all aspects of student thinking and backgrounds. We address concerns raised by Li et al. regarding AI bias and emphasize the importance of empirical testing and evidence-based arguments in referring to bias. We assert that AI-based formative assessment does not necessarily lead to inequity and can, in fact, contribute to more equitable educational experiences. Furthermore, we discuss how AI can facilitate the diversification of representational modalities in assessment practices and highlight the potential benefits of AI in saving teachers’ time and providing them with valuable assessment information. We call for a shift in perspective, from viewing AI as a problem to be solved to recognizing its potential as a collaborative tool in education. We emphasize the need for future research to focus on the effective integration of AI in classrooms, teacher education, and the development of AI systems that can adapt to diverse teaching and learning contexts. We conclude by underlining the importance of addressing AI bias, understanding its implications, and developing guidelines for best practices in AI-based formative assessment.
Conference Paper
Full-text available
The term Learner Profile has proliferated over the years, and more recently, with the increased advocacy around personalising learning experiences. Learner profiles are at the center of personalised learning, and the characterisation of diversity in classrooms is made possible by profiling learners based on their strengths and weaknesses, backgrounds and other factors influencing learning. In this paper, we discuss three common approaches of profiling learners based on students' cognitive knowledge, skills and competencies and behavioral patterns, all latter commonly used within Learning Analytics (LA). Although each approach has its strengths and merits, there are also several disadvantages that have impeded adoption at scale. We propose that the broader adoption of learner profiles can benefit from careful combination of the methods and practices of three primary approaches, allowing for scalable implementation of learner profiles across educational systems. In this regard, LA can leverage from other aligned domains to develop valid and rigorous measures of students' learning and propel learner profiles from education research to more mainstream educational practice. LA could provide the scope for monitoring and reporting beyond an individualised context and allow holistic evaluations of progress. There is promise in LA research to leverage the growing momentum surrounding learner profiles and make a substantial impact on the field's core aim-understanding and optimising learning as it occurs.
Article
Full-text available
In this commentary, we respond to the recent article Applying machine learning to automatically assess scientific models by Zhai et al. (2022). The authors present automated assessment as a solution to the problem of limited time for assessment in middle school science classrooms. Drawing from our collective expertise in science assessment, machine learning (ML), artificial intelligence (AI), and culturally relevant and linguistically responsive pedagogy, we argue that there are significant limitations to the current applications of AI for formative assessment practices. Although we believe that these limitations extend to all students, we are particularly concerned about the implications for students from nondominant cultural and linguistic backgrounds. We first share our understanding of AI's role in formative assessment, with reference to the paper by Zhai and colleagues. Next, we ask whether AI can effectively assess students' emergent sensemaking and then consider whether we should use AI for purposes of formative assessment. Finally, we discuss how we can better use AI for formative assessment.
Article
Full-text available
In this paper, we argue that a particular set of issues mars traditional assessment practices. They may be difficult for educators to design and implement; only provide discrete snapshots of performance rather than nuanced views of learning; be unadapted to the particular knowledge, skills, and backgrounds of participants; be tailored to the culture of schooling rather than the cultures schooling is designed to prepare students to enter; and assess skills that humans routinely use computers to perform. We review extant artificial intelligence approaches that–at least partially–address these issues and critically discuss whether these approaches present additional challenges for assessment practice.
Article
Educators in large-scale online courses tend to lack the necessary resources to generate and provide adequate feedback for all students, especially when students’ learning outcomes are evaluated through student writing. As a result, students welcome peer feedback and sometimes generate self-feedback to widen their perspectives and obtain feedback, but often lack the support to do so. This study, as part of a larger project, sought to address this prevalent problem in large-scale courses by allowing students to write essays as an expression of their opinions and response to others, conduct peer and self-evaluation, using provided rubric and Artificial Intelligence (AI)-enabled evaluation to aid the giving and receiving of feedback. A total of 605 undergraduate students were part of a large-scale online course and contributed over 2500 short essays during a semester. The research design uses a mixed-methods approach, consisting qualitative measures used during essay coding, and quantitative methods from the application of machine learning algorithms. With limited instructors and resources, students first use instructor-developed rubric to conduct peer and self-assessment, while instructors qualitatively code a subset of essays that are used as inputs for training a machine learning model, which is subsequently used to provide automated scores and an accuracy rate for the remaining essays. With AI-enabled evaluation, the provision of feedback can become a sustainable process with students receiving and using meaningful feedback for their work, entailing shared responsibility from teachers and students, and becoming more effective.
Chapter
Assessment for Learning in Africa was a 3-year (2016–2019) study in Tanzania and South Africa. It studied how to develop sustainable teacher capacity for use of assessment for learning (AfL) in challenging educational settings. In Tanzania the study included a teacher development program in selected schools in an informal settlement in Dar es Salaam. This paper reports from the qualitative data from the teacher development. Teachers considered covering the syllabus for examinations as the main purpose of teaching. Teachers' practice showed creative approaches to manage the large classes, provide students an opportunity to participate in the classroom. While, teachers' espoused a nuanced understanding of AfL, their practice was largely teacher-directed with little evidence of students taking ownership of learning. The study makes significant recommendations for policy and practice in education.
Chapter
Interest in how feedback is used to enhance learning processes has increased during the last several decades. The aim of this paper is to discuss the relevance of a dialogic feedback practice and propose a framework for effective classroom feedback as responsive pedagogy. A model based on the framework is introduced to visualize the relationship between its six dimensions: goals, feedback strategy, feedback content, feedback dialog, self-efficacy, and emotions. Feedback agency is central for understanding how the dimensions interact. Implications are a need for classroom feedback dialogs in which the motivational and emotional aspects of students' self-regulated learning are actively promoted.
Article
As instruction shifts away from traditional approaches, online learning has grown in popularity in K-12 and higher education. Artificial intelligence (AI) and learning analytics methods such as machine learning have been used by educational scholars to support online learners on a large scale. However, the fairness of AI prediction in educational contexts has received insufficient attention, which can increase educational inequality. This study aims to fill this gap by proposing a fair logistic regression (Fair-LR) algorithm. Specifically, we developed Fair-LR and compared it with fairness-unaware AI models (Logistic Regression, Support Vector Machine, and Random Forest). We evaluated fairness with equalized odds that caters to statistical type I and II errors in predictions across demographic subgroups. The results showed that the Fair-LR could generate desirable predictive accuracy while achieving better fairness. The findings implied that the educational community could adopt a methodological shift to achieve accurate and fair AI to support learning and reduce bias.