PreprintPDF Available

Embrace Opportunities and Face Challenges: Using ChatGPT in Undergraduate Students' Collaborative Interdisciplinary Learning

Authors:
Preprints and early-stage research may not have been peer reviewed yet.

Abstract and Figures

ChatGPT, launched in November 2022, has gained widespread attention from students and educators globally, with an online report by Hu (2023) stating it as the fastest-growing consumer application in history. While discussions on the use of ChatGPT in higher education are abundant, empirical studies on its impact on collaborative interdisciplinary learning are rare. To investigate its potential, we conducted a quasi-experimental study with 130 undergraduate students (STEM and non-STEM) learning digital literacy with or without ChatGPT over two weeks. Weekly surveys were conducted on collaborative interdisciplinary problem-solving, physical and cognitive engagement, and individual reflections on ChatGPT use. Analysis of survey responses showed significant main effects of topics on collaborative interdisciplinary problem-solving and physical and cognitive engagement, a marginal interaction effect between disciplinary backgrounds and ChatGPT conditions for cognitive engagement, and a significant interaction effect for physical engagement. Sentiment analysis of student reflections suggested no significant difference between STEM and non-STEM students' opinions towards ChatGPT. Qualitative analysis of reflections generated eight positive themes, including efficiency, addressing knowledge gaps, and generating human-like responses, and eight negative themes, including generic responses, lack of innovation, and counterproductive to self-discipline and thinking. Our findings suggest that ChatGPT use needs to be optimized by considering the topics being taught and the disciplinary backgrounds of students rather than applying it uniformly. These findings have implications for both pedagogical research and practices.
Content may be subject to copyright.
1
Embrace Opportunities and Face Challenges: Using ChatGPT in Undergraduate Students'
Collaborative Interdisciplinary Learning
Gaoxia Zhu*
Assistant Professor
National Institute of Education (NIE)
Nanyang Technological University
E-mail: gaoxia.zhu@nie.edu.sg
Xiuyi Fan
Senior Lecturer
School of Computer Science and Engineering
Nanyang Technological University
E-mail: xyfan@ntu.edu.sg
Chenyu Hou
Ph.D. Student
Nanyang Technological University
E-mail: CHENYU004@e.ntu.edu.sg
Tianlong Zhong
Ph.D. Student
Nanyang Technological University
E-mail: tianlong001@e.ntu.edu.sg
Peter Seow
Research Scientist
National Institute of Education (NIE)
Nanyang Technological University
E-mail: peter.seow@nie.edu.sg
Annabel Chen Shen-Hsing
Professor
Division of Psychology
Nanyang Technological University
E-mail: annabelchen@ntu.edu.sg
Preman Rajalingam
Director
Centre for Teaching, Learning & Pedagogy
2
Nanyang Technological University
E-mail: prajalingam@ntu.edu.sg
Low Kin Yew
Associate Professor (Practice)
College of Business (Nanyang Business School)
Nanyang Technological University
E-mail: alowky@ntu.edu.sg
Tan Lay Poh
Associate Professor
School of Materials Science & Engineering
Nanyang Technological University
E-mail: lptan@ntu.edu.sg
Gaoxia Zhu and Xiuyi Fan contributed equally to this article.
* Corresponding author
Learning Sciences and Assessment Academic Group, National Institute of Education (NIE),
Nanyang Technological University.
E-mail: gaoxia.zhu@nie.edu.sg
Declaration
Competing interests: The authors declare no potential conflict of interest in the work.
Availability of data and materials: Because of confidentiality agreements and ethical concerns,
the data used in this study will not be made public. These data will be made available to other
researchers on a case-by-case basis.
Funding: This study was supported by the NTU Edex Teaching and Learning Grants.
Acknowledgments: The authors are indebted to the students who participated in this study.
3
Abstract
ChatGPT, launched in November 2022, has gained widespread attention from students and
educators globally, with an online report by Hu (2023) stating it as the fastest-growing consumer
application in history. While discussions on the use of ChatGPT in higher education are abundant,
empirical studies on its impact on collaborative interdisciplinary learning are rare. To investigate
its potential, we conducted a quasi-experimental study with 130 undergraduate students (STEM
and non-STEM) learning digital literacy with or without ChatGPT over two weeks. Weekly
surveys were conducted on collaborative interdisciplinary problem-solving, and physical and
cognitive engagement, and individuals reflected on their ChatGPT use. Analysis of survey
responses showed significant main effects of topics on collaborative interdisciplinary problem-
solving and physical and cognitive engagement, a marginal interaction effect between disciplinary
backgrounds and ChatGPT conditions for cognitive engagement, and a significant interaction
effect for physical engagement. Qualitative analysis of students' reflections generated eight
positive themes about ChatGPT use, including efficiency, addressing knowledge gaps, and
generating human-like responses, and eight negative themes, including generic responses, lack of
innovation, and counterproductive to self-discipline and thinking. Our findings suggest that
ChatGPT use needs to be optimized by considering the topics being taught and the disciplinary
backgrounds of students rather than applying it uniformly. These findings have implications for
both pedagogical research and practices.
Keywords: Undergraduates, higher education, collaborative interdisciplinary learning, ChatGPT
Introduction
Released in November 2022, ChatGPT (Generative Pretrained Transformer, based on GPT-
4
3.5), a large language model pre-trained on massive text data using Reinforcement Learning from
Human Feedback technique (Thorp, 2023), has received tremendous attention, interest, surprise,
applause, and at the same time, concerns all around the world. In higher education, ChatGPT is
shaking up the landscape. On the one hand, it brings new opportunities, such as triggering
instructors to use ChatGPT to design innovative assessments and activities for teaching and
learning purposes. On the other hand, the use of ChatGPT poses challenges, such as potential costs
and efforts for educators to evaluate the relevance and accuracy of generated information and
threats to student essays as evaluation assignments because ChatGPT-generated text takes only
seconds to produce and may not trigger plagiarism detectors (Rudolph et al., 2023).
However, with a few exceptions (e.g., Ali, 2023; Shoufan, 2023; Yan, 2023), which we will
discuss below, there is a lack of empirical research on the impact of ChatGPT in learning. Moreover,
there is no ChatGPT study on collaborative interdisciplinary learning. Collaborative
interdisciplinary learning involves using knowledge, methods, and insights from various
disciplines to solve complex problems in challenging situations (Bybee, 2013; Ivanitskaya et al.,
2002). In the first year of their university programs, undergraduate students may not have sufficient
knowledge of their own or other disciplines, which hinders their ability to engage in
interdisciplinary discussions and learning with other students. ChatGPT can be a valuable tool to
complement students' lack of disciplinary knowledge because it is trained on a vast amount of data
from different disciplines and has conversational capabilities and the ability to take on a specified
persona or identity (Qadir, 2022).
In this work, we conducted an empirical research on the impact of ChatGPT with 130
undergraduate students enrolled in a digital literacy course in an Asian public university. The study
lasted for two weeks, and the students worked on two learning topics, namely, Artificial
5
Intelligence (AI) and Blockchain. To examine the impact of ChatGPT on students' collaborative
interdisciplinary learning and their perceptions, we collected student responses on class
engagement and problem-solving skills using weekly surveys and online written self-reflections
after each class. We analyzed the student data to identify positive and negative themes relating to
their learning using ChatGPT. The benefits of this study are two-folded: (1) providing valuable
insights into the potential benefits and challenges of using ChatGPT in collaborative
interdisciplinary learning and (2) informing the development of effective teaching guidelines for
students from different disciplines.
Literature Review
ChatGPT in Education
Chatbots are software programs that engage in real-time conversations with users (Clarizia
et al., 2018). In education, they serve roles such as teaching agents, peer agents, teachable agents,
and motivational agents (Kuhail et al., 2022). Chatbots can integrate multiple sources of
information, answer students' questions immediately and motivate them (Okonkwo & Ade-Ibijola,
2021). They have been used in computing education, language learning, and mathematics
education (Winkler et al., 2020; Jeon, 2021; Rodrigo et al., 2012). However, criticism such as
lacking user-centered design, fictional conversations, and ethical issues are also there (Kuhail et
al., 2022; Murtarelli et al., 2021).
ChatGPT is a powerful chatbot with human-like text processing abilities. Some are
apprehensive about its use, while others are optimistic. In education, Kasneci et al. (2023) listed
risks and challenges with using LLMs, including ethical and practical concerns. Tlili et al. (2023)
raised concerns about ChatGPT's accuracy, fairness, and user privacy, while Qadir (2022) noted
potential biases and ethical issues. Cheating and plagiarism are also concerns (Guo et al., 2023).
6
Rudolph et al. (2023) suggested that ChatGPT's limitations include understanding context,
emotion, creativity, misinformation, response quality variation, and the danger of jailbreaking.
Despite these concerns, there is a growing consensus that LLMs will become increasingly
prevalent in daily life and learning (Looi & Wong, 2023; McMurtrie, 2022; Thorp, 2023; Tlili et
al., 2023). Previous studies suggest that LLMs can have positive effects on providing automated
assessments and adaptive feedback (Moore et al., 2022; Sailer et al., 2023; Zhu et al., 2020),
stimulating curiosity (Abdelghani et al., 2022), increasing engagement (Tai & Chen, 2020), and
supporting programming tasks and code explanations in computing education (Sarsa et al., 2022).
ChatGPT has shown potential effectiveness in supporting language learning (Ali et al., 2023; Yan,
2023), medical education (Kung et al., 2023; Sallam, 2023), and engineering education (Qadir,
2022; Shoufan, 2023).
As ChatGPT is a relatively new technology, existing empirical research on its impacts on
learning is limited. Rudolph et al. (2023) conducted a literature review on ChatGPT and higher
education in January 2023 and found only two peer-reviewed articles and eight unreviewed
preprints. The studies examined various applications of ChatGPT in learning, such as using it to
write academic papers (Zhai, 2022) and having conversations with it (Qaidr, 2022). Other studies
explored learners' perceptions of ChatGPT in language learning (Ali et al., 2023) and its potential
to support L2 writing tasks (Yan, 2023). While some of the studies found that ChatGPT can be
motivating and helpful to learning, others expressed concerns about its impact on academic
integrity and educational equality (Yan, 2023; Shoufan, 2023). For example, some learners
reported that ChatGPT improved their reading and writing skills but not their listening and
speaking skills (Ali et al., 2023). Additionally, some students expressed concerns about the
potential negative impacts of ChatGPT on academic integrity, personal life, and job prospects
7
(Shoufan, 2023). Overall, while some studies suggest that ChatGPT may have benefits for learning,
there are also concerns about its use. More research is needed to fully understand the impacts of
ChatGPT on learning.
Collaborative Interdisciplinary Learning
Research has consistently shown that collaborative interdisciplinary learning offers a
multitude of benefits to learners. For instance, it enables learners to personalize their organization
of knowledge, develop critical thinking and metacognitive skills, and engage in meaningful
collaboration (Alberta Education, 2015; Ivanitskaya et al., 2002; Ledford, 2015). Furthermore, this
approach helps learners apply their knowledge and skills to real-world problems and enhances
their ability to solve complex challenges effectively. Collaborative interdisciplinary learning is
typically characterized by the integration of multiple disciplines, active collaboration towards
shared goals, and the creation of innovative solutions across disciplinary boundaries. This
approach also fosters the advancement of knowledge by identifying and solving new problems,
explaining phenomena, and designing products (Klaassen, 2019; MacLeod & van der Veen, 2020).
The boundaries of different disciplines create a space where students can co-construct and
develop new knowledge as well as contribute and integrate diverse ideas, but navigating this space
is challenging because of disciplinary boundaries and conflicting epistemology (Akkerman &
Bakker, 2011; MacLeod & van der Veen, 2020; Stentoft, 2017). Furthermore, forming
collaborative interdisciplinary learning groups is challenging, as it is impossible to cover all the
disciplinary knowledge needed for various learning problems and challenges. Difficulties include
curriculum structures that focus more on single subjects, limited student representation from
different disciplines, and students' limitations in terms of interdisciplinary knowledge, skills and
methods (Authors, 2022; Ivanitskaya et al., 2002). To address these practical challenges, we
8
explored whether ChatGPT could serve as additional group members to complement students' lack
of disciplinary areas during collaborative interdisciplinary learning.
The Current Study
Despite the benefits of collaborative interdisciplinary learning in higher education, its
implementation can be challenging. One potential solution to overcome this challenge is to
leverage ChatGPT's ability to quickly access information from multiple disciplines. However,
there is a limited amount of empirical research on the use of ChatGPT for education, and students'
attitudes towards chatbots must also be taken into consideration (Okonkwo & Ade-Ibijol, 2021).
Therefore, it is critical for researchers to conduct more studies to examine the appropriate use of
ChatGPT, its potential benefits, challenges, and risks, as well as students' perceptions, particularly
for collaborative interdisciplinary learning. Factors such as the learning content and students'
disciplinary backgrounds and attitudes should be considered when incorporating ChatGPT in
education. These investigations can inform the design of activities and materials for integrating
ChatGPT into teaching and developing acceptable teaching guidelines (Qadir, 2022).
The purpose of our empirical study was to investigate the potential of ChatGPT as a peer
agent to support collaborative interdisciplinary learning of two digital literacy topics, AI and
Blockchain, among STEM and non-STEM undergraduate students. We designed a quasi-
experiment involving 130 students from four tutorial groups and assessed their collaborative
interdisciplinary problem-solving abilities, physical and cognitive engagement, and attitudes
toward ChatGPT. Specifically, we aimed to answer the following research questions:
1. How do students' discipline backgrounds, ChatGPT conditions, and learning topics affect
their collaborative interdisciplinary problem-solving abilities, physical and cognitive
engagement?
9
2. Do STEM and non-STEM students differ in their attitudes toward using ChatGPT in
collaborative interdisciplinary learning? What themes emerge in their reflections on
ChatGPT use?
Our study aimed to provide valuable insights into the potential benefits and challenges of
using ChatGPT in collaborative interdisciplinary learning and inform the development of effective
teaching guidelines that have specifically considered students' disciplinary backgrounds.
Methods
Participants
This study was conducted in March 2023 in a digital literacy course offered to first or
second-year undergraduate students at a public university in Asia. The participants were in four
tutorial classes taught by the same instructor, with 48 students in tutorial 1 (i.e., T1), 48 in T2, 47
in T3, and 40 in T4 (183 students in total), respectively. The project was approved by the
Institutional Review Board at the authors' institution, and students within these four tutorial classes
were recruited. Among the 183 students, 130 (36 females, 54 males, 40 not indicated or did not fill
in the survey, mean age=20.5) participated in this study. The participants enrolled in STEM
programs (72 students from the College of Engineering and College of Science) and non-STEM
programs (58 students from the College of Humanities and Social Sciences and College of
Business). We then assigned students into groups of five to six members and ensured each group
has a good representation of students from different schools to facilitate their interdisciplinary
learning.
Instructional Design
The research team, comprising instructors, researchers, and postgraduate students with
backgrounds in computer science, learning sciences, and psychology, developed two activities on
10
AI and Blockchain topics. When designing the activities, we were guided by the characteristics of
typical interdisciplinary problems synthesized from relevant literature (Kuo et al., 2019; Lam et
al., 2014; MacLeod & van der Veen, 2020; Spelt et al., 2009). These considerations are outlined
below.
Firstly, we aimed to create a real-world scenario for the students to make the task or
problem authentic and relevant. For instance, in the AI module, we created a debating competition
scenario in which interdisciplinary student teams prepared for a debate competition organized by
the university debating society. The debate topic was "AI versus Internet: which one has a more
profound impact on our society?"
Secondly, we encouraged collaborative contributions from various disciplines and made it
explicit that students were expected to think from different disciplinary perspectives. For example,
in the Blockchain module, students were tasked with studying successful Blockchain cases in
different fields worldwide, learning from their experiences, and proposing commercial gaps that
Blockchain could fill. We emphasized that the groups should not be constrained by a specific
industry but can focus on the potential of ideas to make good use of their diverse disciplinary
background asset.
Thirdly, we expected students to make authentic cognitive advancements, such as
explaining a phenomenon, solving a problem, or designing a product that requires high-level
cognitive efforts. In the AI module, we asked groups to collaboratively prepare for the debate by
laying down arguments and counterarguments for each side.
Fourthly, the activities should not be easily solved by ChatGPT or other AI tools to ensure
students' efforts in and control of their learning. For example, the debate topic of "AI versus
Internet, which one has a more profound impact on our society?" does not have a straightforward
11
answer. Therefore, students have to collaboratively work on the activities to deepen their
understanding of the use of AI and the Internet in different fields, consider various factors and
weigh different evidence and examples, and construct coherent and pervasive arguments while
considering potential counterarguments from groups on the other side.
Finally, we provided scaffolding to support students in breaking down the tasks and using
ChatGPT, as interdisciplinary learning tasks can be challenging for students to work on, and
ChatGPT is a new tool for education. We broke down the tasks into smaller steps and provided
scaffolding questions to guide students' learning and thinking. For instance, in the AI module, we
segmented the debating task into four steps: identifying a framework, analyzing and evaluating
applications, building arguments, and presenting argument points.
We used the same task in all four tutorial groups, either with or without ChatGPT support.
We assigned T1, T2, T3, and T4 to either ChatGPT or non-ChatGPT conditions based on two
criteria: 1) including a control group without using ChatGPT, and 2) giving each tutorial at least
one opportunity to participate in the ChatGPT condition during the two-week experiment. Table 1
shows the arrangement of conditions for each tutorial. It is important to note that the ChatGPT
condition included two options: normal ChatGPT, which allowed students to use it freely without
constraints or instructions, and ChatGPT Persona, which involved defining a persona (e.g., Jack,
a 23-year-old Singaporean first-year undergraduate from the business school) to provide additional
disciplinary knowledge to the group and involve them in discussions and problem-solving.
Table 1.
Condition arrangements for students in different tutorials in the two-week quasi-experiment
Tutorials
AI week
Blockchain week
T1
ChatGPT Persona
ChatGPT Persona
T2
Non-ChatGPT
ChatGPT
12
T3
ChatGPT
Non-ChatGPT
T4
ChatGPT
ChatGPT
A flipped classroom learning approach was implemented for the course delivery. Prior to
each class, students were instructed to watch an hour-long pre-recorded video on the week's topic
(i.e., AI and Blockchain) and complete a short self-assessment test. During class, in each tutorial,
small groups with STEM and Non-STEM students were provided with two hours to engage in
collaborative interdisciplinary learning activities, with or without the use of ChatGPT based on
their assigned treatment conditions. The students participated in face-to-face activities with the aid
of Miro, a visual collaboration platform (https://miro.com, see Figure 1). Following each class,
students were required to write individual reflections.
Figure 1.
Content Board of Introduction to Artificial Intelligence
Learning Environment
In Miro, we used navigation and content boards for each tutorial and group, respectively,
13
as shown in Figure 1. We included stickers of various shapes and colors to track the contributions
of students from different schools and ensure inclusivity for color-blind students as well as
provided group note stickers to encourage collaboration. We also provided a breakdown of the
steps and detailed questions for each step to facilitate the learning process. To remind students to
record their interactions with ChatGPT, we separated their response areas and ChatGPT areas, and
placed stickers for ChatGPT's input on the template.
Data Collection
This study included three data sources: (1) students' age, gender and disciplinary
background information collected at the beginning of the semester, (2) weekly surveys on students'
collaborative interdisciplinary problem-solving, and physical and cognitive engagement and (3)
167 pieces of written reflections on the use of ChatGPT by individuals. Next, we would elaborate
on how the surveys and individual reflections were collected.
Collaborative interdisciplinary problem-solving survey
In each week's tutorial, the participants were asked to complete surveys on their
collaborative interdisciplinary problem-solving and engagement using google forms after the
group activities and students' individual reflections. Collaborative interdisciplinary problem-
solving was measured by three questions adopted from the Integration subscale in the Team
Collaboration Questionnaire (Cole et al., 2018) on a five-point Likert scale (1 represents strongly
disagree and 5 strongly agree). The questions include "I investigated an issue with my team to find
an acceptable solution," "I integrated my ideas with my team to come up with a joint solution,"
and "The solution my team and I found to the problem satisfies our needs." The Cronbach's alpha
coefficient for this subscale is .91 (Cole et al., 2018).
Physical and cognitive engagement survey
Engagement (i.e., physical and cognitive engagement) was measured by four questions
14
adapted from Student Engagement (Gerald et al., 2015). This questionnaire measures students'
physical and cognitive engagement in class on a five-point Likert scale (1 denotes strongly
disagree and 5 strongly agree). Questions to measure physical engagement include "I devoted a lot
of energy toward the group activities today" and "I tried my hardest to perform well in the group
activities today"; cognitive engagement is measured by "When working on the group activities, I
paid a lot of attention" and "when working on the group activities, I concentrated on the work."
The Cronbach's alpha coefficient is .93 for physical engagement and .96 for cognitive engagement
(Gerald et al., 2015).
Individual reflection on the use of ChatGPT
To gain insight into students' experience with ChatGPT, we collected their reflections
immediately after completing learning tasks that involved ChatGPT. In each tutorial, we asked
individuals to respond to predetermined questions in a designated reflection area on the Miro board,
as shown in Figure 1. To ensure the learning processes and components were as similar as possible
for students in the non-ChatGPT condition, we also asked them to reflect on two questions in Miro
after their learning activities. Table 2 displays the questions that guided students' reflections in
both the ChatGPT and non-ChatGPT conditions. However, it should be noted that this study
focuses exclusively on reflections from the ChatGPT condition to gain insights into students'
opinions and reflected themes.
Table 2.
Individual reflection questions for the ChatGPT and non-ChatGPT conditions
Individual reflection questions for the
ChatGPT condition
Individual reflection questions for the non-
ChatGPT condition
1. How did you respond to ChatGPT's
responses in this module? E.g., when do
you agree or disagree, and why?
2. What is the most valuable thing you found
from ChatGPT in this module?
1. List one thing that your group has done
well in relation to all other groups.
2. List one thing you will improve in your
next debate exercise.
15
3. What is the least valuable thing you found
from ChatGPT in this module?
Data Analysis
ANOVA
We employed ANOVA and factorial ANOVA to answer RQ1 and investigate the main and
interaction effects of students' disciplinary backgrounds, ChatGPT conditions, and learning topics
on collaborative interdisciplinary problem-solving, physical engagement, and cognitive
engagement. Due to sample size and statistical power considerations, we categorized students'
disciplinary backgrounds into STEM (engineering and science) and non-STEM programs
(humanities and social sciences, and business). We also classified conditions as ChatGPT
(ChatGPT and ChatGPT Persona) and non-ChatGPT. First, we examined the main effect of topics
(AI & Blockchain), ChatGPT conditions, and disciplinary backgrounds on students' collaborative
interdisciplinary problem-solving, physical engagement, and cognitive engagement. We then
conducted a factorial ANOVA to investigate the interaction effects of students' learning topics,
ChatGPT conditions, and disciplinary backgrounds on the dependent variables.
Sentiment analysis and qualitative analysis
To answer RQ2, which investigates potential differences in STEM and non-STEM
students' attitudes towards using ChatGPT during their collaborative interdisciplinary learning and
the themes emerging from their reflections, we performed sentiment analysis using the DistilBERT
model (Sanh et al., 2019). Each reflection was labeled as either positive or negative by the model.
The third author reviewed the sentiment analysis results and highlighted the ones different from
her interpretation, resulting in 17 out of 167 reflections being highlighted. The first author then
reviewed the highlighted reflections and discussed them with the third author until they reached a
consensus. The agreed sentiment analysis results were subjected to a t-test to examine if STEM
16
and non-STEM students had different attitudes toward using ChatGPT.
To examine the themes that emerged from students' reflections, the first author coded the
reflections on using ChatGPT during collaborative interdisciplinary learning. First, the author
separated the student reflections from T1 into relatively positive and negative ones. Next, the
author synthesized the coding and identified common themes through an inductive reasoning
approach, a "bottom-up" analytical strategy (Thorne, 2000). Inductive reasoning allows
researchers to use the data to generate ideas by interpreting and structuring the meanings derived
from data (Thorne, 2000). Specifically, the first author first labeled the theme "Efficiency" which
was mostly mentioned by students and then read through the remaining data to see if they fit the
existing identified themes. If yes, the relevant quotes would be labeled with the theme; otherwise,
new themes will be added until all the data has been assigned a theme. Using the same process,
the author analyzed data from other tutorials to examine whether new themes needed to be added
until all reflections had been analyzed. The third author then read through all the reflections,
checked the reflections against the themes, and confirmed the themes identified by the first author
was appropriate and comprehensive.
Results
RQ1: Main and interaction effects of students' disciplinary backgrounds, ChatGPT
conditions and learning topics
Table 3 displays the descriptive statistics for collaborative interdisciplinary problem-
solving, physical engagement, and cognitive engagement of STEM and non-STEM students over
two weeks of classes in ChatGPT and non-ChatGPT conditions. STEM students had higher
average scores in collaborative interdisciplinary problem-solving, physical engagement, and
cognitive engagement under the ChatGPT condition than non-STEM students. The only exception
17
is that in the Blockchain module, non-STEM students had higher cognitive engagement (mean
score of 3.91) than STEM students (mean score of 3.83). Non-STEM students had relatively higher
average scores on the dependent variables in the non-ChatGPT condition.
Table 3.
Descriptive statistics of physical engagement, cognitive engagement and collaborative
interdisciplinary problem-solving
Topics
ChatGPT
Non-ChatGPT
STEM
STEM
Non-STEM
AI &
Internet
Number of
Participants
69
27
21
Number of
Responses
45
10
9
Collaborative
Interdisciplinary
Learning
M(SD)
4.24 (.59)
4.06 (.51)
4.22 (1.16)
Physical
Engagement
M(SD)
4.21 (.60)
4.20 (.63)
4.55 (.72)
Cognitive
Engagement
M(SD)
4.177 (.06)
4.15 (.33)
4.55 (.72)
Blockchain
Number of
Responses
46
17
12
Collaborative
Interdisciplinary
Learning
M(SD)
3.84 (.69)
3.82 (.71)
4.13 (.59)
Physical
Engagement
M(SD)
3.83 (.82)
3.52 (.81)
4.08 (.66)
Cognitive
Engagement
M(SD)
3.83 (.85)
3.76 (.73)
4.16 (.57)
18
As shown in Table 4, the factorial ANOVA results indicate significant main effects of
topics on collaborative interdisciplinary problem-solving (F(1, 202) = 13.19, p < .01, 󰇜,
cognitive engagement (F(1, 202) = 5.91, p < .05, 󰇜, and physical engagement (F(1, 202)
= 15.15, p < .001, 󰇜. A post hoc Tukey test shows students' collaborative interdisciplinary
learning, physical engagement, and cognitive engagement were consistently higher for the AI &
Internet topic. Different ChatGPT conditions did not have a main effect on the dependent variables.
Table 4.
ANOVA results: Main and interaction effects of students' disciplinary backgrounds, ChatGPT
conditions, and learning topics
Main Effect
Interaction Effect
Disciplinary
Background
F(1, 202)
Condition
F(1, 202)
Topic
F(1, 202)
Condition ×
Disciplinary
Background
F(1, 202)
Topic ×
Disciplinary
Background
F(1, 202)
Topic ×
Condition
F(1, 202)
Topic ×
Disciplinary
Background
× Conditions
F(1, 202)
Collaborative
Interdisciplinary
Problem-solving
.00
.20
12.40***
1.83
.10
.85
.06
Physical
Engagement
.00
.45
15.15****
6.29**
.31
1.05
.05
Cognitive
Engagement
.12
1.46
5.91**
3.37*
1.24
.52
.36
Note. Topics: AI & Internet versus Blockchain; Conditions: ChatGPT versus non-ChatGPT;
Disciplinary Backgrounds: STEM versus non-STEM.
*p < .1; **p < .05; ***p <.01; ****p < .001
Regarding interaction effects, we found a marginal interaction effect between students'
disciplinary background and ChatGPT condition for cognitive engagement (F(1, 202) = 3.37, p =
0.06, 󰇜, and a significant interaction effect for physical engagement (F(1, 202) = 6.29, p
< .05, 󰇜. As plotted in Figure 2, the post hoc comparisons using the t-test with Bonferroni
correction show that non-STEM students were more physically and cognitively engaged during
19
collaboratively learning in the non-ChatGPT condition. No significant difference in the level of
engagement was found for STEM students between ChatGPT and non-ChatGPT conditions.
Figure 2.
Interaction effects between disciplinary backgrounds and ChatGPT conditions on physical
engagement and cognitive engagement
RQ2: STEM and non-STEM Students' opinions on the use of ChatGPT and the themes that
emerged in their reflections
In our sentiment analysis, we assigned a value of "1" to all positive posts and "-1" to
negative posts, and conducted a t-test to compare the sentiment scores of STEM and non-STEM
students. The results showed that there was no significant difference in the opinions of STEM and
non-STEM students on the use of ChatGPT in collaborative interdisciplinary learning (t(137.2)
= .92, p = .35). However, we wanted to gain a deeper understanding of their perspectives, so we
analyzed all individual reflections from STEM and non-STEM students. Table 5 presents the eight
positive and eight negative themes that emerged from our analysis. Each theme is described in
detail below.
Table 5.
Positive and negative themes generated from students' individual reflections
Positive themes
Negative themes
P1: Efficiency
N1:Inaccuracy, fact-checking and
verification required
P2: Accurate and insightful, unbiased,
N2: Misleading and biased decision-making
19
reasonable and justified, structured and
comprehensive and targeted responses
P3: Providing different perspectives and
supporting brainstorming and idea generation
N3: Generic but not contextualized, brief but
not profound; therefore, additional probing
or research is required
P4: Addressing knowledge gaps and
complementing thinking
N4: Depending on prompts and having usage
context limits
P5: Inspiration and refinement through further
research and probing
N5: Repetition, wordiness and no innovative
or original ideas
P6: Not judgmental and can generate human-
like responses
N6: Harming self-discipline and thinking
P7: Safeguards for its misuse
N7: Human labor replacement
P8: Providing language support
N8: Technical issue
Positive themes
P1: Efficiency. In students' reflections, a frequently mentioned theme is the efficiency of
ChatGPT. The participants were surprised at how quickly and efficiently ChatGPT responded to
their input/prompts and consolidated information from different sources. As a result, it saves
students' time. For instance, in the AI module, a STEM student was surprised by ChatGPT's speed
by exclaiming, "ChatGPT is able to give us our desired information at O(1) speed … Insane time
complexity!" Similarly, in the same module, a non-STEM student reflected that "ChatGPT
reduces the time needed for humans to analyze the different sources of data and compile it together
to form a concrete argument."
P2: Accurate and insightful, unbiased, reasonable and justified, structured and
comprehensive and targeted responses. Students found ChatGPT's responses accurate,
insightful, logical, reasonable, unbiased, structured, comprehensive, and customized. One STEM
student praised ChatGPT's accuracy in the AI module, while a non-STEM student appreciated its
ability to offer "a lot of insightful responses from a different perspective." Students noted
ChatGPT's unbiased and neutral approach, with one stating, "ChatGPT took a neutral side."
Students also appreciated ChatGPT's justification, comprehensiveness, and structure. One STEM
student in the Blockchain module noted that ChatGPT "helps to put structure to scattered ideas and
20
thoughts," while a non-STEM student appreciated the bot's comprehensive responses with reasons
to justify them. Some students believed ChatGPT customized its answers to their questions.
P3: Providing different perspectives and supporting brainstorming and idea
generation. Participants commonly mentioned ChatGPT's ability to provide different perspectives,
which they found helpful. One STEM student in the AI module noted, "ChatGPT provides a very
broad answer that gives the users a good idea of the different information surrounding the questions
they asked." In the Blockchain module, a STEM student found it useful as an idea generator, saying,
"ChatGPT offers a broad range of content that even I wouldn't think of." ChatGPT's affordances
make it appropriate for supporting users to brainstorm ideas and start their work efficiently and
comprehensively. For instance, a student reflected in the AI module that "with the key
points/factors that ChatGPT generates, it was easy to get started on the research."
P4: Addressing knowledge gaps and complementing thinking. Some students found
ChatGPT to be useful for addressing knowledge gaps and complementing thinking. According to
a STEM student in the AI module, "It can address gaps in technical knowledge, allowing for those
who don't have that knowledge to apply it similarly." Another STEM student in the Blockchain
module reflected, "ChatGPT uncovered information that I usually would not think of; it helps me
to see other perspectives/applications." Similarly, some non-STEM students found that ChatGPT
"found other applications of blockchain technology other than crypto and finance" and "can help
to complement our ideas and thinking."
P5: Inspiration and refinement through further research and probing. Some students
found that ChatGPT inspires them and allows them to build on its responses and do further research
to improve the responses. One non-STEM student in the AI module responded, "The responses are
very insightful (when I have really no ideas), leading me to think more and build more opinions
21
on them." Similarly, a STEM student wrote, "I feel like ChatGPT inspired me to think broadly, as
if I was talking to another person. ChatGPT allowed me to dive into another thought process,
giving me a more complete look into the issue." ChatGPT's ability to retain conversation history
allows users to "ask it things that you have previously mentioned." However, some students
indicate the need for further research rather than total dependence on ChatGPT.
P6: Not judgemental and can generate human-like responses. Participants like
ChatGPT because it can generate human-like responses and serve as a supporting and non-
judgmental teammate. One STEM student indicated, "ChatGPT inspired me to think broadly, as if
I was talking to another person." Another STEM student added, "ChatGPT is very informative and
can construct logical explanations eerily similar to how a human can process thought and decision
making." A non-STEM student suggested what they liked most about ChatGPT was its "ability to
generate human-like responses." Another non-STEM student discussed that being "committed,
knowledgeable, caring, willing to participate, does not judge you no matter how stupid your
question is makes ChatGPT a good teammate."
P7: Safeguarding for its misuse. Although rarely mentioned by the participants, another
distinct theme that students like about ChatGPT is its affordance of safeguarding its misuse.
Specifically, in the AI module, a STEM student wrote, "ChatGPT has safeguards for its misuse or
its use for harmful/criminal activities."
P8: Providing language support. Although the two activities do not emphasize spelling
or grammar issues, after the activities, a few students reflected that ChatGPT supported their
language use ("It also useful when it comes to language matters") or helped them check grammar
("We will agree to use ChatGPT during the research stage and use it for grammar checking").
Negative themes
N1: Inaccuracy, fact-checking and verification are required. One major concern
22
students expressed regarding ChatGPT is its possibility of misinterpreting inputs and providing
unreliable or false answers. In the AI module, a non-STEM student expressed concern: "ChatGPT
may sometimes be unable to understand our question and hence not generate an accurate answer."
Similarly, in the Blockchain module, students reflected that "credibility needs to be verified" and
"not everything they explain is always correct." Moreover, some students acknowledged that
ChatGPT's responses might not be up-to-date and that it could be challenging to fact-check them.
As a result, students emphasized the need for fact-checking to ensure the accuracy of ChatGPT's
responses and to prevent misinformation. However, they also recognized that this could be difficult,
given ChatGPT's extensive knowledge and the need to understand the topic to fact-check its
responses.
N2: Misleading or biased decision-making. Students expressed concerns about
ChatGPT's potential for providing inaccurate or unreliable responses, leading to misunderstanding
information and biased decision-making. In the AI module, one student noted, "ChatGPT may
provide an insight more inclined to whoever is feeding ChatGPT the information." Similarly, in
the Blockchain module, a student mentioned that "ChatGPT is limited to information before 2021
and only has sparse knowledge on real-world events after that year." Another student highlighted
that ChatGPT could return information that supports the user's perspective, potentially ignoring
other sides of the argument. Additionally, some users may not be able to detect mistakes made by
ChatGPT due to their limited understanding of the topic. As a result, students emphasized the
importance of fact-checking and verifying ChatGPT's responses to prevent misinformation and
bias.
N3: Generic but not contextualized, brief but not profound; therefore, additional
probing or research is required. Students have concerns about the limitations of ChatGPT, with
23
some citing its tendency to provide generic, vague, and superficial responses that lack context,
specificity, and depth. One student noted that "ChatGPT doesn't do well in terms of giving exact
case studies, but only gives a generic answer to questions." Some students also felt that ChatGPT's
responses lacked depth, with comments such as "The responses are usually very brief and not very
in-depth" and "the lack of insights in the information." Therefore, students suggested that
additional research or using tools like "google scholar, scientific websites..." could help obtain
more in-depth analyses.
N4: Depending on prompts and having usage context limits. Students have raised
concerns about ChatGPT's limitations, particularly in generating comprehensive responses. Two
non-STEM students in the AI module noted, "ChatGPT does not provide information fully at one
go, several prompts are needed to force AI to think" and "an answer may not really answer the
question if not prompted correctly." Participants generally agree that learners need to adjust their
prompts to get comprehensive responses from ChatGPT. Some students also pointed out that
ChatGPT has limitations in evaluating complex concepts, tasks requiring personal opinions, and
topics requiring nuanced perspectives. For example, a STEM student in the AI module commented,
"ChatGPT is not very useful to draw insights on concepts that are tough to evaluate, lacking a
certain form of intelligence when it comes to personal opinion matters."
N5: Repetition, wordiness and no innovative or original ideas. Some students are
concerned about the repetitiveness and wordiness of ChatGPT's responses. They suggest refining
ChatGPT's responses by filtering out unneeded information and avoiding excessive repetition. For
instance, students recommended asking ChatGPT to "narrow down/cut out unneeded information."
Students also expressed frustration with ChatGPT's limitations in generating new ideas and
original content, as one student noted that ChatGPT struggles to "come up with new ideas, in terms
24
of innovation and original ideas."
N6: Impairing self-discipline and thinking. The participants raised concerns about how
ChatGPT negatively affected their critical thinking and effort to generate ideas. In the AI module,
a STEM student noted, "ChatGPT gives concise answers which take away the need for critical
thinking." Similarly, in the Blockchain module, a STEM student mentioned that "it stops me from
thinking," and another student reflected, "ChatGPT sounded so confident so I just believed in him
and took his responses and based my thoughts on it, showing the influence and impact AI has on
us." The students recognized that relying too much on ChatGPT could make them lazy to think for
themselves, with one STEM student noting, "repeatedly using ChatGPT can be a bad habit as it
makes us lazy to think (by) ourselves." Others expressed similar concerns, stating that "ChatGPT
makes me tempted not to expend effort to generate ideas myself" and "My brain didn't think much
because AI is too professional." However, the participants also acknowledged the need for self-
discipline, with one saying they had been "too dependent on it," and another acknowledging "there
is the urge to be reliant on ChatGPT."
N7: Human labor replacement. A few participants were worried that ChatGPT was so
capable of providing wholesome and structured responses that it might replace their jobs. In the
AI module in the ChatGPT persona condition in which student groups could define persona to
support and complement their debating teams, two students reflected that "What extra value can
we bring? I might as well make 6 different personas and ask ChatGPT to debate. I think it will do
a better job" and "ChatGPT may replace jobs. This is a concern that can be readily addressed."
N8: Technical issue. A few participants encountered technical or stability issues with
ChatGPT (e.g., network errors or unknown reasons) in the Blockchain module. They included
these issues in their reflections and indicated they used Google instead because "For me, ChatGPT
25
didn't really work due to network errors as they told me. Only managed to answer 1 question before
it failed again" and "ChatGPT not working due to some unknown reason, still faster to google."
Discussions
The impact of ChatGPT on students' in-class collaborative interdisciplinary problem-
solving and engagement was examined. While self-reported engagement and collaborative
interdisciplinary learning did not differ between ChatGPT and non-ChatGPT conditions,
quantitative analysis revealed higher levels of collaborative interdisciplinary problem-solving,
physical engagement, and cognitive engagement in the AI module compared to the Blockchain
module. This result is consistent with prior research on the importance of instructional design in
computer-supported collaborative learning (Luckin & Cukurova, 2019). In detail, the AI module
involved formulating arguments to defend a side in a debate, whereas the Blockchain module
required conducting a case analysis, which mainly involved factual information that students could
adopt from websites or ChatGPT responses. Our result aligns with a meta-analysis showing that
computer-supported collaborative learning can harm case-based learning (Jeong et al., 2019),
where learners might overly rely on website information and compromise interactions with peers
(Kemp et al., 2019). Hence, instructors need to consider preserving students' agency and critical
thinking while maximizing the advantages of ChatGPT (e.g., efficiency, comprehensive, and
targeted responses) when designing learning activities.
We found that non-STEM students tended to be more physically and cognitively engaged
in the learning activities without help from ChatGPT. In contrast, STEM students showed no
significant differences in collaborative interdisciplinary problem-solving and engagement between
ChatGPT and non-ChatGPT conditions. Previously, Jeong et al. (2019) found that the disciplinary
backgrounds of students moderate the effect of technology on learning outcomes. Thongsri et al.
26
(2019) found that non-STEM students, especially those with low computer self-efficacy, have a
lower behavioral intention to use new learning technologies. We conjecture that non-STEM
students might spend extra time getting familiar with ChatGPT functions and were left out during
group discussions. STEM students, on the other hand, have high digital literacy (Thongsri et al.,
2019) and are more comfortable integrating new technology, such as ChatGPT, into their learning
process. They naturally treat ChatGPT as a new form of seeking information and can integrate the
ChatGPT responses into their problem-solving process as they do with traditional web searches.
However, more research is needed to understand how students from different disciplines respond
to ChatGPT during learning activities and how the unevenness of digital literacy influences
collaborative problem-solving.
The sentiment analysis results showed no significant difference in the perceptions towards
using ChatGPT between non-STEM and STEM students. However, it is important to consider how
we framed the reflection questions, as shown in Table 2. By asking participants to reflect on the
most valuable and least valuable aspects of using ChatGPT in the module, we may have
inadvertently biased the responses towards positive and negative perceptions, regardless of the
disciplinary background. In contrast, Shoufan (2023) asked undergraduate students an open-ended
question, "What do you think of ChatGPT? Think deeply and write down whatever comes into
your mind!" and found that almost 67% (56/171) of the comments were positive. To gain a deeper
understanding of students' perceptions towards using ChatGPT in learning, future studies may
consider using more open-ended and less biased methods, such as semi-structured interviews.
We qualitatively analyzed students' reflections on their experience using ChatGPT and
identified several positive themes. First, ChatGPT provided students with different perspectives,
supported brainstorming and idea generation, and addressed knowledge gaps. Second, ChatGPT
27
complemented thinking and inspired students to refine their ideas through further research and
probing. Third, students appreciated ChatGPT's non-judgmental nature and ability to generate
human-like responses. These findings are consistent with McMurtrie's (2023) suggestion that AI
tools may "help ignite the imaginative process." Shoufan (2023) also identified similar positive
themes, such as ChatGPT being "helpful for learning" and "helpful for work" and generating
human-like quality of conversations.
Based on students' reflections, some negative impacts of ChatGPT on learning were
identified, including the potential for misleading or biased decision-making, repetitive or wordy
responses, and a lack of innovative or original ideas. Tlili et al. (2023) found that learners reported
similar concerns about ChatGPT's responses being potentially misleading or even eroding their
creative and critical thinking skills. Qadir (2022) also noted that ChatGPT can generate
misinformation due to limitations in LLM models, training data quality, and prompt inputs.
OpenAI itself warns that ChatGPT can give seemingly confident but potentially unreliable answers
to complex questions (OpenAI, 2022), and CEO Sam Altman has cautioned against overreliance
on ChatGPT, stating that it is not yet reliable enough for anything important (Altman, 2022). As
Rudolph et al. (2023) suggested, ChatGPT may be less effective at handling content that requires
critical thinking and analysis and should be seen as a tool to complement and enhance learning
rather than replace teachers' and students' roles. Students in this study also highlighted the
importance of developing literacy to evaluate AI-generated content. They emphasized that AI
cannot replace the learning process and self-discipline is necessary for critical thinking.
The prevailing notion is that students will plagiarize content generated by ChatGPT and
integrate it into their academic assignments. However, the findings from the study indicate that
students are aware of the limitations of ChatGPT, as evidenced by the negative feedback they
28
provide, without the need for explicit instruction from educators. The findings are positive
indicators from an instructional standpoint, with implications for more effective pedagogy. Rather
than prohibiting the use of ChatGPT, educators can initiate a dialogue with students about the
challenges associated with its use. Allowing students to utilize ChatGPT in the classroom and
facilitating discussions about its potential benefits and drawbacks could be a constructive approach.
On the other hand, only a small number of students reflected on ethical considerations
associated with using ChatGPT in their learning. OpenAI's efforts to avoid offensive outputs may
have led students to overlook ethical considerations. OpenAI uses reinforcement learning, human
feedback, recursive reward modeling, and moderation models to ensure the safe use of ChatGPT
(Ouyang et al., 2022; Markov et al., 2022). However, ethical considerations should still be
addressed in the use of AI technology, and future research should explore how to better incorporate
these considerations into the learning experience.
ChatGPT and other AI models have the potential to revolutionize education, but it is
important to carefully consider their impact on students' self-discipline and critical thinking, as
well as potential risks such as repetitive and conventional ideas. Additionally, factors such as
learning content, instructional design, and students' differences should be considered when
implementing ChatGPT in education. As such, there is a need for collective efforts to research
ChatGPT and other forthcoming AI tools at philosophical, epistemic, and empirical levels to ensure
they do more good than harm in the rapidly changing era.
Implications
ChatGPT has emerged as a significant technological advancement in education. While
students embrace ChatGPT, they are also aware of its potential drawbacks. This study has practical
and epistemological implications for future research and practice in higher education.
29
Epistemologically, educators need to rethink critical questions, such as what competencies students
should develop in the era of powerful AI, how to consider knowledge generated by ChatGPT, and
how to anticipate potential ethical issues. At a practical level, researchers need to explore activities
and assessments that can help students develop and evaluate required competencies and balance
the relationships between students, instructors, and AI in classrooms. These questions and issues
need to be addressed to ensure that ChatGPT and other AI tools are used effectively in education.
Limitations and future directions
The current study has several limitations that should be addressed in future research. First,
although the sample size of the study is decent, the participants were only first-year undergraduate
students who participated in two relatively similar activities on AI and Blockchain in a digital
literacy course at a single Asian university. To generalize the findings, future research should
survey more senior students with different cultural backgrounds and students participating in
different learning activities in various courses over a longer period of time. Additionally,
researchers should consider the nuances between using ChatGPT in general and using ChatGPT
persona, which allowed students to define their "teammate." Second, the study focused on students'
perceived benefits and drawbacks of ChatGPT after hands-on experience in class, but did not
analyze their interaction data with ChatGPT. Therefore, future studies should investigate how
ChatGPT may influence learning positively or negatively by analyzing students' interaction data
with ChatGPT. Moreover, researchers should design learning activities that leverage ChatGPT and
assess students' learning in its context. Finally, although the study followed the five characteristics
of typical interdisciplinary problems for both the AI and Blockchain modules, there might still be
some nuanced differences between the two tasks. Future research should consider these subtle
differences between task design and topics when applying ChatGPT or other AI tools.
30
Conclusion
ChatGPT has garnered significant attention in higher education due to its potential to
provide personalized and interdisciplinary learning opportunities. However, despite researchers'
hypotheses about ChatGPT's impact, there is a lack of empirical research on its application in
controlled classroom settings, particularly in interdisciplinary contexts. To address this research
gap, we conducted a two-week quasi-experimental study with 130 undergraduate students,
examining their collaborative interdisciplinary problem-solving, engagement, and perceptions of
ChatGPT in a digital literacy course at an Asian public university. Our findings suggest that to
optimize the use of ChatGPT, educators need to consider both the disciplinary backgrounds of
students and the topics being taught (or how the activity is designed), as students with different
backgrounds react differently to ChatGPT across different topics. Furthermore, our analysis of
students' reflections identified positive and negative themes regarding using ChatGPT, providing
valuable insights for educators seeking to integrate ChatGPT into their teaching. Overall, this work
contributes to the limited empirical research on ChatGPT's use in higher education and provides
practical insights into optimizing its application.
31
References
Abdelghani, R., Wang, Y. H., Yuan, X., Wang, T., Sauzéon, H., & Oudeyer, P. Y. (2022). GPT-3-driven
pedagogical agents for training children's curious question-asking skills. arXiv preprint arXiv:2211.14228.
Akkerman, S. F., & Bakker, A. (2011). Boundary Crossing and boundary objects. Review of Educational Research,
81(2), 132169. https://doi.org/10.3102/0034654311404435
Alberta Education. (2015). Interdisciplinary Learning.
https://www.learnalberta.ca/content/kes/pdf/or_ws_tea_elem_05_interdis.pdf
Ali, J. K. M., Shamsan, M. A. A., Hezam, T. A., & Mohammed, A. A. (2023). Impact of ChatGPT on Learning
Motivation: Teachers and Students' Voices. Journal of English Studies in Arabia Felix, 2(1), 41-49.
Clarizia, F., Colace, F., Lombardi, M., Pascale, F., & Santaniello, D. (2018). Chatbot: An education support system
for student. Cyberspace Safety and Security, 291302. https://doi.org/10.1007/978-3-030-01689-0_23
Cole, M. L., Cox, J. D., & Stavros, J. M. (2018). SOAR as a mediator of the relationship between emotional
intelligence and collaboration among professionals working in teams: Implications for entrepreneurial teams.
SAGE Open, 8(2), 2158244018779109.
Gerald F. Burch, Nathan A. Heller, Jana J. Burch, Rusty Freed & Steve A. Steed (2015). Student Engagement:
Developing a Conceptual Framework and Survey Instrument, Journal of Education for Business, 90:4, 224-
229, DOI: 10.1080/08832323.2015.1019821
Guo, B., Zhang, X., Wang, Z., Jiang, M., Nie, J., Ding, Y., ... & Wu, Y. (2023). How Close is ChatGPT to Human
Experts? Comparison Corpus, Evaluation, and Detection. arXiv preprint arXiv:2301.07597.
Hu, K. (2023). ChatGPT sets record for fastest-growing user base. Reuters.
https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/
Ivanitskaya, L., Clark, D., Montgomery, G., & Primeau, R. (2002). Interdisciplinary Learning: Process and
Outcomes. Innovative Higher Education, 27(2), 95111. https://doi.org/10.1023/A:1021105309984
Jeong, H., Hmelo-Silver, C. E., & Jo, K. (2019). Ten years of computer-supported collaborative learning: A meta-
analysis of CSCL in STEM education during 20052014. Educational research review, 28, 100284.
Jeon, J. (2021). Chatbot-Assisted Dynamic Assessment (CA-DA) for L2 vocabulary learning and diagnosis.
Computer Assisted Language Learning, 127. https://doi.org/10.1080/09588221.2021.1987272
Kasneci, E., Seßler, K., Küchemann, S., Bannert, M., Dementieva, D., Fischer, F., Gasser, U., Groh, G.,
Günnemann, S., Hüllermeier, E., Krusche, S., Kutyniok, G., Michaeli, T., Nerdel, C., Pfeffer, J., Poquet, O.,
Sailer, M., Schmidt, A., Seidel, T., … Kasneci, G. (2023). Chatgpt for good? on opportunities and challenges
of large language models for education. https://doi.org/10.35542/osf.io/5er8f
Klaassen, R. G. (2018). Interdisciplinary education: A case study. European Journal of Engineering Education,
43(6), 842859. https://doi.org/10.1080/03043797.2018.1442417
Kuhail, M. A., Alturki, N., Alramlawi, S., & Alhejori, K. (2022). Interacting with educational Chatbots: A
systematic review. Education and Information Technologies, 28(1), 9731018.
https://doi.org/10.1007/s10639-022-11177-3
Kung, T. H., Cheatham, M., Medenilla, A., Sillos, C., De Leon, L., Elepaño, C., ... & Tseng, V. (2023). Performance
of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models. PLoS
digital health, 2(2), e0000198.
Ledford, H. (2015). How to solve the world's biggest problems. Nature, 525(7569), 308311.
https://doi.org/10.1038/525308a
Looi Chee Kit Wong Lung Hsiang, Kit, L. C., Hsiang, W. L., Looi Chee Kit & Wong Lung Hsiang, & Bookmark
Bookmark Share WhatsApp Telegram Face. (n.d.). Commentary: Want chatgpt to do your homework? learn
how to use it first. CNA. Retrieved March 22, 2023, from
https://www.channelnewsasia.com/commentary/chatgpt-schools-students-teachers-education-singapore-ai-
literacy-3338881
Luckin, R., & Cukurova, M. (2019). Designing educational technologies in the age of AI: A learning sciences‐
driven approach. British Journal of Educational Technology, 50(6), 2824-2838.
MacLeod, M., & van der Veen, J. T. (2020). Scaffolding interdisciplinary project-based learning: A case study.
European Journal of Engineering Education, 45(3), 363377.
https://doi.org/10.1080/03043797.2019.1646210
Markov, T., Zhang, C., Agarwal, S., Eloundou, T., Lee, T., Adler, S., ... & Weng, L. (2022). A holistic approach to
undesired content detection in the real world. arXiv preprint arXiv:2208.03274.
32
McMurtrie, B. (2023, January 5). Teaching: Will ChatGPT change the way you teach?. The Chronicle of Higher
Education. https://www.chronicle.com/newsletter/teaching/2023-01-05
Moore, S., Nguyen, H. A., Bier, N., Domadia, T., & Stamper, J. (2022). Assessing the quality of student-generated
short answer questions using GPT-3. Lecture Notes in Computer Science, 243257.
https://doi.org/10.1007/978-3-031-16290-9_18
Murtarelli, G., Gregory, A., & Romenti, S. (2021). A conversation-based perspective for shaping ethical human
machine interactions: The particular challenge of Chatbots. Journal of Business Research, 129, 927935.
https://doi.org/10.1016/j.jbusres.2020.09.018
Okonkwo, C. W., & Ade-Ibijola, A. (2021). Chatbots applications in education: A systematic review. Computers
and Education: Artificial Intelligence, 2, 100033. https://doi.org/10.1016/j.caeai.2021.100033
OpenAI API. (2022). Retrieved May 1, 2023, from https://platform.openai.com/docs/chatgpt-education
Ouyang, L., Wu, J., Jiang, X., Almeida, D., Wainwright, C., Mishkin, P., ... & Lowe, R. (2022). Training language
models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35,
27730-27744.
Park, S. Y. (2009). An analysis of the technology acceptance model in understanding university students' behavioral
intention to use e-learning. Journal of Educational Technology & Society, 12(3), 150-162.
Qadir, J. (2022). Engineering education in the era of chatgpt: Promise and pitfalls of Generative AI for Education.
https://doi.org/10.36227/techrxiv.21789434.v1
Rodrigo, M. M., Baker, R. S., Agapito, J., Nabos, J., Repalam, M. C., Reyes, S. S., & San Pedro, M. O. (2012). The
effects of an interactive software agent on student affective dynamics while using ;an intelligent tutoring
system. IEEE Transactions on Affective Computing, 3(2), 224236. https://doi.org/10.1109/t-affc.2011.41
Rudolph, J., Tan, S., & Tan, S. (2023). ChatGPT: Bullshit spewer or the end of traditional assessments in higher
education? Journal of Applied Learning and Teaching, 6(1), Article 1.
https://doi.org/10.37074/jalt.2023.6.1.9
Sailer, M., Bauer, E., Hofmann, R., Kiesewetter, J., Glas, J., Gurevych, I., & Fischer, F. (2023). Adaptive feedback
from artificial neural networks facilitates pre-service teachers' diagnostic reasoning in simulation-based
learning. Learning and Instruction, 83, 101620. https://doi.org/10.1016/j.learninstruc.2022.101620
Sanh, V., Debut, L., Chaumond, J., & Wolf, T. (2019). DistilBERT, a distilled version of BERT: smaller, faster,
cheaper and lighter. arXiv preprint arXiv:1910.01108.
Sallam, M. (2023, March). ChatGPT utility in healthcare education, research, and practice: Systematic review on the
promising perspectives and valid concerns. In Healthcare (Vol. 11, No. 6, p. 887). MDPI.
Sarsa, S., Denny, P., Hellas, A., & Leinonen, J. (2022). Automatic generation of programming exercises and code
explanations using large language models. Proceedings of the 2022 ACM Conference on International
Computing Education Research V.1. https://doi.org/10.1145/3501385.3543957
Shoufan, A. (2023). Exploring Students' Perceptions of ChatGPT: Thematic Analysis and Follow-up Survey. IEEE
Access.
Stentoft, D. (2017). From saying to doing interdisciplinary learning: Is problem-based learning the answer? Active
Learning in Higher Education, 18(1), 5161. https://doi.org/10.1177/1469787417693510
Tai, T.-Y., & Chen, H. H.-J. (2020). The impact of google assistant on adolescent EFL learners' willingness to
communicate. Interactive Learning Environments, 118. https://doi.org/10.1080/10494820.2020.1841801
Thongsri, N., Shen, L., & Bao, Y. (2020). Investigating academic major differences in perception of computer self-
efficacy and intention toward e-learning adoption in China. Innovations in Education and Teaching
International, 57(5), 577-589.
Thorne, S. (2000). Data analysis in qualitative research. Evidence-based nursing, 3(3), 68-70.
Thorp, H. H. (2023). CHATGPT is fun, but not an author. Science, 379(6630), 313313.
https://doi.org/10.1126/science.adg7879
Tlili, A., Shehata, B., Adarkwah, M. A., Bozkurt, A., Hickey, D. T., Huang, R., & Agyemang, B. (2023). What if the
devil is my guardian angel: Chatgpt as a case study of using Chatbots in education. Smart Learning
Environments, 10(1). https://doi.org/10.1186/s40561-023-00237-x
Winkler, R., Hobert, S., Salovaara, A., Söllner, M., & Leimeister, J. M. (2020). Sara, the lecturer: Improving
learning in online education with a scaffolding-based conversational agent. Proceedings of the 2020 CHI
Conference on Human Factors in Computing Systems. https://doi.org/10.1145/3313831.3376781
Yan, D. (2023). Impact of ChatGPT on learners in a L2 writing practicum: An exploratory investigation. Education
and Information Technologies, 1-25.
Zhai, X. (2022). CHATGPT user experience: Implications for education. SSRN Electronic Journal.
https://doi.org/10.2139/ssrn.4312418
33
Zhu, M., Liu, O. L., & Lee, H.-S. (2020). The effect of automated feedback on revision behavior and learning gains
in formative assessment of scientific argument writing. Computers & Education, 143, 103668.
https://doi.org/10.1016/j.compedu.2019.103668
... The release of ChatGPT (OpenAI, 2023) shows new premises to address these challenges, given its power to be involved in reasonable conversations of various disciplines with learners and its ability to take and maintain a defined persona and identity (Qadir, 2023;Zhu et al., 2023). A persona is a detailed representation of a user's characteristics, behaviors, and goals (Cooper, 1999), and it is widely used in the Human-Computer Interaction field. ...
... However, it should also be noted that we did not find significant differences in integration or cognitive advancement among the conditions using ChatGPT or not. These results indicate that as researchers (Cooper, 2023;Zhu et al., 2023) suggested, Chat-GPT will not provide a "one-size-fits-all" solution for educational issues such as interdisciplinary learning; instead, when cultivating relatively higher-level skills like integrating ideas, critical thinking, and making cognitive advancement, students' efforts and teachers' deliberate activity design are still critical. While taking advantage of ChatGPT's benefits, students and teachers must be cautious of its potential threats to students' critical thinking and deep understanding/learning (Dwivedi et al., 2023). ...
Article
Full-text available
The significance of interdisciplinary learning has been well-recognized by higher education institutions. However, when teaching interdisciplinary learning to junior undergraduate students, their limited disciplinary knowledge and underrepresentation of students from some disciplines can hinder their learning performance. ChatGPT’s ability to engage in human-like conversations and massive knowledge grounded in different disciplines holds promise in enriching undergraduate students with the disciplinary knowledge that they lack. In this exploratory study, we engaged 130 undergraduate students in a three-condition quasi-experiment to examine how ChatGPT influences their demonstrated and perceived interdisciplinary learning quality, as measured by their online posts and surveys, respectively. The content analysis results show that overall, students’ online posts could be coded into four interdisciplinary learning dimensions: diversity, disciplinary grounding, cognitive advancement, and integration. The means of the first three dimensions were close to the middle level (ranging from 0.708 to 0.897, and the middle level is 1), whereas the mean score of integration was relatively small (i.e., 0.229). Students under the ChatGPT condition demonstrated improved disciplinary grounding. Regarding their perceived interdisciplinary learning quality, we did not find significant differences across the three conditions in the pre- or post-surveys. The findings underscore ChatGPT’s ability to enhance students’ disciplinary grounding and the significance of further fostering their integration skills.
... ChatGPT has gained attention in the online community and has been discussed extensively [13]. It has been used to help medical students with writing requirements and has seen applications in various fields [14]. Table 2 revealed that both teachers and students exhibit a good level of awareness regarding ChatGPT's capabilities in the domain of creativity and writing. ...
Conference Paper
Full-text available
This research seeks to bridge an existing knowledge gap by investigating the awareness level of ChatGPT among faculty and students in selected universities in the Philippines, with the overarching goal of formulating effective strategies to harness its potential benefits while addressing associated concerns. This study employs a descriptive cross-sectional research survey design that uses simple random sampling techniques and descriptive statistics to evaluate the prevalence and depth of awareness within specific demographic segments. The results indicate a moderate level of awareness across various domains among teachers and students. Notably, teachers exhibit slightly greater awareness of ChatGPT's capabilities, which may extend to other generative AI tools, presenting an opportunity for integrating such technologies into educational curricula to enhance teaching tools and elevate student engagement. In light of these findings, the study recommends that higher education institutions organize educational seminars and workshops to facilitate teachers' and students' effective utilization of AI tools. Furthermore, regular evaluations should be instituted to ensure that students, teachers, and school administrators remain well-informed about emerging AI technologies.
... Terry, 2023). Empirical studies corroborate these accounts, finding students are aware of and largely positive towards using generative AI to support academic tasks (Bonsu & Baffour-Koduah, 2023;Chan & Hu, 2023). Similar results emerged from experimental studies where students were asked to use generative AI for class (Sudirman & Rahmatillah, 2023;Zhu et. al., 2023). Additionally, students recognize the potential for inaccuracies in generative AI outputs and believe they must review the output (Shoufan, 2023). The majority of the literature reviewed was focused on undergraduate students, which Crompton and Burke (2023) found was the trend of studies about students and AI in higher education. ...
Article
Full-text available
Objectives – To understand how many undergraduate and graduate students use generative AI as part of their academic work, how often they use it, and for what tasks they use it. We also sought to identify how trustworthy students find generative AI and how they would feel about a locally maintained generative AI tool. Finally, we explored student interest in trainings related to using generative AI in academic work. This survey will help librarians better understand the rate at which generative AI is being adopted by university students and the need for librarians to incorporate generative AI into their work. Methods – A team of three library staff members and one student intern created, executed, and analyzed a survey of 360 undergraduate and graduate students at Harvard University. The survey was distributed via email lists and at cafes and libraries throughout campus. Data were collected and analyzed using Qualtrics. Results – We found that nearly 65% of respondents have used or plan to use generative AI chatbots for academic work, even though most respondents (65%) do not find their outputs trustworthy enough for academic work. The findings show that students actively use these tools but desire guidance around effectively using them. Conclusion – This research shows students are engaging with generative AI for academic work but do not fully trust the information that it produces. Librarians must be at the forefront of understanding the significant impact this technology will have on information-seeking behaviors and research habits. To effectively support students, librarians must know how to use these tools to advise students on how to critically evaluate AI output and effectively incorporate it into their research.
... This study employed a quasi-experimental design 43,44 because ChatGPT technology is novel and few individuals are familiar with it. ...
Article
Full-text available
Background Artificial Intelligence (AI) applications are widely researched for their potential in effectively improving the healthcare operations and disease management. However, the research trend shows that these applications also have significant negative implications on the service delivery. Purpose To assess the use of ChatGPT for mental health support. Methods Due to the novelty and unfamiliarity of the ChatGPT technology, a quasi-experimental design was chosen for this study. Outpatients from a public hospital were included in the sample. A two-week experiment followed by semi-structured interviews was conducted in which participants used ChatGPT for mental health support. Semi-structured interviews were conducted with 24 individuals with mental health conditions. Results Eight positive factors (psychoeducation, emotional support, goal setting and motivation, referral and resource information, self-assessment and monitoring, cognitive behavioral therapy, crisis interventions, and psychotherapeutic exercises) and four negative factors (ethical and legal considerations, accuracy and reliability, limited assessment capabilities, and cultural and linguistic considerations) were associated with the use of ChatGPT for mental health support. Conclusion It is important to carefully consider the ethical, reliability, accuracy, and legal challenges and develop appropriate strategies to mitigate them in order to ensure safe and effective use of AI-based applications like ChatGPT in mental health support.
... This environment (i.e., non-judgmental and unbiased) is a suitable setting to facilitate open communication and discussion that assist students in respecting and accepting clients. This aligns with a study by [56], which revealed undergraduate students who realized ChatGPT's non-judgmental nature and capability to create responses like humans. Additionally, ChatGPT helps students focus in counseling sessions through its ability to provide procedures, strategies, and guidance to enhance these skills. ...
Article
Full-text available
ChatGPT is an artificial intelligence (AI)-based chatbot that has expanded in popularity among users in several fields. It could transform education and support achieving sustainable development goals. Providing counseling and mental health services is a vital national goal for maintaining sustainable goals. As evidence of innovative technology’s ability to support sustainability in the digital era, no study has been found to investigate ChatGPT in counseling and mental health education based on students’ perceptions and experiences. Consequently, this study aimed to explore undergraduates’ perceptions in terms of the benefits and challenges of using ChatGPT in counseling and mental health education. The study adopted a descriptive quantitative approach using a purposive sample comprising 210 respondents who enrolled in the counseling and mental health program at the University of Jordan and experienced ChatGPT in their learning. The results reported that 81.9 percent of respondents perceived ChatGPT as a beneficial learning tool for counseling and mental health (M = 4.10); they perceived their ability to support most of the primary counseling skills and therapeutic conditions as the counseling dispositions and behaviors expected for reflection on meaning skills, empathic care, setting boundaries with colleagues, openness to feedback, congruence and genuineness, ability to record-keeping and task completion, and non-verbal skills. Furthermore, the findings reported that students experienced moderate challenges and concerns while utilizing ChatGPT (M = 3.21). The study’s findings suggest educators in counseling and mental health employ ChatGPT in the teaching and learning process. In addition, it proposes that decision-makers at the University of Jordan consider student concerns about ChatGPT.
... With their ability to generate information across a wide array of topics, chatbots can promote interdisciplinary learning, allowing students to see connections between different fields of study (Zhu et al., 2023). They can provide relevant information from various disciplines, thereby fostering a holistic approach to learning. ...
Article
Full-text available
As the application of Artificial Intelligence (AI) continues to permeate various sectors, the educational landscape is no exception. Several AI in education (AIEd) applications, like chatbots, present an intriguing array of opportunities and challenges. This paper provides an in-depth exploration of the use and role of AI in education and research, focusing on the benefits (the good) and potential pitfalls (the bad and ugly) associated with the deployment of chatbots and other AIEDs. The opportunities explored include personalised learning, facilitation of administrative tasks, enriched research capabilities, and the provision of a platform for collaboration. These advantages are balanced against potential downsides, such as job displacement, misinformation, plagiarism, and the erosion of human connection. Ethical considerations, particularly concerning data privacy, bias reinforcement, and the digital divide, are also examined. Conclusions drawn from this analysis stress the importance of striking a balance between AI capabilities and human elements in education, as well as developing comprehensive ethical frameworks for AI deployment in educational contexts.
Article
This study aims to analyze a project-based learning (PBL) case utilizing ChatGPT and explore its potential as an instructional support tool. To this end, a PBL course integrating ChatGPT was designed and implemented in the general education course titled ‘Self-Directed Learning Strategies for University Students’ at a university in Seoul during the first semester of 2024. Throughout the course, students used ChatGPT to select presentation topics, conduct research, draft scripts, and deliver their final presentations. To evaluate the effectiveness of ChatGPT integration, a survey was conducted and peer evaluations of project outcomes were analyzed. The findings indicate that students benefited from ChatGPT in various aspects, including organizing presentation content, gathering information, drafting scripts, refining arguments, and generating ideas. Additionally, peer evaluations revealed high ratings for the completeness and coherence of the presentations. These results suggest that ChatGPT can serve as an effective instructional aid by supporting students in the processes of information retrieval and presentation development. This study highlights the potential of ChatGPT to enhance the effectiveness of PBL and offers foundational insights for the design and implementation of AI-assisted educational models.
Article
Full-text available
ChatGPT has sparked both excitement and skepticism in education. To analyze its impact on teaching and learning it is crucial to understand how students perceive ChatGPT and assess its potential and challenges. Toward this, we conducted a two-stage study with senior students in a computer engineering program (n=56). In the first stage, we asked the students to evaluate ChatGPT using their own words after they used it to complete one learning activity. The returned responses (3136 words) were analyzed by coding and theme building (36 codes and 15 themes). In the second stage, we used the derived codes and themes to create a 27-item questionnaire. The students responded to this questionnaire three weeks later after completing other activities with the help of ChatGPT. The results show that the students admire the capabilities of ChatGPT and find it interesting, motivating, and helpful for study and work. They find it easy to use and appreciate its human-like interface that provides well-structured responses and good explanations. However, many students feel that ChatGPT's answers are not always accurate and most of them believe that it requires good background knowledge to work with since it does not replace human intelligence. So, most students think that ChatGPT needs to be improved but are optimistic that this will happen soon. When it comes to the negative impact of ChatGPT on learning, academic integrity, jobs, and life, the students are divided. We conclude that ChatGPT can and should be used for learning. However, students should be aware of its limitations. Educators should try using ChatGPT and guide students on effective prompting techniques and how to assess generated responses. The developers should improve their models to enhance the accuracy of given answers. The study provides insights into the capabilities and limitations of ChatGPT in education and informs future research and development.
Article
Full-text available
Technology-enhanced language learning has exerted positive effects on the performance and engagement of L2 learners. Since the advent of tools based on recent advancement in artificial intelligence (AI), educators have made major strides in applying state-of-the-art technologies to writing classrooms. In November 2022, an AI-powered chatbot named ChatGPT capable of automatic text generation was introduced to the public. The study tried to apply ChatGPT’s text generation feature in a one-week L2 writing practicum. The study adopted a qualitative approach to investigate students’ behaviors and reflections in their exposure to ChatGPT in writing classrooms. The developmental features in learning activities and reflective perceptions were triangulated for the piloting evaluation of the impact of ChatGPT on L2 writing learners. The findings revealed the affordance and potential applicability of the tool in L2 writing pedagogy. Additionally, the tool also showcased an automatic workflow that could maximize the efficiency in composing writing. However, participants generally expressed their concern with its threats to academic honesty and educational equity. The study impelled the reconceptualization of plagiarism in the new era, development of regulatory policies and pedagogical guidance to regulate proper utilization of the tool. Being a pioneering effort, the study accentuated future research directions for more insights into the application of ChatGPT in L2 learning, and the establishment of corresponding pedagogical adjustments.
Article
Full-text available
ChatGPT is an artificial intelligence (AI)-based conversational large language model (LLM). The potential applications of LLMs in health care education, research, and practice could be promising if the associated valid concerns are proactively examined and addressed. The current systematic review aimed to investigate the utility of ChatGPT in health care education, research, and practice and to highlight its potential limitations. Using the PRIMSA guidelines, a systematic search was conducted to retrieve English records in PubMed/MEDLINE and Google Scholar (published research or preprints) that examined ChatGPT in the context of health care education, research, or practice. A total of 60 records were eligible for inclusion. Benefits of ChatGPT were cited in 51/60 (85.0%) records and included: (1) improved scientific writing and enhancing research equity and versatility; (2) utility in health care research (efficient analysis of datasets, code generation, literature reviews, saving time to focus on experimental design, and drug discovery and development); (3) benefits in health care practice (streamlining the workflow, cost saving, documentation, personalized medicine, and improved health literacy); and (4) benefits in health care education including improved personalized learning and the focus on critical thinking and problem-based learning. Concerns regarding ChatGPT use were stated in 58/60 (96.7%) records including ethical, copyright, transparency, and legal issues, the risk of bias, plagiarism, lack of originality, inaccurate content with risk of hallucination, limited knowledge, incorrect citations, cybersecurity issues, and risk of infodemics. The promising applications of ChatGPT can induce paradigm shifts in health care education, research, and practice. However, the embrace of this AI chatbot should be conducted with extreme caution considering its potential limitations. As it currently stands, ChatGPT does not qualify to be listed as an author in scientific articles unless the ICMJE/COPE guidelines are revised or amended. An initiative involving all stakeholders in health care education, research, and practice is urgently needed. This will help to set a code of ethics to guide the responsible use of ChatGPT among other LLMs in health care and academia.
Article
Full-text available
Motivation drives most human activities, including foreign language learning, and it is affected by several intrinsic and extrinsic factors. This study investigated how ChatGPT impacts learning English. A quantitative research design was used to collect data from 80 teachers and students who had access to the ChatGPT in its very beginning phase in early 2023. The sample, selected in a non-probability sampling technique, responded to an online questionnaire. Findings showed that ChatGPT generally motivates learners to develop reading and writing skills. The respondents had neutral attitudes towards the effect of ChatGPT on developing listening and speaking skills. The findings suggest that ChatGPT-based teaching is motivational. ChatGPT should be used as a learning tool instead of fearing its negative impacts, which require further detailed investigations. Further research is required to explore more advantages of ChatGPT for other aspects of the language and illuminate its negative effects to help educators reduce them in English programs.
Article
Full-text available
Artificial Intelligence (AI) technologies have been progressing constantly and being more visible in different aspects of our lives. One recent phenomenon is ChatGPT, a chatbot with a conversational artificial intelligence interface that was developed by OpenAI. As one of the most advanced artificial intelligence applications, ChatGPT has drawn much public attention across the globe. In this regard, this study examines ChatGPT in education, among early adopters, through a qualitative instrumental case study. Conducted in three stages, the first stage of the study reveals that the public discourse in social media is generally positive and there is enthusiasm regarding its use in educational settings. However, there are also voices who are approaching cautiously using ChatGPT in educational settings. The second stage of the study examines the case of ChatGPT through lenses of educational transformation, response quality, usefulness, personality and emotion, and ethics. In the third and final stage of the study, the investigation of user experiences through ten educational scenarios revealed various issues, including cheating, honesty and truthfulness of ChatGPT, privacy misleading, and manipulation. The findings of this study provide several research directions that should be considered to ensure a safe and responsible adoption of chatbots, specifically ChatGPT, in education.
Article
Full-text available
We evaluated the performance of a large language model called ChatGPT on the United States Medical Licensing Exam (USMLE), which consists of three exams: Step 1, Step 2CK, and Step 3. ChatGPT performed at or near the passing threshold for all three exams without any specialized training or reinforcement. Additionally, ChatGPT demonstrated a high level of concordance and insight in its explanations. These results suggest that large language models may have the potential to assist with medical education, and potentially, clinical decision-making.
Article
The ability of children to ask curiosity-driven questions is an important skill that helps improve their learning. For this reason, previous research has explored designing specific exercises to train this skill. Several of these studies relied on providing semantic and linguistic cues to train them to ask more of such questions (also called divergent questions). But despite showing pedagogical efficiency, this method is still limited as it relies on generating the said cues by hand, which can be a very long and costly process. In this context, we propose to leverage advances in the natural language processing field (NLP) and investigate the efficiency of using a large language model (LLM) for automating the production of key parts of pedagogical content within a curious question-asking (QA) training. We study generating the said content using the "prompt-based" method that consists of explaining the task to the LLM in natural text. We evaluate the output using human experts annotations and comparisons with hand-generated content. Results suggested indeed the relevance and usefulness of this content. We then conduct a field study in primary school (75 children aged 9–10), where we evaluate children’s QA performance when having this training. We compare 3 types of content: 1) hand-generated content that proposes "closed" cues leading to predefined questions; 2) GPT-3-generated content that proposes the same type of cues; 3) GPT-3-generated content that proposes "open" cues leading to several possible questions. Children were assigned to either one of these groups. Based on human annotations of the questions generated, we see a similar QA performance between the two "closed" trainings (showing the scalability of the approach using GPT-3), and a better one for participants with the "open" training. These results suggest the efficiency of using LLMs to support children in generating more curious questions, using a natural language prompting approach that affords usability by teachers and other users not specialists of AI techniques. Furthermore, results also show that open-ended content may be more suitable for training curious question-asking skills.
Article
We present a holistic approach to building a robust and useful natural language classification system for real-world content moderation. The success of such a system relies on a chain of carefully designed and executed steps, including the design of content taxonomies and labeling instructions, data quality control, an active learning pipeline to capture rare events, and a variety of methods to make the model robust and to avoid overfitting. Our moderation system is trained to detect a broad set of categories of undesired content, including sexual content, hateful content, violence, self-harm, and harassment. This approach generalizes to a wide range of different content taxonomies and can be used to create high-quality content classifiers that outperform off-the-shelf models.
Article
In less than 2 months, the artificial intelligence (AI) program ChatGPT has become a cultural sensation. It is freely accessible through a web portal created by the tool's developer, OpenAI. The program-which automatically creates text based on written prompts-is so popular that it's likely to be "at capacity right now" if you attempt to use it. When you do get through, ChatGPT provides endless entertainment. I asked it to rewrite the first scene of the classic American play Death of a Salesman, but to feature Princess Elsa from the animated movie Frozen as the main character instead of Willy Loman. The output was an amusing conversation in which Elsa-who has come home from a tough day of selling-is told by her son Happy, "Come on, Mom. You're Elsa from Frozen. You have ice powers and you're a queen. You're unstoppable." Mash-ups like this are certainly fun, but there are serious implications for generative AI programs like ChatGPT in science and academia.