ArticlePDF Available

Chatbots for future docs: exploring medical students’ attitudes and knowledge towards artificial intelligence and medical chatbots

Taylor & Francis
Medical Education Online
Authors:

Abstract and Figures

Artificial intelligence (AI) in medicine and digital assistance systems such as chatbots will play an increasingly important role in future doctor – patient communication. To benefit from the potential of this technical innovation and ensure optimal patient care, future physicians should be equipped with the appropriate skills. Accordingly, a suitable place for the management and adaptation of digital assistance systems must be found in the medical education curriculum. To determine the existing levels of knowledge of medical students about AI chatbots in particular in the healthcare setting, this study surveyed medical students of the University of Luebeck and the University Hospital of Tuebingen. Using standardized quantitative questionnaires and qualitative analysis of group discussions, the attitudes of medical students toward AI and chatbots in medicine were investigated. From this, relevant requirements for the future integration of AI into the medical curriculum could be identified. The aim was to establish a basic understanding of the opportunities, limitations, and risks, as well as potential areas of application of the technology. The participants (N = 12) were able to develop an understanding of how AI and chatbots will affect their future daily work. Although basic attitudes toward the use of AI were positive, the students also expressed concerns. There were high levels of agreement regarding the use of AI in administrative settings (83.3%) and research with health-related data (91.7%). However, participants expressed concerns that data protection may be insufficiently guaranteed (33.3%) and that they might be increasingly monitored at work in the future (58.3%). The evaluations indicated that future physicians want to engage more intensively with AI in medicine. In view of future developments, AI and data competencies should be taught in a structured way during the medical curriculum and integrated into curricular teaching.
This content is subject to copyright. Terms and conditions apply.
RESEARCH ARTICLE
Chatbots for future docs: exploring medical students’ attitudes and
knowledge towards artificial intelligence and medical chatbots
Julia-Astrid Moldt
a
, Teresa Festl-Wietek
a
, Amir Madany Mamlouk
b
, Kay Nieselt
c
, Wolfgang Fuhl
c
and Anne Herrmann-Werner
a,d
a
University of Tuebingen, Tuebingen, Germany;
b
Institute for Neuro- and Bioinformatics, University of Luebeck, Luebeck, Germany;
c
Institute for Bioinformatics and Medical Informatics, University of Tuebingen, Germany;
d
Department of Internal Medicine VI/
Psychosomatic Medicine and Psychotherapy, University Hospital Tuebingen, Tuebingen, Germany
ABSTRACT
Artificial intelligence (AI) in medicine and digital assistance systems such as chatbots will play an
increasingly important role in future doctor patient communication. To benefit from the
potential of this technical innovation and ensure optimal patient care, future physicians should
be equipped with the appropriate skills. Accordingly, a suitable place for the management and
adaptation of digital assistance systems must be found in the medical education curriculum. To
determine the existing levels of knowledge of medical students about AI chatbots in particular in
the healthcare setting, this study surveyed medical students of the University of Luebeck and the
University Hospital of Tuebingen. Using standardized quantitative questionnaires and qualitative
analysis of group discussions, the attitudes of medical students toward AI and chatbots in
medicine were investigated. From this, relevant requirements for the future integration of AI
into the medical curriculum could be identified. The aim was to establish a basic understanding of
the opportunities, limitations, and risks, as well as potential areas of application of the technology.
The participants (N = 12) were able to develop an understanding of how AI and chatbots will
affect their future daily work. Although basic attitudes toward the use of AI were positive, the
students also expressed concerns. There were high levels of agreement regarding the use of AI in
administrative settings (83.3%) and research with health-related data (91.7%). However, partici-
pants expressed concerns that data protection may be insufficiently guaranteed (33.3%) and that
they might be increasingly monitored at work in the future (58.3%). The evaluations indicated
that future physicians want to engage more intensively with AI in medicine. In view of future
developments, AI and data competencies should be taught in a structured way during the
medical curriculum and integrated into curricular teaching.
ARTICLE HISTORY
Received 15 December 2022
Revised 6 February 2023
Accepted 16 February 2023
KEYWORDS
Medical students; artificial
intelligence; applications in
education;; human-
computer interface;
teaching/learning strategies;
chatbot
Introduction
The healthcare system is undergoing a digital transfor-
mation, and artificial intelligence (AI) will play
a significant role in defining everyday medical practice
[1]. The location- and time-independence of digital appli-
cations have created new opportunities for medicine and
health communication that are also changing the doctor –
patient relationship [2]. The growing importance of
e-health applications, wearables and AI applications
such as chatbots can empower patients to collect their
own health data [3,4].
Furthermore, the digital networking of patients,
hospitals, physicians and other healthcare services is
enabling a shift from a physician-centric approach to
more patient-centered treatment [5]. To exploit the
potential of this technical innovation and ensure
optimized care for patients, future doctors must be
equipped with the appropriate skills [6]. Future phy-
sicians will not only need to be flexible in responding
to different healthcare contexts but will also require
the competence to adequately deal with procedures
and applications involving AI and the accompanying
big data [7]. The growing complexity of medicine and
increasing specialization of knowledge require the
integration of AI as well as the interaction with digital
assistance systems already in the curriculum of med-
ical studies [8–10]. According to current literature,
although AI competencies are essential for medical
practice, they are not comprehensively taught in
medical education [7,11,12].
Medical curriculum in Germany
A look at the national competence-based learning
objectives catalog for medicine (NKLM) [13] shows
that the teaching of competencies in the area of
medical apps and artificial intelligence is still under-
represented. The national competence-based learning
objectives catalog for medicine is currently being
further developed on the basis of the ‘Master Plan
CONTACT Julia-Astrid Moldt julia-astrid.moldt@med.uni-tuebingen.de TIME Tübingen Institute for Medical Education, Elfriede-Aulhorn-
Straße 10, 72076, Tuebingen, Germany
MEDICAL EDUCATION ONLINE
2023, VOL. 28, 2182659
https://doi.org/10.1080/10872981.2023.2182659
© 2023 The Author(s). Published by Informa UK Limited, trading as Taylor & Francis Group.
This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/), which permits
unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
for Medical Studies 2020’ [14]. It is to be made
compulsory at all medical faculties when the new
Medical Licensing Regulations come into force. The
focus is on the question of which competencies junior
doctors should acquire as part of the core curriculum
of their medical studies, including medical commu-
nication skills, interprofessional teamwork, scientific
work and digital competencies [15]. The importance
of integrating AI into medical studies is already being
explored extensively in research and literature, but
curricular implementation [16,17] is in its early
stages. The emerge of medical chatbots could have
the potential to improve the efficiency, quality and
accessibility of healthcare services by providing quick
information and connecting patients to healthcare
providers and will become increasingly important
for both doctors and patients during the treatment
process. However, there are also raising concerns
about chatbots potentially eroding diagnostic prac-
tice, being driven by business logic, and having lim-
itations and incompleteness that may harm patients
in emergency situations [18,19]. Understanding how
these technologies work and how they can be used
can help medical students better understand the
broader landscape of healthcare and be better pre-
pared for the future of medicine [20]. As far as the
authors know, the experience and attitude towards
the use of chatbots in medicine among medical stu-
dents has not yet been recorded in Germany.
Therefore, the question arose for us, how to suc-
cessfully prepare medical professionals for this new
scenario during their medical training. This study is
intended as basic research relating to AI in the med-
ical curriculum. We aimed to provide an overview of
the current knowledge and attitudes of medical stu-
dents regarding AI technologies to determine what is
necessary for future implementation in the curricu-
lum. Therefore, the following questions arose regard-
ing the implementation of AI in medical education:
What attitudes do medical students have about
artificial intelligence and the use of chatbots in
medicine?
Does their view of chatbots change after learn-
ing more about medical chatbots in the course?
Material and methods
Study design and participants
This study followed the Standards for Reporting
Qualitative Research [21] and presents a mixed-method
study as part of a project supported by the Federal
Ministry of Education and Research, Germany (BMBF).
It was carried out at the Medical Faculty of Tuebingen
with the support of the Institute for Neuro- and
Bioinformatics, University of Luebeck. A hybrid course
named ‘Chatbots for Future Docs’ was developed for
medical students of all semesters and was offered as an
elective course between January and March 2022. N = 12
medical students learned about conditions of doctor
patient communication in general, possible uses of chat-
bots in healthcare, the ethical framework of AI, how
chatbots work in general and what it takes to create
a ready-to-go bot. The course combined classical teaching
of theoretical knowledge (represented by asynchronous
digital short inputs) and practical competence acquisition
in the form of exercises and group work during
a synchronous online format via Zoom.
The attitudes of participants regarding the topic of
AI in medicine were quantitatively assessed by means
of a standardized questionnaire before and after the
course via the learning management software ILIAS.
Qualitative methods such as group discussions sup-
plemented the detailed, subjective and individual atti-
tudes of the medical students to the topic.
Ethics
The study received ethical approval from the Ethics
Committee of Tuebingen Medical Faculty (no. 824/
2021BO2). Participation was voluntary, and students
provided their informed consent and received no com-
pensation. All responses and data were kept anonymous.
Measurement and process of study
We combined quantitative and qualitative research
methods in this study to achieve an in-depth analysis
of the collected data [22].
To conduct the quantitative survey, the medical stu-
dents were asked to complete a self-developed standar-
dized questionnaire at the beginning of the course to
assess subjective attitudes toward AI in medicine. The
questionnaire was divided into three subsections, with
items including five-point Likert scales and multiple-
choice questions. The first subsection aimed to gather
demographic data including age, gender and general
attitudes toward technology and AI. Attitudes toward
technology use were assessed using a shortened version
of the validated short scale for assessing individual
differences in the willingness of technology, as per
Neyer et al. [23] (technology commitment).
The second subsection examined the openness of med-
ical students regarding the future use of AI in healthcare
in particular, which in the future may be able to answer
health-related questions, perform certain tests or exam-
inations, diagnose health conditions and suggest and
apply treatments. The third subsection addressed the
attitudes of the medical students toward chatbots in
medicine. Students were asked to revisit two questions
from the questionnaire on chatbots in medicine after
the course, as this was the main interest of the course.
The material for qualitative content analysis was
based on group work conducted by the students,
2J.-A. MOLDT ET AL.
which involved discussing questions about the use of
mental health chatbots in medicine (see A1, appendix).
Data analysis
Quantitation
Statistical analysis of the questionnaire was per-
formed using IBM SPSS version 27. Mean values
(M), standard deviations (SD), sum values, frequen-
cies and percentages of the relevant items were
obtained. The Pearson correlation coefficient was
calculated to capture linear relationships between
the relevant variables. We used the Wilcoxon test to
analyze the statistical significance of the differences
between the two repeated items after the course; the
level of significance was p < .05.As one student did
not complete the course, we used the mean imputa-
tion procedure for the missing values after testing for
normal distribution using the Kolmogorov-Smirnov
test procedure [24,25].
Qualitative analysis
The results of the students’ group work were
recorded, transcribed and coded by three different
authors (JAM, TFW, AM). Following discussions in
regular meetings, findings were summarized and
a category system consisting of main and subcate-
gories (according to Mayring’s qualitative content
analysis) was agreed upon [26]. Selected text passages
were used as quotations to illustrate each category
[26]. Inductive category formation was performed to
reduce the content of the material to its essentials
(bottom-up process). In the first step, the goal of the
analysis and the theoretical background were deter-
mined. From this, questions about chatbots in med-
icine were developed and presented to the students
for discussion in a group exercise. Two main topics
were identified, namely positive and negative atti-
tudes toward chatbots in medicine. In the second
step, we worked through the individual statements
of the students systematically and derived various
categories: user group, technical implementation,
acceptance, and use in medicine.
Results
The course was attended by 12 medical students from
the clinical and preclinical study sections, who were
able (at least partly) to give a broad explanation of AI.
For further information, please see Table 1.
Main results questionnaire
Part 1: general attitudes toward technology, and
socio demography
The participating students stated that they could
explain the main features and applications of AI
(50.0%) and that they could at least give a broad
definition of AI (50.0%). Evaluation of the technol-
ogy commitment according to Neyer et al. 2016
showed that the study participants displayed
a moderate willingness to use technology (M = 21;
SD = 3.1, min. 18, max. 31).
Part 2: attitudes toward AI in medicine
Many students were not afraid of the future
changes that AI could bring. The majority (83.3%)
were not afraid of losing their jobs because of AI or
of being overwhelmed by using AI (83.3%).
However, some students agreed that they were
afraid of increasing surveillance at their future
workplaces (58.3% agreement vs. 25.0% disagree-
ment) and decreasing transparency regarding the
use of personal data (33.3% agreement vs. 33.3%
disagreement) (Figure 1 Plot A).
As shown in Figure 1 Plot B, the majority of
students supported the use of AI in various areas of
medicine. Nevertheless, there was less support for
its use in human-centered areas such as therapy or
follow-up treatment, while the use of AI in more
technical areas such as diagnostics or administra-
tive tasks was seen as mostly beneficial.
Students saw the use of AI in medicine primar-
ily as an opportunity to reduce the administrative
burden on physicians. At the same time, however,
they also believed that the new technology would
be able to make diagnoses faster and more accu-
rate in the future and ultimately also make access
to medical advice more sustainable (Figure 2
Plot A).
Part 3: attitudes toward chatbots in medicine
At the beginning of the course 692% of the stu-
dents never have used a medical chatbot before.
However, the students had great confidence in
chatbots, especially for organizational tasks. In
task areas with high levels of responsibility, such
as diagnosis or treatment suggestions, opinions
tended to diverge and were more critical, but the
students were not completely opposed to the use of
AI in general (Figure 2 Plot B).
Table 1. Students’ previous knowledge about AI and demo-
graphic data.
Yes, I could explain the main
features and applications of AI
6/12
(50.0%)
I could give a general
explanation, but I don’t know
anything more specific than
that
6/12
(50.0%)
Age Mean
24.8
Standard
deviation 2.9
Min/Max
21/29
Gender Male
5 (41.7%)
Female
6 (50.0%)
Other
1 (8.3%)
MEDICAL EDUCATION ONLINE 3
As shown in Tables A2 and A3 in the appendix,
no major differences in attitudes held before and
after the course were detected. However, students
were significantly (p = 0.02) less critical about priv-
acy and the use of medical chatbots after the
course. Prior to the course, 500% of students
believed that data privacy and security could not
be fully guaranteed, 41.7% were undecided, and
only 8.3% did not believe this. After the course,
only 36.4% were still critical of data protection,
18.2% were undecided and 45.5% did not believe
that data protection cannot be guaranteed with
Figure 1. Attitudes of medical students toward AI in medicine (fears about AI in various areas of medicine).
4J.-A. MOLDT ET AL.
med. Chatbots (see Table A3). Only a few state-
ments showed a clear positive or negative tendency.
For example, the majority agreed that using chat-
bots saved time (83.3% agreement pre; 72.7% agree-
ment post) and money (75.0% agreement pre;
54.5% post). However, medical students felt that
chatbots were not yet sufficiently established and
that long-term success had yet to materialize
(81.8%). There was also a fear of communication
problems (81.8%) and loss of personal contact with
Figure 2. Attitudes of medical students toward AI in medicine (statements about the use of AI and chatbots).
MEDICAL EDUCATION ONLINE 5
patients (63.6%) because of the lack of maturity of
the technology.
Findings of the qualitative content analysis
The main findings of the present study concern the
students’ views on chatbots in medicine. Four main
themes/categories were identified: user group, technical
implementation, acceptance, and use in medicine
(Table 2).
Theme 1: user group
The medical students were positive that chatbots are
accessible to a broad user group as a result of their
time- and location-independent availability. They
also believed that possible language barriers or other
hurdles could also be overcome. Nevertheless, they
were concerned that certain groups of people (elderly
people, visually impaired people) may not be familiar
with modern technology or chatbots. The medical
students also proposed a minimum age of 18 years
and an alternative language mode specifically aimed
at children.
Theme 2: technical implementation
The students felt that with the use of emojis, artificial
delays, small talk and customization, there was an
opportunity to provide the most human-like and
realistic conversation possible with a chatbot.
Furthermore, the development of the chatbot should
take into account the context in which it would be
used (administrative or personal assistant). However,
according to the students’ evaluations, technical
implementation was not sufficiently mature or flex-
ible because of the limited and rigid algorithms used.
Theme 3: acceptance and trust
The use of chatbots as therapy tools or interactive
diaries for patients was considered by the students to
be a good opportunity to reduce anonymity, shyness
and shame about disclosing personal and painful
information. However, it was thought that doctor
patient communication could deteriorate as a result
of non-verbal communication with a chatbot and
interpersonal information may get lost. Thus, the
first personal conversation between doctor and
patient was considered indispensable.
Use in medicine
The students perceived the use of therapy chatbots in
medicine as potentially supportive and complemen-
tary diagnostic tools that could be used indepen-
dently by patients but did not understand them as
a substitute for therapy. As an example, the students
mentioned the possibility of technical errors or the
risk of increasing social withdrawal in depressed per-
sons due to the lack of personal contact. Therefore,
human contact and empathy were considered essen-
tial for the success of therapy, which could be
reduced by the use of chatbots.
Teaching evaluation
The participants enjoyed gaining a fundamental
insight into the topic of chatbots and AI, which was
completely new for some of them. In addition, the
interactive design and hybrid format contributed to
the positive evaluation of the course. In general, the
students gained a new perspective on chatbots and
their associated problems, as well as a basic under-
standing of how chatbots function, where they can be
used and the effort involved in implementing them.
Table 2. Overview of categories, subcategories and corresponding examples based on qualitative analysis.
category Subcategory Example
User group Positive
Attitudes
Enabling of individual chat selection (in case someone needs small talk or similar)
Possible individual language selection
Target group-oriented language possible
Negative
Attitudes
Specific user groups are not confident with modern techniques and chatbots
Technical
implementation
Positive
Attitudes
Communication tools, e.g., emojis and artificial delays could be helpful to simulate human-like con-
versation
Negative
Attitudes
By now, algorithms are too strict to enable flexible and individualized answer design
Smalltalk is not expedient
Acceptance and
trust
Positive
Attitudes
Anonymity reduces reluctance and shame when disclosing personal and hurtful information
Negative
Attitudes
Skepticism regarding the ability of the intermediary instance chatbot to support physicians as this could
limit the direct doctor–patient relationship
Informational gain is questionable due to lost information from nonverbal communication
Initial personal talk remains indispensable
Data security
Use in medicine Positive
Attitudes
Reflecting through chatting as a therapeutic tool but not as a substitute for therapy
Interactive diary
Negative
Attitudes
Critical toward chatbots as a therapy tool
Human contact and empathy are crucial for therapeutic success
6J.-A. MOLDT ET AL.
Discussion
In the context of digitalization in healthcare, applica-
tions that use AI are becoming increasingly impor-
tant [12]. This study contributed to the
understanding of the current level of knowledge and
attitudes regarding AI, particularly chatbots, in
medicine.
Medical chatbots are an example of how artificial
intelligence and technology are being integrated into
healthcare. As such, it would be beneficial for medical
students in Germany to learn about this technology
as part of their medical curriculum. This will help
ensure that they are well-equipped to work with and
utilize medical chatbots in their future practice, in
order to provide high-quality and efficient care to
patients. Additionally, learning about medical chat-
bots would provide medical students with a better
understanding of the role that technology plays in
healthcare and how it is likely to continue to shape
the medical profession in the future.
This study revealed that the attitudes and expecta-
tions of medical students were generally optimistic
about the use of AI and chatbots in relation to
a variety of purposes in medicine, including in areas
such as administration, research and diagnostic ima-
ging techniques (Figure 1, Plot B). In particular, the
students would trust the chatbots to perform recur-
ring and supportive activities, such as answering sim-
ple questions, arranging appointments, and providing
basic information. The majority were certain that
they would not be replaced by AI in their jobs or
that they would be less valued in their roles as future
doctors, as other authors have stated in similar stu-
dies [27]. However, they were more critical of the use
of these new technologies in core tasks, such as carry-
ing out personal counseling and specific medical
examinations. The participants also expressed con-
cerns that data protection and privacy may no longer
be adequately guaranteed. Finally, there was a fear
that personal contact with patients could be lost if the
patients were increasingly engaging with technical
systems rather than human personnel.
As other studies have demonstrated, it cannot be
assumed that the generation of people who have
grown up with digital technologies and are proficient
in their use (similar to our cohort) are also aware of
all the options and ethical consequences of the use of
new technologies in their professional field. However,
this is not synonymous with the simultaneous devel-
opment of digital competencies in the professional
field [28,29]. The areas in which AI can be applied
in medicine are diverse and, with the development of
smartphone apps, have reached not only the health-
care system but also the private sphere and there
will likely be more in the future [30]. Accordingly, to
remain empowered, future physicians must be able to
understand how AI works, as well as how to interpret
results, in order to meaningfully support patients
with digital tools at the same time as critically mon-
itoring AI [31]. Digital learning opportunities and the
development of AI skills are essential in medical
education and may help to meet the vast need for
qualifications [32]. We found a great deal of uncer-
tainty and skepticism regarding chatbots due to the
lack of integration of AI topics into the medical
curriculum (as yet), as well as a lack of knowledge
about the basic conditions and legal and ethical
requirements of AI use [33], reflecting findings
from other studies [34–36].
While students saw a great deal of potential for the
use of chatbots in medicine, they had many concerns
about using them in areas that went beyond organi-
zational activities such as making patient appoint-
ments. Above all, they believed that the technology
was not yet sufficiently developed and that in the
context of healthcare, patients needed a human coun-
terpart. Neutral attitudes to chatbots were also evi-
dent from many statements in the questionnaire,
which confirmed the thesis that there is not yet
enough knowledge about the topic of AI in medicine
for the study participants to have developed distinct
opinions. In the questions about chatbots, in particu-
lar, which were repeated after the course, we could
not identify any significant changes in attitudes.
However, our study was able to give medical stu-
dents, as non-computer scientists, a good initial over-
view of how a chatbot works, a basic understanding
of how much data needs to be provided for an AI,
and also potential future uses of AI in medicine and
medical chatbots.
The perceived challenges and concerns of students
relating to the use of AI and chatbots in healthcare
must be addressed and taken seriously before future
physicians are exposed to such tools [37]. AI is still
underrepresented in the medical curriculum, and
students lack the opportunity to engage more inten-
sively with the topic of AI and develop the required
expertise [11,38,39]. Therefore, for us, it was impor-
tant not only to teach digital competencies and
knowledge about AI theoretically but also to incor-
porate them practically into the teaching unit. Thus,
digitization was both included in the teaching and
incorporated as a learning objective.
Limitations
Although the students were very interested in the
topic of chatbots in medicine, and the topic of AI is
also gaining increasing importance in medicine in
general, the results of our study are limited in terms
of representing the perspectives of the student popu-
lation due to the small number of participants. Also,
MEDICAL EDUCATION ONLINE 7
the course duration of three months was too short for
sustainable changes in viewpoints or information
gain to occur. However, the goal of this work was
not to draw representative conclusions for all
German medical students, but rather to understand
the state of knowledge and perceptions of medical
students regarding AI and chatbots in medicine,
which we believe was achieved with our sample.
The next step will be to investigate which AI compe-
tencies should be included in the curriculum.
Conclusion
This study indicated that future physicians in
Germany are willing to engage more intensively
with AI in medicine. In our study, they were able to
develop a basic understanding of how AI and chat-
bots will affect their future daily work. Although their
basic attitude toward the use of clinical AI was posi-
tive, medical students also had concerns, especially
with regard to the lack of data protection and declin-
ing personal contact with patients. With a view to
future developments in the workplace, we can only
emphasize once again how urgently medical curricula
need to be supplemented with these new core com-
petencies so that doctors can help to shape the tech-
nological course of patient treatment in an informed
and self-confident manner.
Acknowledgments
We would like to thank our study assistants Marie
Thiwissen and Felix Pfeiffer for their help.
Disclosure statement
No potential conflict of interest was reported by the
author(s).
Funding
We acknowledge support via financing publication fees
from ‘Deutsche Forschungsgemeinschaft’ and the Open
Access Publishing Fund of the University of Tuebingen.
We also thank the Federal Ministry of Education and
Research, Germany (BMBF) for supporting this project
(16DHBQP041/16DHBQP042).
Ethical approval
The study received ethical approval from the Ethics
Committee of Tuebingen Medical Faculty (no. 824/
2021BO2). Participation was voluntary, and students pro-
vided their written informed consent. All responses and
data were kept anonymous.
Contributorship
JAM was responsible for the study design and implementa-
tion and data collection. AMM was responsible for the tech-
nical information of the chatbot, AHW for the didactical
design of the teaching course. JAM analysed the research
material and conducted the writing for the manuscript.
TFW participated in data analysis and interpretation and
critically revised the manuscript. AHW and AMM made
significant contributions to the study design and critically
revised the manuscript. KN and WF revised the manuscript
and have given final approval of the version to be published.
All authors approved the final version of the manuscript and
agreed to be responsible for all aspects found therein.
ORCID
Julia-Astrid Moldt http://orcid.org/0000-0002-2418-
150X
Teresa Festl-Wietek http://orcid.org/0000-0003-1450-
1757
Amir Madany Mamlouk http://orcid.org/0000-0001-
9709-1620
Kay Nieselt http://orcid.org/0000-0002-1283-7065
Wolfgang Fuhl http://orcid.org/0000-0001-7128-298X
Anne Herrmann-Werner http://orcid.org/0000-0003-
2413-7047
References
[1] Liebrich F. Digitale Medienprodukte in Der Arzt-
Patienten-Kommunikation. Wiesbaden: Springer
Fachmedien; 2017.
[2] Kundu S. How will artificial intelligence change med-
ical training? Communications Medicine. 2021;1(1).
[3] Lin B, Wu S. Digital transformation in personalized
medicine with artificial intelligence and the internet of
medical things. OMICS. 2022 Feb;26(2):77–81.
PubMed PMID: 33887155.
[4] Bates M. Health care chatbots are here to help. IEEE
Pulse. 2019 May-Jun;10(3):12–14. PubMed PMID:
31135345.
[5] Rajpurkar P, Chen E, Banerjee O, et al. AI in health
and medicine. Nat Med. 2022 Jan;28(1):31–38.
PubMed PMID: 35058619.
[6] Ng KH, Wong JHD. A clarion call to introduce arti-
ficial intelligence (AI) in postgraduate medical physics
curriculum. Phys Eng Sci Med. 2022 Mar;45(1):1–2.
PubMed PMID: 35006576.
[7] Han ER, Yeo S, Kim MJ, et al. Medical education
trends for future physicians in the era of advanced
technology and artificial intelligence: an integrative
review. BMC Med Educ. 2019 Dec 11;19(1):460.
PubMed PMID: 31829208; PubMed Central PMCID:
PMCPMC6907217.
[8] Moldt J-A, Festl-Wietek T, Mamlouk AM, et al.
Assessing medical students’ perceived stress levels by
comparing a chatbot-based approach to the Perceived
Stress Questionnaire (PSQ20) in a mixed-methods
study. Digit Health. 2022;8:20552076221139092.
[9] Wiljer D, Hakim Z. Developing an artificial
intelligence-enabled health care practice: rewiring
health care professions for better care. J Med
Imaging Radiat Sci. 2019 Dec;50(4 Suppl 2):S8–14.
PubMed PMID: 31791914.
8J.-A. MOLDT ET AL.
[10] Wartman SA, Combs CD. Reimagining medical edu-
cation in the age of AI. AMA J Ethics. 2019 Feb 1;21
(2):E146–152. PubMed PMID: 30794124.
[11] Chan KS, Zary N. Applications and challenges of
implementing artificial intelligence in medical educa-
tion: integrative review. JMIR Med Educ. 2019 Jun
15;5(1):e13930. PubMed PMID: 31199295; PubMed
Central PMCID: PMCPMC6598417.
[12] Wood EA, Ange BL, Miller DD. Are we ready to
integrate artificial intelligence literacy into medical
school curriculum: students and faculty survey.
J Med Educ Curric Dev. 2021 Jan-Dec
;8:23821205211024078. PubMed PMID: 34250242;
PubMed Central PMCID: PMCPMC8239949.
[13] Fischer MR, Bauer D, Mohn K, et al. Finally finished!
national competence based catalogues of learning
objectives for undergraduate medical education
(NKLM) and dental education (NKLZ) ready for
trial. GMS Z Med Ausbild. 2015;32(3):Doc35.
PubMed PMID: 26677513; PubMed Central PMCID:
PMCPMC4580444.
[14] Erschens R, Herrmann-Werner A, Schaffland TF,
et al. Association of professional pre-qualifications,
study success in medical school and the eligibility
for becoming a physician: a scoping review. PLoS
ONE. 2021;16(11):e0258941. (PubMed PMID:
34762678; PubMed Central PMCID:
PMCPMC8584759. eng).
[15] Plange N, Feltgen N, Arbeitsgemeinschaft L. The
“Nationaler Kompetenzbasierter Lernzielkatalog
Medizin NKLM 2.0”-Implications for medical educa-
tion in ophthalmology. Ophthalmologie. 2022 Nov 21.
PubMed PMID: 36409352.
[16] Kuhn S. Transformation durch Bildung. Deutsches
Ärzteblatt. 2018;2018:115.
[17] Nagy M, Radakovich N, Nazha A. Why machine
learning should be taught in medical schools.Med Sci
Educator. 2022 Apr 1];32(2):529–532.
[18] Nadarzynski T, Miles O, Cowie A, et al. Acceptability
of artificial intelligence (AI)-led chatbot services in
healthcare: a mixed-methods study. Digit Health.
2019 Jan-Dec;5:2055207619871808. PubMed PMID:
31467682; PubMed Central PMCID:
PMCPMC6704417.
[19] Parviainen J, Rantala J. Chatbot breakthrough in the
2020s? an ethical reflection on the trend of automated
consultations in health care.Medicine, Health Care
and Philosophy. 2022 Mar 1;25(1):61–71.
[20] Pryss R, Kraft R, Baumeister H, et al. Using chatbots
to support medical and psychological treatment pro-
cedures challenges, opportunities, technologies, refer-
ence architecture. In: Montag C Baumeister Heditors.
Digital Phenotyping and Mobile Sensing. Studies in
Neuroscience, Psychology and Behavioral Economics.
Cham: Springer International Publishing; 2023. p.
335–346.
[21] O’brien BC, Harris IB, Beckman TJ, et al. Standards
for reporting qualitative research: a synthesis of
recommendations. Acad Med. 2014 Sep;89
(9):1245–1251. PubMed PMID: 24979285; eng.
[22] Kaplan B, Duchon D. combining qualitative and
quantitative methods in information systems research:
a case study. MIS Quarterly. 1988;12(4):571–586.
[23] Neyer F, Felber J, Gebhardt C. Kurzskala zur
Erfassung von Technikbereitschaft [Short scale for
recording technology willingness]. In: GESIS -
Leibniz Institute for Social Sciences.
Zusammenstellung sozialwissenschaftlicher Items
und Skalen (ZIS); 2016.
[24] Baur N, Blasius J. Handbuch Methoden der empiri-
schen Sozialforschung. 1 ed. Wiesbaden: Springer VS;
2014.
[25] Göthlich S. E. Zum Umgang mit fehlenden Daten in
großzahligen empirischen Erhebungen. Albers, S.,
Klapper, D., Konradt, U., Walter, A., Wolf, J., eds.
In: Methodik der empirischen Forschung.
Wiesbaden: Gabler Verlag; 2007. p. 119–134.
[26] Mayring P. Qualitative Inhaltsanalyse. Grundlagen
und Techniken. Weinheim: Deutscher Studien
Verlag; 2000.
[27] Dos Santos D P, Giese D, Brodehl S, et al. Medical
students’ attitude towards artificial intelligence:
a multicentre survey. Eur Radiol. 2019 Apr;29
(4):1640–1646. PubMed PMID: 29980928.
[28] Kirschner PA, De Bruyckere P. The myths of the
digital native and the multitasker. In Teaching and
Teacher EducationVol. 67 p. 135–142; 2017 Oct.
PubMed PMID: WOS:000411543800013; English.
[29] Kuhn S, Kadioglu D, Deutsch K, et al. Data Literacy in
der Medizin. Der Onkologe. 2018;24(5):368–377.
[30] Ryan P, Luz S, Albert P, et al. Using artificial intelli-
gence to assess clinicians’ communication skills. BMJ.
2019 Jan 18;364:l161. PubMed PMID: 30659013.
[31] Wallis C. How artificial intelligence will change
medicine. Nature. 2019 Dec;576(7787):S48. PubMed
PMID: 31853072; eng.
[32] Mosch L, Back D-A, Balzer F, et al. Lernangebote zu
Künstlicher Intelligenz in der Medizin. 2021. https://
zenodo.org/record/5497668
[33] Blum A, Bosch S, Haenssle H, et al. Künstliche intel-
ligenz und smartphone-programm-applikationen
(Apps). Der Hautarzt. 2020;71(9):691–698.
[34] Sit C, Srinivasan R, Amlani A, et al. Attitudes and
perceptions of UK medical students towards artificial
intelligence and radiology: a multicentre survey.
Insights Imaging. 2020 Feb 5;11(1):14. PubMed
PMID: 32025951; PubMed Central PMCID:
PMCPMC7002761. eng.
[35] Chen M, Zhang B, Cai Z, et al. Acceptance of clinical
artificial intelligence among physicians and medical stu-
dents: a systematic review with cross-sectional survey.
Front Med. 2022;9:990604. PubMed PMID: 36117979;
PubMed Central PMCID: PMCPMC9472134. eng.
[36] Valikodath NG, Cole E, Ting DSW, et al. Impact of
artificial intelligence on medical education in
ophthalmology. Transl Vis Sci Technol. 2021 Jun
1;10(7):14. PubMed PMID: 34125146; PubMed
Central PMCID: PMCPMC8212436.
[37] Palanica A, Flaschner P, Thommandram A, et al. phy-
sicians’ perceptions of chatbots in health care: cross-
sectional web-based survey. J Med Internet Res. 2019
Apr 5;21(4):e12887. PubMed PMID: 30950796;
PubMed Central PMCID: PMCPMC6473203. eng.
[38] Bisdas S, Topriceanu CC, Zakrzewska Z, et al.
Artificial Intelligence in medicine: a multinational
multi-center survey on the medical and dental stu-
dents’ perception. Front Public Health.
2021;9:795284. PubMed PMID: 35004598; PubMed
Central PMCID: PMCPMC8739771.
[39] Buhmann J, Felix J, Gächter T, et al. Digitalisierung der
Medizin: Konsequenzen für die Ausbildung.
Bulletin des Médecins Suisses. 2018 Oct 17.
MEDICAL EDUCATION ONLINE 9
Appendices
A1. Questions from the group work about mental health chatbots in medicine
How would you rate chatbots as therapy tools, for example, Woebot? What opportunities do they offer and what are their limitations?
How could they influence the doctor – patient relationship? In your opinion, what would chatbots need to do in order for both patients
and healthcare professionals to benefit from them in a meaningful way?
Which aspects of the use of chatbots in medicine do you see as particularly critical and why?
How do you think a chatbot should be used in terms of dialog design? How relevant are factors such as emoji use, artificial delays,
language and small talk?
What aspects should be considered in terms of linguistic style? Should the dialog be designed differently regardless of the target group?
How would you describe the optimal personality of the chatbot? Do you think a chatbot should have personality and/or show
emotion? How far can a chatbot be humanized in your view?
A2. I have a positive attitude towards the application and development of chatbots in medicine,
because. . .
Item
Agree
entirely
Rather
agree Undecided
Rather
disagree
Disagree
entirely
Mean (SD)
Prä
Wilcoxon
p < 0.05
I believe that the use of chatbots brings more opportunities than
risks.
Prä 8.3% 41.7% 33.3% 16.7% 0% 3.4 n.s.
(1/12) (5/12) (4/12) (2/12) (0/12) (0.9)
Post 27.3% 18.2% 45.5% 9.1% 0% 3.7
(3/11) (2/11) (5/11) (1/11) (0/11) (1.0)
you save time by using chatbots Prä 25.0% 58.3% 16.7% 0% 0% 4.1 n.s.
(3/12) (7/12) (2/12) (0/12) (0/12) (0.7)
Post 45. 5% 27.3% 18.2% 9.1% 0% 4. 1
(5/11) (3/11) (2/11) (1/11) (0/11) (1.0)
you save money by using chatbots Prä 33.3% 41.7% 25.0% 0% 0% 4.1 n.s
(4/12) (5/12) (3/12) (0/12) (0/12) (0.79)
Post 27.3% 27.3% 45.5% 0% 0% 3.8
(3/11) (3/11) (5/11) (0/11) (0/11) (0.8)
the integration of these, the participation in therapy decisions of
patients increases (e.g., less language barriers)
Prä 16.7% 41.7% 25.0% 0% 8.3% 3.6 n.s
(2/12) (5/12) (3/12) (0/12) (1/12) (1.1)
Post 9.1% 36.4% 45.5% 9.01% 0% 3.5
(1/11) (4/11) (5/11) (1/11) (0/11) (0.8)
Chatbots enable inclusion of people who otherwise have
inhibitions about confiding in a doctor
Prä 8.3% 33.3% 33.3% 8.3% 16.7% 3.1 n.s
(1/12) (4/12) (4/12) (1/12) (2/12) (1.2)
Post 18.2% 27.3% 45.5% 9.1% 0% 3.6
(2/11) (3/11) (5/11) (1/11) (0/11) (0.9)
Chatbots are available to everyone regardless of time and
location
Prä 33.3% 58.3% 8.3% 0% 0% 4.3 n.s
(4/12) (7/12) (1/12) (0/12) (0/12) (0.6)
Post 54.6% 36.4% 9.1% 0% 0% 4.5
(6/11) (4/11) (1/11) (0/11) (0/11) (0.7)
Chatbots can draw on a large database and thus have access to
more knowledge than a physician
Prä 50.0% 25.0% 16.7% 8.3% 0% 4.2 n.s
(6/12) (3/12) (2/12) (1/12) (0/12) (1.0)
Post 36.4% 27.3% 27.3% 9.1% 0% 3.9
(4/11) (3/11) (3/11) (1/11) (0/11) (1.0)
Chatbots are neutral listeners without being personally
judgmental
Prä 16.7% 8.3% 41.7% 16.7% 16.7% 2.3 n.s
(2/12) (1/12) (5/12) (2/12) (2/12) (1.3)
Post 9.1%
(1/11)
36.4%
(4/11)
27.3%
(3/11)
27.3%
(3/11)
0%
(0/11)
3.3 (1.0)
10 J.-A. MOLDT ET AL.
A3. I am critical of the use of chatbots in medicine, because. . .
Item
Agree
entirely
Rather
agree Undecided
Rather
disagree
Disagree
entirely
Mean
(SD)
Wilcoxon
p < 0.05
I believe that data protection and data security cannot be fully
guaranteed
Prä 16.7% 33.3% 41.7% 0% 8.3% 3.5 0.02
(2/12) (4/12) (5/12) (0/12) (1/12) (1.1)
Post 0% 36.4% 18.2% 27.3% 18.2% 2.7
(0/11) (4/11) (2/11) (3/11) (2/11) (1.14)
the health data can be easily manipulated Prä 16.7% 33.3% 25.0% 16.7% 8.3% 3.3 n.s.
(2/12) (4/12) (3/12) (2/12) (1/12) (1.2)
Post 9.1% 36.4% 18.2% 18.2% 18.2% 3.0
(1/11) (4/11) (2/11) (2/11) (2/11) (1.3)
the personal contact gets lost Prä 50.0% 16.7% 25.0% 8.3% 0% 4.1 n.s
(6/12) (2/12) (3/12) (1/12) (0/12) (1.1)
Post 27.3% 36.4% 9.1% 27.3% 0% 3.6
(3/11) (4/11) (1/11) (3/11) (0/11) (1.15)
I am not ready to discuss sensitive data with a chatbot. Prä 25.0% 16.7% 25.0% 16.7% 16.7% 3.2 n.s
(3/12) (2/12) (3/12) (2/12) (2/12) (1.5)
Post 36.4% 9.1% 9.1% 36.4% 9.1% 3.3
(4/11) (1/11) (1/11) (4/11) (1/11) (1.5)
there may be possible communication problems due to
immature technology
Prä 50.0% 50.0% 0% 0% 0% 4.5 n.s
(6/12) (6/12) (0/12) (0/12) (0/12) (0.5)
Post 54.6% 27.3% 18.2% 0% 0% 4.4
(6/11) (3/11) (2/11) (0/11) (0/11) (0.8)
Chatbots are not established enough yet and long-term success
is yet to come
Prä 25.0% 33.3% 33.3% 8.3% 0% 3.8 n.s
(3/12) (4/12) (4/12) (1/12) (0/12) (1.0)
Post 36.4% 45.5% 18.2% 0% 0% 4.2
(4/11) (5/11) (2/11) (0/11) (0/11) (0.7)
I am concerned that my data will be transferred and evaluated
from uninvolved third parties
Prä 16.7% 50.0% 16.7% 8.3% 8.3% 3.6 n.s
(2/12) (6/12) (2/12) (1/12) (1/12) (1.2)
Post 18.2% 9.1% 18.2% 36.4% 18.8% 2.8
(2/11) (1/11) (2/11) (4/11) (2/11) (1.4)
MEDICAL EDUCATION ONLINE 11
... In addition, chatbots offer scalable, accessible digital therapy for addressing the mental health needs of students in HE, potentially improving academic outcomes and retention rates (Biro et al., 2023;Lin et al., 2021). Medical chatbots also assist in administrative tasks such as scheduling appointments, managing patient records, providing information and support, enhancing engagement with medical students, and managing stress levels (Moldt et al., 2022;Moldt et al., 2023). ...
... In addition, they have mentioned that they have encountered technical obstacles and that mentors for digital students need continuous support and training. They also emphasized the importance of human support when it is needed (Ilieva et al., 2023;Moldt et al., 2023). On the other hand, there were also contradictory results; for example, Bilquise et al. (2023) revealed that social influence positively impacts the behavioral intention to use an advising chatbot, while Rahim et al. (2022) reported that students do not require social reinforcement to use a chatbot, and users are not influenced by the opinions, suggestions, or recommendations of important others who think they should adopt a chatbot. ...
... These ethical issues are worsened by the opacity of knowledge about AI chatbot algorithms. Students and even instructors often miss a full picture of how the underlying AI works, skepticism about chatbot outputs is common (Moldt et al. 2023). This issue is compounded by the relative opaqueness of various AI models, where their inner functioning is black-boxed which obfuscates how data and decisions are processed. ...
Article
Full-text available
Artificial Intelligence (AI) chatbot 'ChatGPT' in the education sector has changed the learning patterns among students, staff, and lecturers. The rise of AI-driven tools is bringing forth significant ethical concerns, which are considered essential to discuss. The present study was conducted in three public sector universities in Malaysia, where we opted for the quantitative research design and approached the research participants through personal invitation and snowball sampling procedures. A total of 406 respondents were involved in this study, and the data was gathered through a survey, using structured questionnaire. The purpose of this method is to collect diverse data from a variety of participants, including students, staff, and lecturers. The findings show that it is imperative to highlight the need for robust ethical guidelines and a higher education institution framework to ensure that implementation is indispensable. Furthermore, addressing these ethical challenges can harness the potential of chatbots in academia along with ensuring ethical concerns being addressed. The study concluded that academicians had clear understanding about the ethical issues of using ChatGPT or AIs. The study also suggests recommendations regarding the ethical usage by the government of Malaysia. Abstrak Chatbot Kecerdasan Buatan (AI) 'ChatGPT' di sektor pendidikan telah mengubah pola pembelajaran di kalangan mahasiswa, staf, dan dosen. Meningkatnya alat yang digerakkan oleh AI memunculkan berbagai masalah etika yang signifikan, yang dianggap penting untuk dibahas. Studi saat ini dilakukan di tiga universitas sektor publik di Malaysia, di mana kami memilih desain penelitian kuantitatif dan mendekati partisipan penelitian melalui undangan pribadi dan prosedur pengambilan sampel bola salju. Sebanyak 406 responden terlibat dalam penelitian ini, dan data dikumpulkan melalui survei, menggunakan kuesioner terstruktur. Tujuan dari metode ini adalah untuk mengumpulkan beragam data dari berbagai partisipan, termasuk mahasiswa, staf, dan dosen. Temuan menunjukkan bahwa sangat penting untuk menyoroti perlunya pedoman etika yang kuat dan kerangka kerja lembaga pendidikan tinggi untuk memastikan bahwa implementasi sangat diperlukan. Lebih jauh, mengatasi tantangan etika ini dapat memanfaatkan potensi chatbot di dunia akademis sekaligus memastikan masalah etika ditangani. Studi ini menyimpulkan bahwa akademisi memiliki pemahaman yang jelas tentang masalah etika penggunaan ChatGPT atau AI. Studi ini juga menyarankan rekomendasi mengenai penggunaan etika oleh pemerintah Malaysia. Kata kunci: kecerdasan buatan dalam pendidikan; ChatGPT; masalah etika; lembaga pendidikan tinggi 433
... The questionnaire assessed the attitude toward artificial intelligence of the undergraduate medical students who attended the course (n=354). The questionnaire was adapted from Stewart et al. [12], Moldt et al. [13] and Grassini, S. [14]. The attitude questionnaire included 28 questions, as demonstrated in supplementary 2, about four subthemes: undergraduate medical students' optimistic perspectives towards AI in healthcare (6 questions), undergraduate medical students' concerns towards AI in healthcare (8 questions), undergraduate medical students' optimistic perspectives towards AI in medical education (8 questions), and undergraduate medical students' concerns towards AI in medical education (6 questions). ...
... Currently, clinicians are raising concerns about the reliability, limited knowledge or training of chatbots, implementation costs, ethical concerns, and the absence of regulatory approval for integrating AI-based LLM chatbots into routine practice [105][106][107][108][109]. Tackling these challenges requires essential collaboration between chatbot developers and healthcare providers, ongoing professional training, strict regulatory oversight, and the establishment and enforcement of relevant ethical guidelines and regulations. ...
Article
Full-text available
Artificial intelligence (AI) is becoming increasingly influential in ophthalmology, particularly through advancements in machine learning, deep learning, robotics, neural networks, and natural language processing (NLP). Among these, NLP-based chatbots are the most readily accessible and are driven by AI-based large language models (LLMs). These chatbots have facilitated new research avenues and have gained traction in both clinical and surgical applications in ophthalmology. They are also increasingly being utilized in studies on ophthalmology-related exams, particularly those containing multiple-choice questions (MCQs). This narrative review evaluates both the opportunities and the challenges of integrating chatbots into ophthalmology research, with separate assessments of studies involving open- and close-ended questions. While chatbots have demonstrated sufficient accuracy in handling MCQ-based studies, supporting their use in education, additional exam security measures are necessary. The research on open-ended question responses suggests that AI-based LLM chatbots could be applied across nearly all areas of ophthalmology. They have shown promise for addressing patient inquiries, offering medical advice, patient education, supporting triage, facilitating diagnosis and differential diagnosis, and aiding in surgical planning. However, the ethical implications, confidentiality concerns, physician liability, and issues surrounding patient privacy remain pressing challenges. Although AI has demonstrated significant promise in clinical patient care, it is currently most effective as a supportive tool rather than as a replacement for human physicians.
... The findings indicated that ChatGPT can potentially enhance student learning and affect their results. Moldt et al. (2023) studied medical students' perspectives on using chatbots and AI in medicine. The study employed a mixed-methods approach. ...
Article
To better understand the use of ChatGPT as a tool for learning in higher education, this investigation reports student responses to an e-questionnaire investigating their perspectives on it. In particular, it delves into whether there are any significant differences between opinions based on gender and the level of study of 51 postgraduate students in the applied linguistics program at Majmaah University, Saudi Arabia. The study’s findings revealed that postgraduates’ attitudes were reported as being highly positive toward using ChatGPT as a learning aid in higher education. Considering gender differences, the descriptive statistics showed that male Saudi postgraduate students have a more positive attitude toward using ChatGPT as a tool for learning in terms of their study level. The findings indicated no significant difference among first- and second-year postgraduate students toward using ChatGPT as a learning aid, which did not affect the study’s findings. Keywords: ChatGPT; gender differences; learning aid; postgraduate; postgraduates’ perspectives
Preprint
Full-text available
Background With the increase of medical demand, the teaching quality of medical universities was highly valued by all sectors of society. How to innovate the teaching management mode in the new era was an important problem faced by all universities. In the past decade, many research hotspots have emerged in medical teaching management. The research of the latest literature in this period can help us grasp the frontier dynamics of this field, which was of great significance to the development of higher medical education. Methods From Web of Science,we searched for studies published between 2014 and 2023 using the terms “medical education” and “teaching management”, ended up with 1407 eligible papers. CiteSpace software was used to visually analyze the research institutions, authors and keywords of the studies. Results The number of research on medical teaching management had shown a steady increase trend in the past decade. BMC MEDICAL EDUCATION was the journal with the largest number of publications. Durning, Steven J., of the Uniformed Services University of the Health Sciences - USA, was the most prolific author. Literature research in the field of medical teaching management had many research directions but not deep enough. Conclusions As the core author group had not yet been formed and the team effect cannot be brought into play. Organized teaching research that can efficiently integrate resources and cooperatively complete tasks was urgently needed.Literature research in the field of medical teaching management had many research directions but not deep enough. This study identified three characteristics of medical teaching management research in recent years: people-oriented, application of high technology, and teaching reform. These topics can provide useful information for educators and researchers to continue in-depth research.
Article
The integration of Artificial Intelligence (AI) into healthcare is set to fundamentally transform the role of medical professionals. This article takes a dive into the future specifying several strategies to prepare medical and other health professions students for an AI-driven healthcare system. It highlights the current and potential applications of AI in healthcare and discusses, in the immediate term, the necessity of incorporating AI education into medical curricula, including hands-on training and interdisciplinary collaboration. With a view to a more distant future, it also addresses probable ethical considerations, the evolving roles of healthcare professionals, a need for contingency, and emphasises the importance of maintaining clinical skills amidst AI advancements. These strategies all aim to equip future medical graduates with competencies in knowledge, skills, and ethical grounding required to thrive in an AI-dominated healthcare environment. This framework is intended to stimulate discussion and provide a foundation upon which more detailed, context-specific strategies can be developed by educators, policymakers, and other stakeholders. https://link.springer.com/article/10.1007/s40670-024-02257-9
Article
Full-text available
Objective Digital transformation in higher education has presented medical students with new challenges, which has increased the difficulty of organising their own studies. The main objective of this study is to evaluate the effectiveness of a chatbot in assessing the stress levels of medical students in everyday conversations and to identify the main condition for accepting a chatbot as a conversational partner based on validated stress instruments, such as the Perceived Stress Questionnaire (PSQ20). Methods In this mixed-methods research design, medical-student stress level was assessed using a quantitative (digital- and paper-based versions of PSQ20) and qualitative (chatbot conversation) study design. PSQ20 items were also shortened to investigate whether medical students’ stress levels can be measured in everyday conversations. Therefore, items were integrated into the chat between medical students and a chatbot named Melinda. Results PSQ20 revealed increased stress levels in 43.4% of medical students who participated ( N = 136). The integrated PSQ20 items in the conversations with Melinda obtained similar subjective stress degree results in the statistical analysis of both PSQ20 versions. Qualitative analysis revealed that certain functional and technical requirements have a significant impact on the expected use and success of the chatbot. Conclusion The results suggest that chatbots are promising as personal digital assistants for medical students; they can detect students’ stress factors during the conversation. Increasing the chatbot's technical and social capabilities could have a positive impact on user acceptance.
Article
Full-text available
Background Artificial intelligence (AI) needs to be accepted and understood by physicians and medical students, but few have systematically assessed their attitudes. We investigated clinical AI acceptance among physicians and medical students around the world to provide implementation guidance. Materials and methods We conducted a two-stage study, involving a foundational systematic review of physician and medical student acceptance of clinical AI. This enabled us to design a suitable web-based questionnaire which was then distributed among practitioners and trainees around the world. Results Sixty studies were included in this systematic review, and 758 respondents from 39 countries completed the online questionnaire. Five (62.50%) of eight studies reported 65% or higher awareness regarding the application of clinical AI. Although, only 10–30% had actually used AI and 26 (74.28%) of 35 studies suggested there was a lack of AI knowledge. Our questionnaire uncovered 38% awareness rate and 20% utility rate of clinical AI, although 53% lacked basic knowledge of clinical AI. Forty-five studies mentioned attitudes toward clinical AI, and over 60% from 38 (84.44%) studies were positive about AI, although they were also concerned about the potential for unpredictable, incorrect results. Seventy-seven percent were optimistic about the prospect of clinical AI. The support rate for the statement that AI could replace physicians ranged from 6 to 78% across 40 studies which mentioned this topic. Five studies recommended that efforts should be made to increase collaboration. Our questionnaire showed 68% disagreed that AI would become a surrogate physician, but believed it should assist in clinical decision-making. Participants with different identities, experience and from different countries hold similar but subtly different attitudes. Conclusion Most physicians and medical students appear aware of the increasing application of clinical AI, but lack practical experience and related knowledge. Overall, participants have positive but reserved attitudes about AI. In spite of the mixed opinions around clinical AI becoming a surrogate physician, there was a consensus that collaborations between the two should be strengthened. Further education should be conducted to alleviate anxieties associated with change and adopting new technologies.
Article
Full-text available
Background: The emerging field of artificial intelligence (AI) will probably affect the practice for the next generation of doctors. However, the students' views on AI have not been largely investigated. Methods: An anonymous electronic survey on AI was designed for medical and dental students to explore: (1) sources of information about AI, (2) AI applications and concerns, (3) AI status as a topic in medicine, and (4) students' feelings and attitudes. The questionnaire was advertised on social media platforms in 2020. Security measures were employed to prevent fraudulent responses. Mann-Whitney U-test was employed for all comparisons. A sensitivity analysis was also performed by binarizing responses to express disagreement and agreement using the Chi-squared test. Results: Three thousand one hundred thirty-three respondents from 63 countries from all continents were included. Most respondents reported having at least a moderate understanding of the technologies underpinning AI and of their current application, with higher agreement associated with being male (p < 0.0001), tech-savvy (p < 0.0001), pre-clinical student (p < 0.006), and from a developed country (p < 0.04). Students perceive AI as a partner rather than a competitor (72.2%) with a higher agreement for medical students (p = 0.002). The belief that AI will revolutionize medicine and dentistry (83.9%) with greater agreement for students from a developed country (p = 0.0004) was noted. Most students agree that the AI developments will make medicine and dentistry more exciting (69.9%), that AI shall be part of the medical training (85.6%) and they are eager to incorporate AI in their future practice (99%). Conclusion: Currently, AI is a hot topic in medicine and dentistry. Students have a basic understanding of AI principles, a positive attitude toward AI and would like to have it incorporated into their training.
Article
Full-text available
Background Literature, individual experiences and common considerations suggest that prior professional qualification can be an advantage for later career development. For instance, in Germany, professional pre-qualification has been honored by medical faculties in selection procedures for several years. However, a systematic evaluation of this relationship lacks. This scoping review summarizes existing literature and addresses the role of prior professional pre-qualifications on objective or subjective study success and the choice of a specialization. Methods The scoping review was performed oriented on the PRISMA guidelines. PsycINFO and PubMed databases were searched for relevant studies that included data of medical students with and without professional pre-qualifications. To answer the underlying research questions, this scoping review also includes studies that examine professional pre-qualifications in association with non-cognitive "soft" criteria. Results and further directions 1055 items were identified and reviewed by two independent reviewers with final 11 studies were included for this scoping review. The results of identified studies that report possible effects of prior pre-professional qualifications are inconclusive but suggest that prior professional qualifications tend not to have rather an advantage on study success. Medical school success for students with prior professional qualifications tended to be below average in the preclinical setting, and there were no differences in the clinical setting compared with students without prior professional qualifications. The influence of professional pre-qualifications has not yet been adequately studied without the moderator variable “waiting time” and “A-levels grade”. The scoping review indicates insufficient number of articles stating a co-relation of prior pre-qualifications and subjective data. Again, the results found are not sufficient to state a clear relationship between professional pre-qualifications and the choice of a specific speciality preference. However, professional pre-qualifications, both in medicine and as "practical experience in rural areas", tend to be beneficial for the choice of becoming a rural physician. Large-scale cross-sectional and longitudinal studies are needed to investigate the influence of professional pre-qualifications on different study trajectory parameters.
Book
Full-text available
Der zunehmende Einsatz Künstlicher Intelligenz (KI) in der Medizin wird tiefgreifende Veränderungen hervorrufen. Angesichts dessen benötigen (zukünftige) Mediziner*innen Kompetenzen, die sie zur nutzbringenden Anwendung von KI-Systemen in Praxis und Forschung befähigen. Wir erheben den Bestand an Lernangeboten zum Thema KI in der Medizin für Medizinstudierende und Ärzt*innen in Deutschland, definieren relevante KI-Kompetenzen und Formate und identifizieren Bedarfslücken und Herausforderungen der Implementierung von Lernangeboten in der medizinischen Aus-, Weiter- und Fortbildung. Schließlich legen wir Lösungsstrategien bzw. Empfehlungen für die erfolgreiche KI-Qualifizierung (zukünftiger) Mediziner*innen vor.
Article
Full-text available
Many experts have emphasised that chatbots are not sufficiently mature to be able to technically diagnose patient conditions or replace the judgements of health professionals. The COVID-19 pandemic, however, has significantly increased the utilisation of health-oriented chatbots, for instance, as a conversational interface to answer questions, recommend care options, check symptoms and complete tasks such as booking appointments. In this paper, we take a proactive approach and consider how the emergence of task-oriented chatbots as partially automated consulting systems can influence clinical practices and expert–client relationships. We suggest the need for new approaches in professional ethics as the large-scale deployment of artificial intelligence may revolutionise professional decision-making and client–expert interaction in healthcare organisations. We argue that the implementation of chatbots amplifies the project of rationality and automation in clinical practice and alters traditional decision-making practices based on epistemic probability and prudence. This article contributes to the discussion on the ethical challenges posed by chatbots from the perspective of healthcare professional ethics.
Chapter
The advent of chatbots (also called conversational agents) has been recognized to support various treatment procedures in the medical and psychological fields. Chatbots may be particularly useful before and after medical procedures when patients are at home. For example, while being in the preparation phase of a colonoscopy, a chatbot might answer patient questions more quickly than a doctor. It is more and more discussed whether chatbots may be the first entry point for (urgent) medical questions instead of the consultation of a medical expert, as there exist already well-established algorithms for such situations. For example, if a new medical symptom occurs, a chatbot might serve as the first expert to relieve a patient’s situation. Such situations are mainly driven by the trend that patients often have to wait very long for appointments with a proper medical expert due to capacity problems of the healthcare system. While the usage of supporting at home actions of patients with chatbot technologies is typically welcomed by medical experts, the use of this technology to replace them in their core competence, namely diagnosis and therapy, is generally seen highly critical. Apart from the domain side, it must be carefully considered what currently available chatbot technologies can do and not do. Moreover, it has also to be considered, how existing technologies can be established in such highly critical (e.g., if a chatbot gets the message of a patient that indicates to commit suicide) and interdisciplinary field, involving ethical questions as well as questions of responsibility and accountability. This work raises aspects that might be the basis for medical as well as technical experts to better collaborate for proper chatbot solutions. Concretely, an architecture is proposed that should serve as a reference for various medical and psychological scenarios. When using suitable technical solutions based on such architectures, we argue that many opportunities emerge, which can mitigate identified challenges significantly.
Article
The rapid development of machine learning (ML) applications in healthcare promises to transform the landscape of healthcare. In order for ML advancements to be effectively utilized in clinical care, it is necessary for the medical workforce to be prepared to handle these changes. As physicians in training are exposed to a wide breadth of clinical tools during medical school, this offers an ideal opportunity to introduce ML concepts. A foundational understanding of ML will not only be practically useful for clinicians, but will also address ethical concerns for clinical decision making. While select medical schools have made effort to integrate ML didactics and practice into their curriculum, we argue that foundational ML principles should be taught broadly to medical students across the country.
Article
Artificial intelligence (AI) is poised to broadly reshape medicine, potentially improving the experiences of both clinicians and patients. We discuss key findings from a 2-year weekly effort to track and share key developments in medical AI. We cover prospective studies and advances in medical image analysis, which have reduced the gap between research and deployment. We also address several promising avenues for novel medical AI research, including non-image data sources, unconventional problem formulations and human–AI collaboration. Finally, we consider serious technical and ethical challenges in issues spanning from data scarcity to racial bias. As these challenges are addressed, AI’s potential may be realized, making healthcare more accurate, efficient and accessible for patients worldwide. AI has the potential to reshape medicine and make healthcare more accurate, efficient and accessible; this Review discusses recent progress, opportunities and challenges toward achieving this goal.