VanCauwenbergeetal. BMC Medical Ethics (2022) 23:50
“Many roads lead toRome andtheArticial
Intelligence onlyshows me one road”:
aninterview study onphysician attitudes
clinical decision support systems
Daan Van Cauwenberge1,2, Wim Van Biesen2,3, Johan Decruyenaere2,4, Tamara Leune2,3 and Sigrid Sterckx1,2*
Research regarding the drivers of acceptance of clinical decision support systems (CDSS) by physicians is still rather
limited. The literature that does exist, however, tends to focus on problems regarding the user-friendliness of CDSS.
We have performed a thematic analysis of 24 interviews with physicians concerning speciﬁc clinical case vignettes,
in order to explore their underlying opinions and attitudes regarding the introduction of CDSS in clinical practice, to
allow a more in-depth analysis of factors underlying (non-)acceptance of CDSS. We identiﬁed three general themes
from the results. First, ‘the perceived role of the AI’, including items referring to the tasks that may properly be assigned
to the CDSS according to the respondents. Second, ‘the perceived role of the physician’, referring to the aspects
of clinical practice that were seen as being fundamentally ‘human’ or non-automatable. Third, ‘concerns regarding
AI’, including items referring to more general issues that were raised by the respondents regarding the introduc-
tion of CDSS in general and/or in clinical medicine in particular. Apart from the overall concerns expressed by the
respondents regarding user-friendliness, we will explain how our results indicate that our respondents were primarily
occupied by distinguishing between parts of their job that should be automated and aspects that should be kept
in human hands. We refer to this distinction as ‘the division of clinical labor.’ This division is not based on knowledge
regarding AI or medicine, but rather on which parts of a physician’s job were seen by the respondents as being central
to who they are as physicians and as human beings. Often the respondents’ view that certain core parts of their job
ought to be shielded from automation was closely linked to claims concerning the uniqueness of medicine as a
domain. Finally, although almost all respondents claimed that they highly value their ﬁnal responsibility, a closer inves-
tigation of this concept suggests that their view of ‘ﬁnal responsibility’ was not that demanding after all.
Keywords: Clinical decision support systems, AI, Medicine, Automation, Division of clinical labour, Responsibility
© The Author(s) 2022. Open Access This ar ticle is licensed under a Creative Commons Attr ibution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the
original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or
other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line
to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory
regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this
licence, visit http:// creat iveco mmons. org/ licen ses/ by/4. 0/. The Creative Commons Public Domain Dedication waiver (http:// creat iveco
mmons. org/ publi cdoma in/ zero/1. 0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
Over the last decade, automated clinical decision support
(CDS) using artiﬁcial intelligence (AI) has gained increas-
ing interest. A CDS system (CDSS) is usually deﬁned
as a computer-based information system that supports
decision making in patient care by integrating clinically
relevant patient information and presenting it to the
1 Department of Philosophy and Moral Sciences, Bioethics Institute Ghent,
Ghent University, Ghent, Belgium
Full list of author information is available at the end of the article
Page 2 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
healthcare worker. e more widespread introduction
of electronic health records (EHRs) facilitates the emer-
gence of CDSS as it allows incorporation of clinical deci-
sion support at diﬀerent levels and for diﬀerent purposes.
It is frequently assumed that AI-based CDSS will improve
health care , although studies proving this hypothesis
are scarce [2–4]. Moreover, the few cases where such AI-
based CDSS have been incorporated in EHRs failed to
improve clinically relevant outcomes .
More speciﬁcally, the introduction of EHR has also
been associated with increased burnout in healthcare
workers and decreased patient satisfaction. It has been
argued that EHRs add administrative burden (the “death
by a thousand clicks” ), and that they stand in the way
of true involvement between healthcare worker and
patient . e implementation of CDSS into everyday
care is thus considered a major step forward by some ,
and a major challenge to health care and the medical pro-
fession by others [8–10].
A lot of research has identiﬁed factors associated with
acceptance of CDSS by physicians [11, 12]. Technical
characteristics of the software, such as user interface
and transparency, the clinical aspects of the task at hand,
and the expertise of the physician with the CDS device
have been reported as important factors. However, next
to these engineering and technical issues, a substantial
human factor remains, as the human operator’s interac-
tion with the system is the necessary step for enacting the
recommendation of the device . For example, 49–96%
of interruptive medication alerts are simply ignored by
the operator . Little evidence is available on the true
underlying motivations, emotions and argumentations or
their modulating factors driving the acceptance of and
reaction of physicians to the incorporation of CDSS in
to EHRs. Most EHRs available on the market today are
designed from an administrative and informatics back-
ground perspective . Consequently, they rarely con-
sider the speciﬁc requirements of clinical tasks and the
typical reasoning process of healthcare workers .
erefore, we designed a mixed methods qualitative
study to thematically explore the reactions and under-
lying reasoning of physicians when confronted with
vignettes in which hypothetical CDSS incorporated in
an EHR were presented. We hypothesized that many
of the factors for the acceptance of CDSS by physicians
reported in the literature, such as transparency, alert
fatigue, and user friendliness, may have some more
fundamental underlying drivers in common. Uncov-
ering these might help to better understand, and thus
potentially avoid, the sometimes ambiguous behavior
of physicians when confronted with CDSS in order to
enable proper development and implementation of
such systems. While most researchers have focused
on these more common factors, our analysis aims to
uncover these more hidden underlying drivers.
is thematic analysis is part of a larger study per-
formed in a university hospital in a transition to select-
ing, customizing and implementing a new electronic
health record (EHR) system. While the existing EHR
was considered to be “ﬁt for the task”, it does not have a
clinical decision support system (CDSS). A decision to
upgrade to a more modern EHR incorporating such a
CDSS was approved. All participants were thus famil-
iar with the use of an EHR, however, their exposure to
CDSS actually incorporated in an EHR was low.
Following recommendations on Q-sort methodology,
we intended to interview 30 physicians, with purposive
sampling to achieve a mix of gender and level of exper-
tise (trainee, junior staﬀ, senior staﬀ ) [16, 17]. Respond-
ents were interviewed in a silent room, and all sources
of distraction were avoided as much as possible.
We used a mixed methods approach with a Q-sort
based classiﬁcation of pre-deﬁned reactions to clinical
case vignettes (see Additional ﬁle 1: Qsort vignettes),
in combination with a thinking-aloud approach in
which reasoning and attitudes of the participant dur-
ing the classiﬁcation task were solicited. All sessions
were done by the same interviewer (blinded for review)
with expertise in the ﬁeld of computerized CDSS. All
sessions were audiotaped and typed out verbatim
afterwards by another member of the research team
(blinded for review). Q-sort techniques can be used to
explore subjectivity and attitudes in a more systematic
way, and to provide insights into potential patterns and
concepts within this subjectivity . e audiotaping
of the thinking aloud during the actual completion of
the Qsort allowed to not only gain insights in the per-
ceptions of the participants, but also in their underlying
reasoning, emotions and motivations.
irty statements describing potential actions of a
(hypothetical) CDSS in four well deﬁned clinical set-
tings (for the vignettes describing the potential actions
see Additional ﬁle1: Qsort vignettes) were constructed.
Page 3 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
Statements represented variations of diﬀerent factors
already associated with uptake of CDSS by physicians:
(1) transparency of the system; (2) the degree of cer-
tainty regarding the correctness of the advice provided
by the device; (3) the interruption of workﬂow with or
without obstructiveness1; and (4) the type of problem
at hand. In a pilot, statements were evaluated for clar-
ity and lack of ambiguity. In the ﬁrst two vignettes, the
focus was on decision support for medication orders,
covering formulary alerts, drug-drug interaction alerts
and drug-condition alerts . In the third vignette, a
diagnostic problem was raised, assessing automated
CDS for order entry as well as for presenting and inter-
preting results of the ordered diagnostic radiology
tests. In the last vignette, more advanced CDS for han-
dling a complex clinical problem was presented.
A thematic extraction of underlying opinions, attitudes
and arguments of the participants was performed based
on the audiotapes of the thinking-aloud during the inter-
emes and concepts issued by the participants were
grouped and re-grouped by two members of the research
team (blinded for review) until all concepts were placed
under a non-overlapping header. Two diﬀerent triangu-
lation session with all team members were performed to
reach a consensus on the thematic analysis. Quantitative
results of the Qsort will be published separately.
Our ﬁndings are based on an analysis of the audio
recordings of twenty-four interviews. e respondents
all worked at Ghent University Hospital. Fifteen respond-
ents identiﬁed as female and all were at diﬀerent stages of
Based on our analysis we identiﬁed three overarching
themes, which can be further divided into smaller themes
and subthemes: (1) Perceived role of the AI; (2) Perceived
role of the physician; and (3) Concerns regarding AI.
e ﬁrst two themes focus on the roles that were
ascribed to either AI or the physician, respectively.
Regardless of their general opinion towards the introduc-
tion of AI in the context of medicine, respondents were
in favour of assigning certain tasks to the AI. When argu-
ing in favour or against certain positions, the respond-
ents almost never consciously or explicitly formulated
ethical arguments, or referred to ethical principles such
as justice, fairness, beneﬁcence, or even patient auton-
omy.2 Of course this is not to say that their arguments
did not reﬂect ethical beliefs, but rather that they did
not explicitly formulate them as such. e respondents’
mostly used a two-step argumentation. First, it was, often
implicitly, assumed that either the AI or the physician
had a certain role to play in the medical practice. Second,
it was argued that a given position was either good or
bad in relation to this presupposed role. Some respond-
ents limited the number of tasks the AI was allowed to
perform by emphasizing the unique and indispensable
role of the physician in clinical practice. roughout this
paper we will refer to the allocation of clinical tasks to
certain actors as the division of clinical labour.
e third theme consists of concerns regarding AI
that cannot be reduced to a discussion on the division of
clinical labour, but that instead relates to other issues of
concern. Importantly, in this last category respondents
sometimes did use negative ethical statements such as
‘bad’ or ‘evil.’
Overarching theme 1: perceived role oftheAI
e ﬁrst overarching theme concerns all roles that were
assigned to the AI. ese vary from very concrete tasks,
for example all administrative tasks, to more general
Fig. 1 Overarching theme 1 (above), themes (below)
1 By ‘obstructiveness’ we refer to the extent to which the direct control of the
operator is obstructed by the system. Obstructiveness does not necessarily
mean that the operator is in no way able to override the system, but rather
that their workﬂow is being interrupted.
2 An important exception was the use of the words ‘evil,’ ‘bad’ or ‘horrible.’
Unlike their positive counterparts, these negative statements were sometimes
used by respondents. We will return to this subject when discussing the third
overarching theme that emerged from our analysis.
Page 4 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
goals, such as increasing safety or eﬃciency. is ﬁrst
theme can be divided into six smaller themes: (1) safety;
(2) eﬃciency; (3) learning from the system; (4) adminis-
tration; (5) organisation of data; (6) and the ePhysician
Several respondents indicated that they believed the inte-
gration of AI in clinical practice would increase safety.
ere were two ways in which the respondents believed
AI would increase safety. First, they argued that the AI
would broaden their knowledge by suggesting illnesses or
therapies they would not have thought of otherwise.
AI will not replace us, but it will certainly help us. …
It will increase our performance by reminding us of
rare illnesses and combining all relevant data. (R9)
Second, they argued that the AI would force physi-
cians to better reﬂect on their choices by implement-
ing obstructive messages and rendering certain actions
impossible. For example:
is [obstructing certain actions] could be a good
thing when you are dealing with cowboys, people
who think they know it best. [Obstructing them]
could be a good thing to protect people against
It is dangerous when you are able to just close a pop-
up without changing anything. e pop-ups should
ensure that the behaviour has changed. (R4)
Almost all respondents indicated they believed AI would
make clinical practice more eﬃcient. is emphasis on
eﬃciency was especially prevalent in their judgement of
the many ways in which a CDSS could be implemented.
ey always preferred the version that emphasized eﬃ-
ciency and speed in general.
I want the system to increase my eﬃciency and pro-
One respondent suggested that the implementation of
AI would not increase eﬃciency. Although the respond-
ent initially argued that AI would make clinical practice
more eﬃcient, she later corrected her earlier statement.
I fear that our work is never done. [Even after imple-
menting AI] I think there will be new ineﬃciencies.
But now I am being philosophical (laughs). (R23)
Many respondents stated that they did not simply
want to hand over certain tasks to the AI, but that they
themselves wanted to learn from and improve their skills
through the system. Most often this view was expressed
by negative statements presented in the vignettes.
Respondents regularly disliked certain statements,
because they felt they would not ‘learn’ anything from
them. For example:
I ﬁnd it strange that it does not tell us why it is pro-
moting a certain course of action. What is the origin
of the decision? is way you do not learn anything
[from the system]. (R9)
Generally, the respondents thought it was important
that they should not just take the backseat and let the AI
do the work. Some of the respondents argued that not
only their job performance, but also their reasoning pro-
cess should be improved by the AI.
Many of the respondents created a dichotomy between
administrative tasks on the one hand and the core medi-
cal tasks of the physician on the other. None of the
respondents was very clear about the kind of tasks that
belonged to these categories. Yet, almost all of them
characterised the administrative tasks as being simpler
and therefore easier to automate.
[Administrative tasks] are banal, they are trivial.
ey are very easy and should, obviously, be inte-
grated [in the system]. If I were in charge of [ICT],
this would be the ﬁrst step I would take. (R6)
Indeed, most respondents had no issue with automat-
ing administrative tasks. ey only expressed doubts
about the automation of the tasks they perceived as being
essential to their job, such as diagnosis or the prescrip-
tion of therapy. Sometimes, this distinction was made
[Unlike with medical decisions] I do trust the AI
when it takes administrative decisions. ose do not
look diﬃcult to me. (R14)
Some respondents suggested that the AI would be espe-
cially useful as an organiser of data.3 ey mostly pre-
ferred the AI to collect, organise and present data to the
3 e organisation of data diﬀers from the previous subtheme. Administrative
tasks are already existing tasks that the physician would like to be automated.
For instance, notifying the physician when the hospital does not have a certain
medicine. ‘Organising data’ on the other hand, refers to a set of tasks that did
not exist before the introduction of AI and data analysis, more speciﬁcally the
acquisition, organisation, and analysis of data on a scale that was not possible
before the introduction of AI.
Page 5 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
physician. For example, when asked which job he would
most like the AI to perform, respondent 3 answered:
e best [job the AI could perform] would be the
organisation of my data based on their relevance
and importance. (R3)
Crucially, all the roles taken on by AI mentioned so far
are merely supportive. All of the previous themes sug-
gested that the AI and the physician should become col-
leagues of sorts, but they disagreed about exactly what
kind of jobs the AI should perform. Moreover, some
respondents indicated that they believed the AI would
come to perform every task of the physician, essentially
becoming an electronic physician or ePhysician:
If the system would really have all of the necessary
information, I believe they would often be more reli-
able [than a human]. (R16)
Almost all respondents at a given point in the inter-
view considered the prospect of their job becoming fully
Overarching theme 2: perceived role ofthephysician
e second overarching theme is the counterpart of the
ﬁrst. It concerns all roles that are assigned to the physi-
cian. ese roles are almost always framed as a reaction
to the perceived encroachment of the AI. erefore the
ﬁrst two themes can be seen as a confrontation of view-
points regarding the division of clinical labour. is over-
arching theme can be divided into three subthemes (see
Earlier we alluded to the fact that many respondents
introduced a dichotomy between tasks they considered
to be easier administrative tasks and those they perceived
to be more diﬃcult clinical tasks. Some respondents
attempted to delineate exactly what diﬀerentiates those
administrative tasks from medical ones by emphasizing
certain clinical skills one needs in order to be able to per-
form these clinical tasks. Crucially, it was always assumed
by the respondents that only humans can master these
clinical skills. ey identiﬁed two main clinical skills as
being both crucial to the medical profession and unpos-
sessable by the AI.
First, it was argued that it is impossible to automate
clinical reasoning, but no clear reasons were provided for
why this is so:
So far, I do not believe that a system exists which can
fully replace the clinical reasoning process. (R16)
Fig. 2 Overarching theme 2 (above), themes (middle), subthemes (below)
4 However, most of them reacted very negatively towards this idea. We will
return to these negative reactions when reporting on the third overarching
Page 6 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
Some respondents, however, tried to explain which
aspects of clinical reasoning process render it diﬃcult to
be automated. ose respondents expressed the view that
it is diﬃcult, if not impossible, to reduce medical reason-
ing to a set of rules. All of them emphasized the unique
nature of every consultation. We will return to this view
when discussing the third overarching theme.
Others compared the AI to one of their colleagues.
Unlike the colleague, the AI is unable to explain how it
came to its decision. It is at most able to give additional
information, but this does not come close to a real dis-
cussion, as the respondents indicated they would have
with their human colleagues. ey all said that it was
necessary to understand the reasoning behind a sug-
gestion in order to accept it, but that the AI is unable to
explain its position:
[With people] you are able to ask why they suggested
something, they are able to give arguments for their
position and I am able to react to those arguments.
But here [with the AI] it ends with a suggestion. (R1)
Second, some respondents focused on certain practical
skills that are essential to the medical profession. Some,
for example, pointed out that interventional therapy, e.g.
surgery, still has to be administered by a human physi-
cian. Of interest, they recognized that also for human
physicians, diﬀerences in skills might determine which
type of intervention will lead to the best result, as some
physicians might have more experience with a particular
technique than with others:
Many roads lead to Rome. e AI only shows me one
road, but both me and my patient beneﬁt from the
road that I know best. (R20)
It was perceived by some respondents that AI does
not know which therapies the physician is most familiar
with. ey argued the AI will sometimes suggest certain
therapies, which it believes to be most eﬃcient or well-
researched, but which the physician on call does not
know or is not able to administer.
Many respondents argued that the physician should have
the ﬁnal responsibility in clinical practice. We use the
words ‘ﬁnal responsibility,’ because the respondents were
willing to delegate many of the physician’s tasks to the AI.
Yet, the physician should always have a supervising role
and, at least, every important decision should be made by
e reasons mentioned as to why the ﬁnal responsibil-
ity should stay with the physician diﬀered widely. None-
theless, they can be broadly divided into three main
categories. First, many respondents wanted to have ﬁnal
responsibility for psychological reasons. By ‘psychologi-
cal’ we mean that the respondent did not mention any
objective reasons, but simply wanted to stay in control:
I do not think the computer system should be
allowed to block you. I have my reasons to do what I
do and maybe I will think about its suggestions, but
I do not want [the] IT [department] to block me at
those moments. … I always want to do what I want.
Second, some respondents indicate that it is episte-
mologically important to stay in charge. e physician
should always know what is going on. is is closely
linked to the demand for more transparency because that
extra information would allow the physician to stay in
e more information they get [from the AI], the
more willing the physician will be to follow [the AI’s
is view diﬀers from the previous view, in which it
was stated that under no circumstance should the AI
override the physician’s own opinion. If the AI is able to
give adequate information on why a certain course of
action has to be taken, the respondent indicates she will
follow the suggestion.
Finally, some respondents emphasized that they
wanted to maintain their autonomy. ey wanted to be
the one to take the decision, because they felt otherwise
the AI would ignore their autonomy as medical profes-
sionals. Again, this is diﬀerent from the two previous,
because an objective reason is mentioned as to why the
physician should stay in charge:
e ﬁnal choice should always stay with the physi-
It is important to emphasize that these categories are
not mutually exclusive. A good example of this is the fol-
I do not want [the system] to order me around. (R1)
Respondent 1 clearly phrased his view in a way that
most clearly aligns with the psychological category, but
the subject matter most closely aligns with the autonomy
category. is may indicate that to some respondents the
lines between these diﬀerent motivations are less clear.
Finally, many respondents argued that their profession
should be ‘enjoyable’. ey often saw the AI as a poten-
tial negative inﬂuence on their day-to-day satisfaction at
the job. Generally they all wanted to avoid unnecessary
annoyance by the AI.
Page 7 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
A recurring theme with regard to frustration was the
respondents’ fear of alert fatigue. Alert fatigue refers to
a situation in which people are bombarded by so many
pop-ups and notiﬁcations that they become numb to
their eﬀect . is results in a lot of frustration. Typi-
cal expressions of this belief are: “Do not bother me with
unimportant news” (R5) or “I do not want 199 notiﬁca-
Finally, many respondents also made clear that they
worried that the introduction of AI in the workplace
would further obfuscate the separation of work and pri-
vate life. It was not simply that they thought the many
notiﬁcations would be frustrating, but that they believed
a clear separation of work and private life is a necessary
precondition for an enjoyable work environment:
e border between work and private life should not
become even more ﬂuid. (R23)
Overarching theme 3: concerns regardingAI
e ﬁrst two overarching themes concerned the role of
the AI and the physician in clinical practice. e third
overarching theme, however, focuses on criticisms of
AI that do not have anything to do with the division of
clinical labour.5 Importantly, some of these general con-
cerns relate to questions regarding the division of clini-
cal labour that we identiﬁed in relation to the ﬁrst and
second overarching themes. However, with regard to
this third main theme we will focus on the issues our
respondents had with the CDSS regardless of the tasks it
would be performing.
We have subdivided these criticisms according to
three perspectives. First, a purely technical perspective.
Second, a perspective that focuses on the relationship
between the user of the AI, in this case a physician, and
the AI. ird, the perspective of the relationship between
the AI and the kind of task it is being used for, in this case
medical tasks. (see Fig.3).
Respondents were often sceptical about a variety of
strictly technical aspects of the AI. We can divide these
concerns into four main themes: (1) data quality; (2)
accuracy of CDSS; (3) the interface of CDSS; and (4)
rigidity. First, many respondents expressed doubts about
the robustness of the data:
e statement [of the AI] is only as strong as the
data [used by the AI] (R2)
Some respondents were more speciﬁc about why the
data used by the AI could be inaccurate. It was argued
that the AI would only work if it possessed all the nec-
essary information regarding the patient, but that his-
torically physicians have not recorded all relevant pieces
of information and therefore their data would create an
inaccurate picture of reality.
In order for the AI to work, you have to feed it data.
But a ﬁle [on a patient made by a human physician]
never contains all [of the necessary] data (R16)
A second theme concerned many respondents’ wor-
ries regarding the AI’s accuracy in general. ey indi-
cated that they did not have any problem with using the
AI, as long as they could be sure the AI would not make
any mistakes. A more extreme version of this argument
was oﬀered by respondent 1. He argued that medicine, as
a domain of research, is unable to guarantee the level of
accuracy necessary for the AI to function properly.
Not only the system, but also medicine [as a ﬁeld
of study] has to have a certain level of accuracy [in
order for these AI to function properly]. (R1)
A third issue concerned the interface of the AI and the
way in which this interface could have a negative impact
on clinical practice. is was one of the most common
Fig. 3 Overarching theme 3 (above), themes (middle), subthemes
(below, ordered vertically)
5 In the second overarching theme we already anticipated some of these
criticisms, because claims that a role should be performed by a human often
rested on an assumption that the AI lacked the capability to perform this task.
Page 8 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
kinds of arguments given by the respondents. ese con-
cerns with the design varied widely from individual to
individual. Respondent 22 said, for example, that she pre-
fers a small amount of text, because otherwise she would
not read the notiﬁcation, whereas respondent 18 told us
that she prefers it when the suggestions of the AI always
ends with a question mark and respondent 14 told us that
when she sees a button, she wants to press it. e only
recurring view was that almost all respondents had a dis-
like for the use of red crosses to indicate that the physi-
cian had made a mistake:
e red [crosses] seem to make fun of you. I like the
other [options], [because] they present you with an
Finally, many respondents said they had concerns
regarding the perceived ‘rigidity’ of the AI. Because this
concern has to do both with the technical aspects of the
AI and the way in which it relates to medicine, we will
explain it below when discussing the latter.
Interaction betweenAI andhumans
is theme deals with concerns respondents had with the
interaction between AI and human users. ey broadly
had two concerns: (1) reinforcement of certain attitudes;
(2) habits of the human user and a fear of replacement by
First, many respondents argued that the AI could have
a reinforcing eﬀect. By reinforcement we understand a
process in which pre-existing human traits are encour-
aged and thus made more prominent by the AI. e
respondents referred to a variety of characteristics that
could be encouraged by the AI. Many respondents men-
tioned laziness in this regard:
I mostly believe the AI will have beneﬁcial eﬀects,
but it should not go too far. For example, alarms
and help with diagnosis seem like good things, but
it would be problematic if we would become too
dependent on them. e danger is that we would all
become very lazy and dependent. (R20)
Importantly, it is argued here that the AI would play
upon pre-existing laziness. By oﬀering quick diagnoses
and alerting physicians when to act, many respondents
fear that physicians would start to follow the AI unques-
tioningly, because that is the easy thing to do. Other
respondents argued that the AI could reinforce the physi-
cian’s pre-existing intuitions or beliefs and thereby create
a ‘tunnel vision’:
I ﬁnd it dangerous when the system would only give
you information about the illness you were already
thinking about. at way the system would rein-
force your ﬁrst thought instead of encouraging wider
reﬂection. … e AI must ﬁght against tunnel vision.
We already pointed out that some respondents believed
the AI would be able to completely replace the physician
and that most physicians were not enthusiastic about
this prospect. e second most common problem the
respondents identiﬁed with AI in relation to humans was
the belief that the AI would replace the human. Many
respondents at some point entertained the thought of the
AI, at some point, being able to completely automate the
tasks of the physicians, but always immediately observed
that this would be a very bad evolution.
When wondering about a future in which AI would be
able to fully automate the physician’s tasks, respondent
1 said: “In that case we [physicians] should just go home.”
When talking about a similar future, respondent 5 said:
“It is just like you [the physician] are a robot who needs
to press on some buttons.” Talking about the future of the
role of AI in medicine, respondent 20 argued: “You see
that they are trying to take the decisions out of our hands.”
ese reactions were often more emotional than their
other responses during the interview and they often used
less formal language.
Other respondents explicitly said that they do not want
a future in which the AI would fully take over their job.
Importantly, this refusal was not based on any criticism
on performance of the AI. Rather, they argued that even
if the AI was able to fully automate every aspect of the
physician’s job, they still would not want it to do so:
I think it would be very dangerous when the AI
would both conduct the diagnosis and prescribe the
therapy all by itself. is would bypass the knowl-
edge of the physician. We do not want the AI to do
everything, otherwise the physicians could just go
Of note, ‘dangerous,’ at ﬁrst glance seems to indicate a
criticism of certain consequences of the AI, but the next
sentence clariﬁes the respondent only criticizes the way
in which the AI would replace the physician, and that it
is this replacement in itself which is seen as dangerous.
Most respondents experienced the automation of diagno-
sis and therapy as being ‘a step too far.’6 An exception to
this rule was respondent 19, a surgeon, who believed that
6 Notice how these boundaries related to the distinction made by most
respondents between easy administrative and more complex clinical tasks.
From this we can conclude that there is a correlation between viewing a task
as being ‘automatable’ and believing a task to be fundamental to one’s job. We
will return to this topic in the Discussion.
Page 9 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
both diagnosis and therapy would become automated,
but the automation of surgery would be “impossible.”
Interaction betweenAI andmedicine
Some people also expressed concerns regarding AI spe-
ciﬁcally because of the way it relates to the domain of
medicine. is theme can be subdivided into two main
concerns: (1) rigidity; (2) the idea that medicine is a
It is believed by some that AI is too rigid, because it has
to work with well-deﬁned concepts that are measurable
and mutually exclusive. is is a technical issue, because
the respondents assumed this rigidity is the result of the
way in which AI functions on a technical level. Yet it is
also a problem in relation to clinical medicine, because
the respondents believed this rigidity was a problem spe-
ciﬁcally in the context of the medical profession:
e algorithm should always leave room for doubt.
In medicine it is always important to doubt. …
Our domain [medicine] is very hard to automate,
because it is diﬃcult to reduce it to well-deﬁned pat-
terns. With us there are way too many dimensions to
take into account. (R23)
is criticism closely resembles some of the views
expressed by the respondents in relation to the second
overarching theme, more speciﬁcally the argument that
AI would not be able to automate the clinical reasoning
process. ese respondents seem to imply that that clini-
cal reasoning is fundamentally non-rigid and needs to
leave room for doubt.
e last issue of concern is the idea that medicine is
diﬃcult, if not impossible, to automate, because it is a
‘unique domain.’ Some respondents just seemed to take
it as an assumption that medicine is fundamentally ‘dif-
ferent’ when it comes to automatability, but they do not
substantiate this claim:
In general I am a technological optimist, but not
when it comes to medicine. (R1)
e respondent does not make clear why medicine dif-
fers from other domains. It could be, because medicine is
hard to reduce to a set of rules, because the data used by
medical AI is rather bad or because medical AI generally
makes the job of the physician less enjoyable. Yet none of
these arguments are mentioned.
In general, whether the respondent saw a domain as
being ‘unique’ depended on whether they were familiar
with the domain. For instance, respondent 17, was more
lenient towards AI that would be used to automate jobs
she indicated she did not herself perform, and respond-
ent 19, a surgeon, often emphasized the extent to which
surgery is a ‘unique’ domain.
e results show that overall the respondents were will-
ing to let the CDSS take over many tasks that were tra-
ditionally seen as part of the physician’s repertoire. ey
believed that the CDSS will make their work safer, more
eﬃcient, that they will improve their own skills by learn-
ing from the system, and that the CDSS will automate
administrative and data processing tasks.
ey did, however, believe that there are limits to the
CDSS’ involvement in clinical practice. ese limits seem
to fall into three categories. First, the respondents men-
tioned some technical issues that need to be ﬁxed in
order for the CDSS to be ready for implementation: the
data need to be robust; the system needs to be accurate;
and the CDSS needs to be user-friendly. Second, they
believed that some aspects of clinical practice are inher-
ently unsusceptible to being automated. ey argued that
the CDSS is overly rigid or lacks certain clinical skills that
are necessary in order to perform the physician’s tasks.
ird, they indicated that they simply did not want spe-
ciﬁc tasks to be automated. In general, they did not give
any reasons for this.
Much of the literature has focused on the technical
issues with AI . Based on our results we would argue,
however, that these issues were not considered to be of
fundamental importance by our respondents. Even if
the AI would be easy to use and as accurate as possible,
they indicated that they would still dislike certain parts
of the physician’s job to be automated via AI. Our the-
matic analysis of a vignette based interview study reveals
that, next to concerns that have been identiﬁed in previ-
ous research, particularly regarding user-friendliness [21,
22] and transparency , more in-depth psychologi-
cal, epistemological and philosophical issues are at stake
when physicians reﬂect on the introduction of CDSS. In
this discussion, we would like to primarily focus on these
more fundamental issues, as it is impossible to explore
in-depth every theme we mentioned in the ‘Results’ sec-
The importance ofuser‑friendliness
Many arguments used by the respondents had some-
thing to do with the user-friendliness of the CDSS. e
respondents were mostly concerned with how the system
would aﬀect their general work experience. ey wanted
to avoid frustration and wanted the CDSS that was best
suited to complement their pre-existing work routines
Earlier research has already shown that one of the most
common complaints of physicians regarding CDSS is that
it is not made to suit their existing habits [24, 25]. Too
often the architecture of the CDSS is such that it causes
disruption of the clinical workﬂow, which results in it
Page 10 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
being disliked by many physicians. We believe that our
respondents’ emphasis on user-friendliness should be
viewed in a similar vein.
The division ofclinical labour
Respondents mostly did not use any purely ethical state-
ments. In general, they assigned a role either to them-
selves or to the AI and, based on this assumption, they
assessed which option best suited these preconceived
roles. e respondents were not directly concerned with
the ethical implications of the use itself of AI, but were
mostly preoccupied with questions concerning the role
of the AI in the medical profession and its relation to the
physician.7 Central to the respondents’ argumentation
was thus often a discussion on the division of medical
From the results it is also apparent that most respond-
ents adopted a dichotomy between easy, automatable
‘administrative tasks’ and diﬃcult, non-automatable
‘clinical tasks.’ Based on our interviews it is not possible
to clearly diﬀerentiate between these two categories, yet
we can make some general observations about the way in
which they were used.
First, the ‘medical tasks’ were implicitly assumed to
be the ‘core’ of the physician’s job, while the ‘administra-
tive tasks’ were seen as non-core tasks. Put diﬀerently,
the ‘clinical tasks’ are what makes a physician a physi-
cian, according to the respondents, while the ‘adminis-
trative tasks’ were assumed to be neither necessary nor
important elements of the deﬁnition of the physician’s
job. Furthermore, the kinds of tasks that were assigned
to each of the two categories depended on the speciﬁc
job performed by the respondent (i.e. their area of medi-
cine) and the respondent’s opinion of their job. Surgeons,
for example, assigned some tasks to the non-essential
category that non-surgeon respondents believed to be
essential to the physician’s job. Moreover, whether or not
psycho-social tasks were seen as ‘clinical’ or ‘administra-
tive’ varied between respondents.
is dichotomy is also very important with regard
to the fear of replacement expressed by most of the
respondents. Generally, the respondents only felt threat-
ened by the CDSS when it would start to be used to auto-
mate core tasks, whereas they were more positive and
even enthusiastic about the automation of non-essential
tasks. is supports our hypothesis that the division
between clinical and administrative tasks is based mostly
on personal experience of one’s job rather than on clear
criteria based or their knowledge of AI. is also implies
that simply taking care of some of the technical issues
with the CDSS will not necessarily improve acceptance
of it by physicians, because some core tasks are consid-
ered to be fundamentally oﬀ-limits when it comes to
Based on this we would argue that the introduction
of CDSS in the medical profession has a lot in common
with the introduction of automation in other professions.
Just like a factory worker has felt threatened by the exist-
ence of robots since the beginning of the twentieth cen-
tury, doctors feel worried by the introduction of AI. We
would like to hypothesize that the medical profession and
‘intellectual professions’ in general will experience a fear
of replacement similar to the fears experienced by work-
ers in manufacturing professions since the introduction
of automation.  erefore, when engaging in ethical
reﬂection about the implementation of CDSS, we should
not just consider the technical aspects of the systems in
question, but also understand this as a modern labour
dispute in which the physician could be seen as a threat-
ened economic actor.
Furthermore, we would submit that automation dis-
rupts the idea of what it means to be a physician. To
many physicians medicine is not simply a job they do to
earn money, but a vocation.  erefore, the encroach-
ment of AI does not just threaten the physician’s eco-
nomic status, but also their self-image and the way they
have chosen to spend a substantial part of their life. It is
an existential issue as much as an economic issue. ese
existential issues should not be treated lightly as bumps
on the road of progress, but should be taken seriously
when contemplating whether or not to automate certain
Claims concerningtheuniqueness ofmedicine
Related to the previous topic is the claim that medicine
is a ‘unique domain.’ For many respondents, this alleged
‘uniqueness’ of medicine was the main reason why it
is impossible to completely automate the job of the
As we have seen, many respondents simply stated this
uniqueness as a given fact. Some respondents, however,
argued that medicine is a special domain because it is
fundamentally ﬂexible, diverse or non-rigid. Put dif-
ferently, they were conveying that it is not possible to
reduce the practice of medicine to a certain set of ‘rules’
or to completely eradicate doubt and uncertainty. Many
respondents emphasized that every case is fundamen-
tally diﬀerent and that this is crucial to understanding the
complexities involved in clinical reasoning in general.
Further research should further explore and criti-
cally investigate the reasons underlying the respond-
ents’ claims about the uniqueness of medicine . For
7 is may explain why the respondents were not that interested in notions
that are central to ethical debates concerning AI, such as privacy.
Page 11 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
example, we believe there are certain important unique
aspects of medicine that were not mentioned explic-
itly by the respondents and that would merit in-depth
ethical, epistemological and political analysis. In their
paper ‘“Just Do Your Job”: Technology, Bureaucracy, and
the Eclipse of Conscience in Contemporary Medicine’,
physicians Blythe and Curlin argue that contemporary
medicine is too often understood according to a market
metaphor. Hospitals are seen as businesses that provide
a ‘service’ to customers. erefore, the physicians have
to become clogs in a large, anonymous, and bureaucratic
machine that produces indistinguishable medical prod-
ucts in accordance with the will of the ‘customer.’ .
Building on the seminal analysis of modernity and
modernisation by sociologist Max Weber in his essay
‘Science as a Vocation’  and its interpretation by
Berger, Berger and Keller , Blythe and Curlin con-
sider this new understanding of medicine to be a result
of the larger phenomenon of modernity spilling-over into
diﬀerent domains. Modernity is characterized by a ‘com-
ponential’ worldview, meaning that the world is under-
stood as a combination of atomised components. All of
these components are or should be interchangeable. In
modernity people understand the world according to the
principles of science and bureaucracy. Medicine too, they
argue, is now often understood as a science. However,
this is problematic, for:
[W]hile medicine is a practice that depends upon
scientiﬁc modes of reasoning and certain features
of the scientiﬁc consciousness, it is decidedly “not a
science”. Rather, “it is a rational, science-using, inter-
level, interpretive activity undertaken for the care of
a sick person. . pp. 439–440
One could argue that medicine is ‘unique’ in the sense
that it is a domain that heavily depends on science and
technology, but fundamentally is not a science. It is, ﬁrst
and foremost, a form of care. e introduction of CDSS
could be seen as the next step in the ‘modernization’ of
medicine and as a threat to medicine’s unique status as a
heavily science-using form of ‘care’.
The “nal responsibility” ofthephysician
Almost all respondents argued that the ﬁnal responsi-
bility8 for clinical actions should stay with the (human)
physician after the introduction of the CDSS in clinical
practice. Earlier research has shown that this freedom
or control is highly valued by users of CDSS.  Our
analysis suggests that taking on this ﬁnal responsibility is
seen as one of the core roles of the human physician. is
could be argued to be the main characteristic that sepa-
rates a human from an electronic physician.9
In order to qualify as ‘having ﬁnal responsibility’ the
physician needs to be able to do three diﬀerent things,
according to our respondents. First, the physician should
have the perception that (s)he is still in charge. Interest-
ingly, the physician’s perception of control is more impor-
tant than their actual control. Second, the physician
should have an idea of what (s)he is doing. By this we do
not mean that the physician must truly understand how
the CDSS came to a recommendation, but rather that
they want some general information about the CDSS’s
suggestion. ird and ﬁnally, the physician wants to be
able to overrule the system’s decision at all times. While
the ﬁrst and second criteria seem manageable to inte-
grate into the CDSS, the third criterion is highly demand-
ing. When implemented, the latter criterion would truly
ensure that the physician would stay in charge.
However, even this last criterion should be further
nuanced. Many respondents indicated that they preferred
a version of the CDSS that would only let people overrule
suggestions if they were able to provide a good reason for
why they were doing so. A good example of this view was
expressed by respondent 2:
My preferred option would be one in which the phy-
sician is able to overrule a suggestion by the AI, but
only when they give the reason why they did so in
order to avoid physicians making unsafe decisions.
e fact that this option was suggested by multiple
respondents indicates that even this proposed ‘right to
overrule’ is not as demanding as it might seem. e phy-
sician is willing to let the CDSS take over a lot of their
traditional tasks. When given the chance to overrule the
CDSS without any questions, they prefer the option in
which the physician is obliged to substantiate a reason
for wanting to overrule the CDSS. erefore, we can con-
clude that the respondents actually wanted a very limited
version of control to be guaranteed by the CDSS.
8 We have opted to use the words ‘ﬁnal responsibility’ instead of more com-
mon terms such as ‘autonomy’ or ‘control’ for multiple reasons. We believe
that both ‘control’ and ‘autonomy’ indicate that one wants a great level of free-
dom in relation to the AI. We want to emphasise, however, that the demands
of the respondents were actually rather limited. ey want a level of ‘respon-
sibility’, which we believe is not the same as control or autonomy, but only in
the ﬁnal stages of the decision process.
9 Although this emphasis on ﬁnal responsibility may reﬂect views regarding
legal liability , none of our respondents explicitly referred to legal issues
and therefore we do not want to make any general claims regarding this
issue. Although we consider ethical and legal questions concerning liability
in contexts involving the use of AI systems to be extremely important, these
questions lie beyond the scope of this paper.
Footnote 9 (continued)
Page 12 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
Furthermore, recent research has suggested that direct
human supervision, as suggested by our respondents, is
neither a necessary nor a suﬃcient criterion for being
meaningfully in control. In a ground-breaking paper by
Filippo Santoni de Sio and Jeroen van den Hoven, it is
argued that we need to abandon the notion that humans
will remain in direct control of autonomous systems in
the twenty-ﬁrst century. Nevertheless, they emphasise,
it is important that the abandonment of direct supervi-
sion should not result in an indiﬀerence towards ques-
tions surrounding responsibility and control. Rather,
they argue that we should strive towards “meaningful
human control” over (semi-)autonomous AI systems,
for reduced control may give rise to responsibility gaps
or accountability vacuums . In short, this principle of
meaningful human control implies that humans, rather
than computers and their algorithms, should ultimately
remain in control of and thus morally responsible for rel-
ey explain that the notion of meaningful human
control implies that “simple human presence” or “being
in the loop” is not suﬃcient, because “one can be present
and perfectly able to inﬂuence some parts of the system
by causal intervention, while (a) not being able to inﬂu-
ence other parts of the causal chains that could come
to be seen as even more relevant from a moral point of
view than the parts one can in fact inﬂuence, [and] (b)
not having enough information or options to inﬂuence
the process”. Moreover, according to this framework of
meaningful human control, “controlling in the sense
of being in the position of making a substantive causal
contribution to [an] activity through one’s intentional
actions might not be a suﬃcient condition for meaning-
ful control either, for instance, if one does not have the
psychological capacity to respond appropriately under
the circumstances and/or [one is] not in the position
to appreciate the real capabilities of the system [one is]
interacting with”. , p. 3.
Santoni de Sio and van den Hoven base their theory of
meaningful human control on a compatibilist account of
responsibility. Compatibilism refers to the philosophi-
cal belief that an actor can be in control of and responsi-
ble for an action even if she has not directly caused that
action. Compatibilists believe that it is suﬃcient to show
that the action was the result of a mental act by the actor
and that the actor would have been able to act diﬀerently.
An interesting way to understand responsibility within
this compatibilist framework is the concept of ‘guidance
control’ proposed by John Fisher and Mark Ravizza .
Fisher and Ravizza claim that two conditions have to be
met within a compatibilist theory in order for an actor to
be morally responsible for an action. First, the decisional
mechanism10 leading up to an action must be respon-
sive to moral or factual input. It must be possible for the
decision-making mechanism to adapt the behaviour of
the actor to the relevant moral features of the situation
at hand. If the actor was unable to avert the action, one
cannot convincingly argue that the actor is responsible
for the action in any meaningful way. Second, the actor
needs to take responsibility for the decisional mecha-
nism, meaning that the actor must be aware of the fac-
tual and moral impact of their actions.
Santoni de Sio and van den Hoven suggest that this
framework is interesting to understand what would be
required in order to talk of meaningful human control in
the context of (semi-)autonomous systems. ey identify
two conditions, similar to those proposed by Fisher and
Ravizza, which need to be met in order to have meaning-
ful human control over (semi-)autonomous systems.
e ﬁrst condition, tracking, refers to the idea that “a
decision-making system should demonstrably and veriﬁ-
ably be responsive to the human moral reasons relevant
in the circumstances… decision-making systems should
track (relevant) human moral reasons.” , p. 7. In prac-
tice, this would mean that autonomous systems would
have to be able to adjust their behaviour based on moral
or factual input. Moreover, establishing whose moral rea-
sons and which moral reasons are relevant in particular
circumstances means establishing which normative prin-
ciples, norms, and values the (semi-) autonomous system
is supposed to follow or reﬂect.
e second condition, tracing, implies that “in order
for a system to be under meaningful human control,
its actions/states should be traceable to a proper moral
understanding on the part of one or more relevant
human persons who design or interact with the system.”
, p. 9. us, as with the second condition of Fisher
and Ravizza, it is important that at least one person in the
design history or use context of the (semi-)autonomous
system is aware of the possible impact of the system and
the moral consequences of this impact. Otherwise, no
one can guarantee or make sure that the system will act
in accordance with moral principles.
If a (semi-)autonomous system acts in a context where
both requirements are not fulﬁlled the system cannot be
said to be under meaningful human control, according to
Santoni de Sio and van den Hoven. Crucially, according
to their framework, the direct supervision of a human, as
proposed by our respondents, is neither a necessary nor a
suﬃcient requirement for meaningful human control .
10 Decisional mechanism refers to the set of causal reactions by which an
action comes about. is mechanism can be mental, physical or digital. Com-
patibilists use this concept in order to highlight their view that free will is not
a necessary precondition for holding someone responsible for an action.
Page 13 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
We can conclude that there is a clear divergence
between this framework and the moral intuitions of our
respondents regarding responsibility and control. Indeed,
our respondents suggest a much less demanding view of
what counts as responsibility and control. Neither the
tracking nor the tracing-requirement have to be fulﬁlled
in our respondents’ view.
First, rather than programming the system to comply
with factual, ethical and legal concerns, our respondents
prefer the end-user, i.e. the physician, to have the right
to overrule the CDSS’ decisions. Furthermore, as we
have pointed out, many respondents argued that physi-
cians should not be allowed to overrule the AI when they
have good reasons to do so and provided that they inform
the AI of those reasons. Second, although our respond-
ents indicated that they value the ability of an AI system
to ‘explain’ its decisions to the physician, they did not
indicate that someone in the design history of the CDSS
needs to be aware of the potential factual and moral
implications of the system. Rather, they emphasized that
they want the end-user, i.e. the physician, to have the
ﬁnal responsibility for medical decisions.
In this paper we have reported the results of a thematic
analysis of twenty-four interviews with physicians. e
results of this analysis were categorised into three over-
arching themes: the perceived role of the AI; the per-
ceived role of the physician; and concerns regarding the
AI. Each of these three main themes was divided into
Based on these themes we elaborated four impor-
tant interpretations of the results. First, we argued that
the respondents focussed on the way in which the AI
would impact their everyday life and happiness. Second,
we discussed the way in which the respondents all cre-
ated a dichotomy between non-essential and core tasks
of a physician. We argued that this dichotomy was linked
to the likelihood of the physician in question being opti-
mistic or not towards the automation of a given task.
ird, we discussed the way in which many respondents
expressed the view that medicine is a ‘unique domain’.
Fourth, we explained how the desire for ﬁnal responsi-
bility was a central concern to many respondents. is
demand, however, should not be understood in a strongly
As hypothesized, these factors were the underlying driv-
ers of much of the discontent with the introduction of
CDSS, while more common factors took a backseat or were
seen as more trivial by the respondents. Although most of
these common factors could be addressed by technically
tweaking the CDSS, we believe that the underlying driv-
ers that we have identiﬁed show that our respondents have
fundamental issues with the automation of certain core
parts of their job—regardless of how well the CDSS may
perform, both from a technical and an ethical perspective.
erefore the acceptance of CDSS is not just a matter of
technical improvements, but would require genuine engage-
ment with and exploration of these underlying factors.
Based on our analysis we should like to make two rec-
ommendations for further research. First, we believe that
the introduction of AI in clinical medicine should not
just be studied from an ethical or a technical perspective.
Indeed, our research has shown that there are important
economic, social, and existential aspects to this techni-
cal transition. ‘Economic’ in the sense that a physician is
an economic actor who feels threatened by the prospect
of automation and whose economic interests should be
taken seriously. ‘Social’ in the sense that physicians do
not work in a vacuum and the social aspects of their job
are important to them. ‘Existential’ in the sense that phy-
sicians are human beings who value the job they do and
who want to do meaningful work.
Second, further research is needed regarding the
‘unique status’ that most respondents ascribed to the
medical ﬁeld. It would be interesting to explore and
critically investigate the reasons underlying the respond-
ents’ claims about the uniqueness of medicine, not only
through the lens of ethics but also from an epistemo-
logical and political perspective. Such explorations could
shed more light on the question as to whether or not clin-
ical medicine really is ‘uniquely unsuited’ to being auto-
mated. We believe that both the highly variable nature of
clinical problems, as emphasised by most of our respond-
ents, and the reconceptualization of medicine as a form
of ‘rational care’ in line with Blythe and Curlin  could
be interesting perspectives to these avenues of research.
The online version contains supplementary material available at https:// doi.
org/ 10. 1186/ s12910- 022- 00787-8.
Additional le1. Qsort vignettes.
DVC performed the ﬁrst thematic analysis together with SS, took part in the
triangulation sessions, and re-analysed the categorisation of the themes in
the second round. He also produced the ﬁrst draft, and integrated comments
of other authors in several rounds of revision of the draft paper. WVB has con-
ceived the concept of the study, helped design the vignettes, and performed
all the interviews. He took part in the triangulation sessions and has written
the ﬁrst draft of the introduction and methods sections. JDC contributed to
the conceptualisation of the project, helped design the vignettes, took part in
the triangulation sessions and reviewed and commented on the diﬀerent con-
secutive draft versions of the paper. TL assisted with the ﬁrst thematic analysis,
took part in the triangulation sessions, designed the ﬁgures, and reviewed
and commented on diﬀerent consecutive versions of the draft paper. She also
Page 14 of 14
VanCauwenbergeetal. BMC Medical Ethic s (2022) 23:50
organized the logistics of the interviews. SS contributed to the concept of the
study, helped design the vignettes. She contributed to the thematic analysis
with DVC, and re-analysed the categorisation of the themes in the second
round. She took part in the triangulation sessions. She provided several rounds
of comments and suggestions on the diﬀerent consecutive draft versions of
the paper. All authors read and approved the ﬁnal version of the manuscript.
This work is part of a larger project funded by Fonds Wetenschappelijk Onder-
zoek (FWO): FWO.OPR.2019.0045.01.
Availability of data and materials
All data are supplied as Additional ﬁle 1. The precise content of the vignettes is
available as Additional ﬁle 1: Qsort vignettes.
Ethics approval and consent to participate
This study was approved by the Ethical Committee of the Ghent University
Hospital (number BC 08294). All participants provided written informed con-
sent. All research and work was performed in accordance with the relevant
guidelines and regulations. We adhered to the Consolidated criteria for report-
ing qualitative research (COREQ) regulations throughout the process.
Consent for publication
None of the authors has competing interests with regard to the content of
this publication. The Authors therefore declare no competing ﬁnancial or
1 Department of Philosophy and Moral Sciences, Bioethics Institute Ghent,
Ghent University, Ghent, Belgium. 2 Consortium for Justiﬁable Digital Health-
care, Ghent University Hospital, Ghent, Belgium. 3 Department of Nephrology,
Ghent University Hospital, Ghent, Belgium. 4 Department of Intensive Care
Medicine, Ghent University Hospital, Ghent, Belgium.
Received: 2 February 2022 Accepted: 20 April 2022
1. Topol EJ. High-performance medicine: the convergence of human and
artiﬁcial intelligence. Nat Med. 2019;25(1):44–56.
2. Beam AL, Manrai AK, Ghassemi M. Challenges to the reproducibility of
machine learning models in health care. JAMA. 2020;323(4):305.
3. Eslami S, de Keizer NF, Abu-Hanna A. The impact of computerized physician
medication order entry in hospitalized patients—a systematic review. Int J
Med Inform. 2008;77(6):365–76.
4. Black AD, Car J, Pagliari C, Anandan C, Cresswell K, Bokun T, et al. The impact
of eHealth on the quality and safety of health care: a systematic overview.
PLoS Med. 2011;8(1):e1000387.
5. Brocklehurst P, Field D, Greene K, Juszczak E, Kenyon S, Linsell L, et al. Com-
puterised interpretation of the fetal heart rate during labour: a randomised
controlled trial (INFANT). Health Technol Assess. 2018;22(9):1–186.
6. Fry E, Schulte F. D eath by a thousand clicks: where electronic health records
went wrong [Internet]. Fortune. [cited 2021 Dec 28]. https:// fortu ne. com/
longf orm/ medic al- recor ds/.
7. Loper PL. The electronic health record and acquired physician autism. JAMA
8. Toll E. The cost of technology. JAMA [Internet]. 2012. https:// doi. org/ 10.
1001/ jama. 2012. 4946.
9. Wachter RM. The digital doctor: hope, hype, and harm at the dawn of medi-
cine’s computer age. New York: McGraw-Hill Education; 2015. p. 330.
10. Gawande A. Why doctors hate their computers. The New Yorker [Internet].
2018. https:// www. newyo rker. com/ magaz ine/ 2018/ 11/ 12/ why- docto rs-
hate- their- compu ters.
11. Kawamoto K, Houlihan CA, Balas EA, Lobach DF. Improving clinical practice
using clinical decision support systems: a systematic review of trials to
identify features critical to success. BMJ. 2005;330(7494):765.
12. Castillo RS, Kelemen A. Considerations for a successful clinical decision sup-
port system. CIN Comput Inform Nurs. 2013;31(7):319–26.
13. Gaube S, Suresh H, Raue M, Merritt A, Berkowitz SJ, Lermer E, et al. Do as
AI say: susceptibility in deployment of clinical decision-aids. npj Digit Med.
14. van der Sijs H, Aarts J, Vulto A, Berg M. Overriding of drug safety alerts
in computerized physician order entry. J Am Med Inform Assoc.
15. Taichman DB, Williams SV, Laine C. Clinical decision making: nurturing our
core skills. Ann Intern Med. 2020;172(9):627.
16. Akhtar-Danesh N, Baumann A, Cordingley L. Q-methodology in nursing
research: a promising method for the study of subjectivity. West J Nurs Res.
17. Bachmann LM, Mühleisen A, Bock A, ter Riet G, Held U, Kessels AG. Vignette
studies of medical choice and judgement to study caregivers’ medical deci-
sion behaviour: systematic review. BMC Med Res Methodol. 2008;8(1):50.
18. Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol.
19. Kempt H, Nagel SK. Responsibility, second opinions and peer-disagreement:
ethical and epistemological challenges of using AI in clinical diagnostic
contexts. J Med Ethics. 2021;48:222–9.
20. Middleton B, Sittig DF, Wright A. Clinical decision support: a 25 year retro-
spective and a 25 year vision. Yearb Med Inform. 2016;25(S01):S103–16.
21. Horsky J, Schiﬀ GD, Johnston D, Mercincavage L, Bell D, Middleton B.
Interface design principles for usable decision support: a targeted review
of best practices for clinical prescribing interventions. J Biomed Inform.
22. Payne TH, Hines LE, Chan RC, Hartman S, Kapusnik-Uner J, Russ AL, et al.
Recommendations to improve the usability of drug-drug interaction clinical
decision support alerts. J Am Med Inform Assoc. 2015;22(6):1243–50.
23. Durán JM, Jongsma KR. Who is afraid of black box algorithms? On the episte-
mological and ethical basis of trust in medical AI. J Med Ethics. 2021;47:329–35.
24. Ford E, Edelman N, Somers L, Shrewsbury D, Lopez Levy M, van Marwijk H,
et al. Barriers and facilitators to the adoption of electronic clinical decision
support systems: a qualitative interview study with UK general practitioners.
BMC Med Inform Decis Mak. 2021;21(1):193.
25. Zicari RV, Ahmed S, Amann J, Braun SA, Brodersen J, Bruneault F, et al. Co-
design of a trustworthy AI system in healthcare: deep learning based skin
lesion classiﬁer. Front Hum Dyn. 2021;13(3):688152.
26. Ford M. Rise of the robots: technology and the threat of a jobless future.
New Yor k: Basic Books; 2016. p. 334.
27. Blythe JA, Curlin FA. “Just do your job”: technology, bureaucracy, and the
eclipse of conscience in contemporary medicine. Theor Med Bioeth.
28. de Melo-Martín I. Vulnerability and ethics: considering our Cartesian hango-
ver. The Lancet. 2009;373(9671):1244–5.
29. Weber M, Owen DS, Strong TB, Livingstone R, Weber M, Weber M. The voca-
tion lectures. Indianapolis: Hackett Publishing; 2004. p. 100.
30. Berger PL, Berger B, Kellner H. The homeless mind: modernization and
consciousness. New York : Vintage Books; 1974. p. 258.
31. Sullivan HR, Schweikart SJ. Are current tort liability doctrines adequate for
addressing injury caused by AI? AMA J Ethics. 2019;21(2):E160–6.
32. Santoni de Sio F, van den Hoven J. Meaningful human control over autono-
mous systems: a philosophical account. Front Robot AI. 2018;5:15.
33. Fischer JM, Ravizza M. Responsibility and control: a theory of moral respon-
sibility. Cambridge studies in philosophy and law. Cambridge: Cambridge
University Press; 2000. p. 277.
34. Umbrello S. Meaningful human control over smart home systems: a value
sensitive design approach. J Philos Stud. 2020;13(37):40–65.
Springer Nature remains neutral with regard to jurisdictional claims in pub-
lished maps and institutional aﬃliations.
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
use bots or other automated methods to access the content or redirect messages
override any security feature or exclusionary protocol; or
share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at