ChapterPDF Available

Abstract and Figures

In this paper, we present the current state-of-the-art of decision making (DM) and machine learning (ML) and bridge the two research domains to create an integrated approach of complex problem solving based on human and computational agents. We present a novel classification of ML, emphasizing the human-in-the-loop in interactive ML (iML) and more specific on collaborative interactive ML (ciML), which we understand as a deep integrated version of iML, where humans and algorithms work hand in hand to solve complex problems. Both humans and computers have specific strengths and weaknesses and integrating humans into machine learning processes might be a very efficient way for tackling problems. This approach bears immense research potential for various domains, e.g., in health informatics or in industrial applications. We outline open questions and name future challenges that have to be addressed by the research community to enable the use of collaborative interactive machine learning for problem solving in a large scale.
Content may be subject to copyright.
Reasoning Under Uncertainty: Towards
Collaborative Interactive Machine Learning
Sebastian Robert1(B
), Sebastian B¨uttner2, Carsten R¨ocker1,2,3,
and Andreas Holzinger3
1Fraunhofer-Institute of Optronics, System Technologies and Image Exploitation,
Application Center Industrial Automation (IOSB-INA), Lemgo, Germany
sebastian.robert@iosb-ina.fraunhofer.de
2Ostwestfalen-Lippe University of Applied Sciences, Lemgo, Germany
{sebastian.buettner,carsten.roecker}@hs-owl.de
3Holzinger Group, HCI-KDD, Institute for Medical Informatics,
Statistics and Documentation, Medical University Graz, Graz, Austria
a.holzinger@hci-kdd.org
Abstract. In this paper, we present the current state-of-the-art of deci-
sion making (DM) and machine learning (ML) and bridge the two research
domains to create an integrated approach of complex problem solving
based on human and computational agents. We present a novel classifica-
tion of ML, emphasizing the human-in-the-loop in interactive ML (iML)
and more specific on collaborative interactive ML (ciML), which we under-
stand as a deep integrated version of iML, where humans and algorithms
work hand in hand to solve complex problems. Both humans and com-
puters have specific strengths and weaknesses and integrating humans
into machine learning processes might be a very efficient way for tack-
ling problems. This approach bears immense research potential for various
domains, e.g., in health informatics or in industrial applications. We out-
line open questions and name future challenges that have to be addressed
by the research community to enable the use of collaborative interactive
machine learning for problem solving in a large scale.
Keywords: Decision making ·Reasoning ·Interactive machine learn-
ing ·Collaborative interactive machine learning
1 Introduction and Motivation
Disregarding the application domain, i.e. whether in the medical domain or in
the industrial context, current developments such as the rapidly growing com-
munication infrastructure, the internet of things and increasing processing power
with services and applications on top of those lead to massive amounts of data
and new possibilities. Traditional analytic tools are not well suited to capturing
the full value of “big data”. Instead ML is ideal for exploiting opportunities hid-
den in data. Highly complex small batch production and personalized medicine
(precision medicine [1]) are two of many possible target scenarios. Both depend
c
Springer International Publishing AG 2016
A. Holzinger (Ed.): ML for Health Informatics, LNAI 9605, pp. 357–376, 2016.
DOI: 10.1007/978-3-319-50478-0 18
358 S. Robert et al.
on computer-intensive data processing prior to its analysis and decision making
processes.
However, to handle and exploit the required data, besides computer algo-
rithms, human capabilities are strongly needed as well. For example, classical
logic in ML approaches permits only exact reasoning, i.e. if A is true then A
is non-false and if B is false then B is non-true. However, even though modern
sophisticated automatic ML approaches can hardly cope with such situations,
human agents can deal with such deficiencies.
Moreover, many ML approaches are based on normative models such as for-
mal probability theory and expected utility (EU) theory. EU theory accounts for
decision under uncertainty and is based on axioms of rational behavior described
by von Neumann and Morgenstern (1944) [2]. Based upon the fact that informa-
tion available in daily problem solving situations is most of the time imperfect,
imprecise and uncertain due to time pressure, disturbance of unknown factors
or randomness outcome of some attributes [3,4], the interaction between human
and computer has to be designed in an optimal way in order to realize the best
possible output. Given that, a combined approach of human and computer input
can be a sustainable approach for effectively revealing structural or temporal
patterns (“knowledge”) and make them accessible for decision making.
At this point, decision theory comes into play and helps us to deal with
bounded rationality and the problem of which questions to pose to human
experts and how to ask those questions. Therefore, new types of human-computer
interaction (HCI) will arise and shape the ecosystem of human, technology and
organization. In particular, adaptive decision support systems that help humans
to solve complex problems and make far-reaching decisions will play a central
role in future work places.
In this paper, we will focus on decision making under uncertainty and bridge
it to ML research and particularly to interactive ML. After discussing the state-
of-the-art in ML and decision making under uncertainty, we provide some prac-
tical aspects for the integration of both approaches. Finally, we discuss some
open questions and outline future research avenues.
2 Glossary and Key Terms
Bias refers to a systematic pattern of deviation from rationality in decision
making processes.
Bounded Rationality – introduced by Herbert A. Simon [5]–isusedtodenote
the type of rationality that people resort to when the environment in which they
operate is too complex relative to their limited mental abilities [6].
Decision Support Systems (DSS) are intended to assist decision makers in taking
full advantage of available information and are a central part of health informat-
ics [7] and industrial applications [8].
Decision Theory is concerned with goal-directed behaviour in the presence of
options [9]. While normative decision theory focuses on identifying the opti-
mal decision to make, assuming a fully-rational decision maker who is able to
Reasoning Under Uncertainty 359
compute with perfect accuracy, descriptive decision theory deals with questions
pertaining to how people actually behave in given choice situations. Prescriptive
decision theory is the logic consequence and tries to exploit some of the logical
consequences of normative theories and empirical findings of descriptive studies
to make better choices [10].
Expected Utility (EU) Theory consists of four axioms, that define a rational deci-
sion maker: completeness, transitivity, independence, and continuity; if those are
satisfied, then the decision making is considered to be rational and the pref-
erences can be represented by a utility function, i.e. one can assign numbers
(utilities) [2].
Heuristics describe approaches to problem solving and decision making which
are not perfect, but sufficient for reaching immediate goals [11].
Human-Computer Interaction (HCI) is a multi-disciplinary research field that
deals with “the design, implementation and evaluation of interactive systems in
the context of the user’s task and work”[12, p. 4]. It can be located at the
intersection of psychology and cognitive science, ergonomics, computer science
and engineering, business, design, technical writing and other fields [12, p. 4].
Judgment and Decision Making (JDM) is a descriptive field of research which
focuses on understanding decision processes on an individual and group level.
Machine Learning (ML) is a research field grounding in computer science that
concentrates on induction algorithms and on other algorithms that can be said
to ‘learn”[13]. While in automatic Machine Learning (aML) representations of
real-world objects and knowledge are automatically generated from data, inter-
active Machine Learning (iML) methods allow humans to interact with com-
puters in some way to generate knowledge and find an optimal solution for a
problem. More specifically, collaborative interactive Machine Learning (ciML)
is a form of iML, where at least one human is integrated into the algorithm
using a specific user interface that allows manipulating the algorithm and its
intermediate steps to find a good solution in a short time.
Perception-Based Classification (PBC) is a classification of data done by humans
based on their visual perception. In the context of ML, PBC has been introduced
byAnkerstetal.[14] who enabled users to interactively create decision trees.
PBC can be seen as one possible way of realizing iML.
Utility Theorem describes that a decision-maker faced with probabilistic (partic-
ularly when probabilities are distorted or unknown) outcomes of different choices
will behave as if she/he is maximizing the expected value [15]; this is the basis
for the expected utility theory.
3 State-of-the-Art
In this section, we will provide an overview of the current research regarding
two fields: First, we will investigate machine learning (ML) and focus especially
360 S. Robert et al.
on the advances in interactive machine learning (iML). Second, we will provide
an overview of the research on JDM under uncertainty. We will further focus on
bridging the research on human decision making and the research on iML. We
will motivate, why the knowledge of and research on human decision-making is
key for the development of future human-oriented ciML systems.
3.1 Machine Learning (ML)
ML is a very practical field with many application areas, though at the same
time well grounded theories with many open research challenges exist. There
are many various definitions, depending on whom to ask; a Bayesian will give
a different answer than a Symbolist [16]; a classical definition is close to and
grounding in computer science that “concentrates on induction algorithms and
on other algorithms that can be said to ‘learn”[13]. This definition is at the same
time the goal of ML which concentrates on the development of “programs that
automatically improve with experience”[17]. Advances in ML have solved many
practical problems, e.g., recognizing speech [18], giving movie recommendations
based on personal references [19] or driving a vehicle autonomously [20].
In the following, we will differentiate between classical ML approaches, that
we will call aML and the newer concepts of iML.
Automatic Machine Learning (aML): Methods and algorithms of machine
learning are often categorized as follows (here the classification of Marsland
[21]):
With supervised learning methods, an algorithm creates a general model from
a training set of examples containing input and output data (targets). With
this model, the output of new unknown input can be predicted.
– Contrary, when using unsupervised learning methods, the output data are
not provided to the algorithm. The algorithm focuses on finding similarities
between a set of input data and classifies the data into categories.
Reinforcement learning is somehow between supervised and unsupervised
learning. It characterizes algorithms that receive feedback, in the case that
their created output data are wrong. By this feedback the algorithm can
explore possibilities and iteratively find better models, respectively outputs.
Finally evolutionary learning methods develop models iteratively by receiving
an assessment of the quality (fitness) of the current model. As the term depicts,
this learning method is inspired by the biological evolution.
The mentioned methods and algorithms all have in common, that they – once
started – run automatically. We therefore call those classical machine learning
methods automatic machine learning. When using aML methods, human involve-
ment is in general very limited and restricted to the following three aspects:
Humans have to prepare the data and remove corrupt or wrong data sets from
the input data (data cleansing).
Reasoning Under Uncertainty 361
When using supervised learning methods, humans are responsible for provid-
ing the output data, e.g., for labeling data in classification tasks.
Another user involvement is the assessment of a certain model and the eval-
uation. Humans can assess the generated model and its results, and decide,
whether a certain model is able to produce good predictions or not.
The traditional approach does not put much emphasize on the human inter-
actions with the ML system. Humans are somehow involved in providing the data
as described above, but the early ML research mostly neglects the question, how
humans can provide data and how they deal with an inaccurate model. From
a practical perspective, this is a huge restriction in automatic machine learning
(aML) systems. The main problems of practical ML applications are often not
the implementation of the algorithm itself, but rather the data acquisition and
cleansing. Often data are corrupt or of bad quality and in most cases data do
not cover all required context information to solve a specific problem [3,4].
Interactive Machine Learning (iML): Compared to aML, iML is a relative
new approach that also considers the human involvement and interactions in
ML and aims at putting the human into the loop of machine learning. In this
section, we will discuss the approaches and concepts that previously have been
described under the term iML. We will distinguish in this section between three
types of iML methods: First, early works in the iML research considered iML
as an alternative way of ML where humans accomplish the model generation,
which basically means that humans replace algorithms. Second, concepts have
been proposed under the term iML that put a human into the training-evaluation
loop, but still execute algorithms automatically. Contrary to aML in this type of
iML algorithms have to be much faster to give rapid feedback to a user. Third,
humans can work hand in hand with algorithms to create a certain model, which
we consider as the most promising concept of iML with the best integration of
users and algorithms.
Humans replacing algorithms: Early work in iML has been done by Ankerst et al.
[14]. They implemented a system called perception-based classification (PBC)
that provides users the means to interactively create decision trees by visual-
izing the training data in a suitable way. By interacting with the visualized
training data, users select attributes and split points to construct the decision
trees. The system cannot automatically generate the tree. Instead, the user of
the system replaces the algorithm and creates the tree manually with the inter-
active application provided. According to their evaluation, the system reaches
the same accuracy as algorithmic classifiers but the human-generated decision
trees have a smaller tree size, which is beneficial in terms of understandability.
Another advantage of the interactive and manual approach is the possibility
of backtracking in case of a suboptimal subtree – a situation that humans can
easily recognize [14]. A huge benefit of this human-centered approach is the inte-
gration of the users’ domain knowledge into the decision tree construction [22].
362 S. Robert et al.
Building on the work of Ankerst et al., Ware et al. [23] developed a similar sys-
tem that replaces the algorithm with users. Their work focuses mainly on an
empirical evaluation of the performance of humans compared to state-of-the-art
algorithms. According to their study, novice users can build trees that are as
accurate as the ones provided from algorithms, but similar to Ankerst et al.
they found, that the tree size is decreased, when humans generate the decision
trees. On the other hand, Ware et al. point out that this manual iML approach
might not be suitable for large data sets and high-dimensional data. This early
variant of interactive machine learning is shown in Fig. 1A.
Humans in the training-evaluation loop: Another variety of iML is the integra-
tion of humans into the training-evaluation loop, when using supervised learning
methods. Fails and Olsen [24] were one of the first, who used the term iML and
proposed this integration for the rapid development of models, if the feature
selection cannot be done by domain-experts due to missing knowledge. They
give an example of the use of iML for the rapid development of perceptual user
interfaces (PUIs), that are developed by interaction designers who are usually
not familiar with computer vision (CV) algorithms. For this purpose, they pro-
vide a tool that gives designers rapid visual feedback of the produced classifiers
and the iterative changes of the selected features for the model generation. The
tool masks the complexity of the feature selection and rather allows users to
assess the output of the model generation and to drive the feature selection
into the right direction. A similar concept has been described by Fiebrink et al.
[25]. They developed Wekinator1, a system that analyses human gestures in the
context of music making. A graphical user interface supports users with the cre-
ation of appropriate training data, the configuration of various ML algorithms
and parameters and allows a real-time evaluation of the trained model by giving
visual or auditory feedback. This real-time evaluation allows a domain user to
rapidly adapt the input data to improve the model. Fogarty et al. [26]presented
CueFlik, a similar iML tool for generating models for image classification tasks.
For the mentioned type of iML, it is essential to have algorithms that have a
very short learning time to be able to give rapid feedback on the results [24].
Addressing this particular aspect in connection with big data, Simard et al. [27]
described a system that is very generic in terms of the data types and tasks and
interactive even when using big data. Their system called ICE (interactive clas-
sification and extraction) allows users to interactively build models consisting
of several millions of items. In [28] they extend their approach and addition-
ally deliver feedback about the performance of the generated model to the user.
With this system they empower users to not only optimize the model in terms
of accuracy, but to optimize in terms of performance as well. While the men-
tioned systems use only one model, in recent years model ensembles became the
standard of ML [16]. Talbort et al. therefore provide a tool that deals with mul-
tiple models and allows users to interactively build combination models [29]. All
mentioned publications in this section use the term iML to describe a concept,
1http://www.wekinator.org/.
Reasoning Under Uncertainty 363
Fig. 1. Classification of interactive machine learning (iML). A: Early iML research
aimed at replacing algorithms and using human pattern recognition capabilities instead.
B: Later iML methods have been proposed that provide a rapid feedback cycle to users.
Models are generated in a very short time and presented to users. Based on the pre-
sented model, users can adapt the input data and rerun the machine learning algorithm.
With this approach the model is iteratively improved. C: Using collaborative interac-
tive machine learning (ciML) humans can manipulate an algorithm during runtime
and improve the model while it is generated. Human and computational agents work
collaboratively on a specific problem.
where humans are in the training-evaluation loop, but cannot interfere with the
algorithm itself – from a human perspective the algorithm is a black-box. The
method of putting humans into the training-evaluation loop is shown in Fig. 1B.
Humans collaborating with algorithms: Sinard et al. define iML as a ML scenario,
where “the teacher can provide [...] information to the machine as the learning
ta sk p rogre sse s ”[27]. De facto, most systems presented in the past realized this
iML by providing means to users to evaluate a certain model and by changing
the training data to optimize the previously generated model. In this section,
364 S. Robert et al.
we present work that goes even one step further and integrates humans into
the process by providing a user interface that allows humans to manipulate the
parameters of the algorithm during its execution. We will call this approach col-
laborative interactive machine learning (ciML). Im this approach, humans can
directly collaborate with an algorithm. With this deep integration, new possi-
bilities of human-computer collaboration in ML might rise. One of the earliest
works, that aimed at the collaboration between human and algorithm in a ML
scenario has been presented by Ankerst et al. [30]. They built up on their earlier
PBCsystem[14] and provide an iML system for building decision trees for a
classification task. While their earlier PBC system only visualized data and left
the decision tree building to the users, algorithms are now integrated into the
system that might (but does not have to) be used. With the options provided,
different types of cooperation can be realized: manual (equivalent to the earlier
PBC), combined or completely automatic model generation. For the decision
tree construction, the system supports with proposing splits, with visualizing
hypothetical splits – up to a defined number of levels (“look-ahead function”),
and with the feature of automatically expanding subtrees. One mentioned goal of
their work is the use of human pattern recognition capabilities in the interactive
decision tree construction by still using algorithmic operations to allow dealing
with huge data sets [30]. Along these lines, Holzinger defines iML as “algorithms
that can interact with agents and can optimize their learning behavior through
these interactions, where the agents can also be human”[31], consequently, he
considers iML as this deeply integrated type of a collaboration between algo-
rithm and human. He discusses another issue that can be addressed with this
deeply integrated form of iML: Sometimes ML needs to deal with rare events,
like occurrences of rare diseases in health informatics, and consequently adequate
training data are missing. He identifies new application areas for ciML within the
health domain, e.g. for subspace clustering, protein folding, or k-anonymization
of patient data and names challenges for the future ciML research. Holzinger
also shows that the solution of complex problems is possible by using ciML. He
presents the integration of users into an ant colony algorithm to solve a traveling
salesman problem (TSP) [32]. A visualization shows the pheromone tracks of the
ants in the TSP and the optimal round-trip found by the algorithm so far. Users
can select edges and add or remove the current amount of pheromones on the
edge between each of the iterations. First experiments show that the process is
sped up in terms of required iterations to find the optimal solution [32]. The
collaborative variant of interactive machine learning is shown in Fig.1C. As the
related work regarding the collaboration between humans and algorithms in iML
shows, there has not been done a lot of research investigating the challenges and
opportunities of a human-algorithm interaction. Application areas of this new
iML approach need to be further identified and the implications of a human
agent in the iML system need to be explored. While humans can bring tacit
knowledge and context information into the process of building models, the
question remains unclear how human decisions effect the output of the iML sys-
tem. However, there has been a lot of research regarding human-decision making
that we will introduce in the next section.
Reasoning Under Uncertainty 365
3.2 Judgement and Decision Research
Generally, the main focus of ML is on dealing with uncertainty and making pre-
dictions. In order to infer unknowns, data sets have to be learned and analysed.
Therefore, most ML approaches are based on normative models such as formal
probability theory and EU theory. EU theory accounts for decisions under uncer-
tainty and is based on axioms of rational behavior, codified by von Neumann and
Morgenstern [2]. It states that the overall utility of an option equals the expected
utility, calculated by multiplying the utility and probability of each outcome [33,
p. 24]. Probability theory in ML is most often used in terms of Bayesian decision
theory [3437], which is build on EU theory as a framework for solving problems
under uncertainty [38, p. 140]. “Individuals who follow these theories are said to
be rational” [39, p. 724].
The successful integration of knowledge of a domain expert in the black-box
as discussed in the iML approach stands or falls with the careful consideration
of people’s actual decision making abilities. It is generally accepted that human
reasoning and decision making abilities can exhibit various shortcomings when
compared with mathematical logic [3]. Hence, the question that arises is, how to
integrate human and computer input, accounting for the imperfections of both
[40, p. 2122]. At this point descriptive decision theory can offer useful insights
for the optimal integration of human judgement in iML approaches.
Descriptive decision theory deals with questions pertaining to how people
behave in given choice situations and what we need to fully predict their behav-
iour in such situations [41, p. 2]. In many cases, this is a difficult task due
to given inconsistencies in people’s choices. These inconsistencies can often be
attributed to irrational behaviour or accidental errors, which can also lead to
deficient decisions [41, p. 6].
Within the last decades, a growing research community within the area of
descriptive decision making is focusing on understanding individual and group
judgement and decision making (JDM) [42,43].2Researchers from various fields
are actively contributing to JDM, e.g. cognitive psychologists, social psycholo-
gists, statisticians and economists [42,45]. They have developed a detailed pic-
ture of the ways in which individuals judgement is bounded [46], e.g., people
violate the axions of EU theory and do not always follow basic principles of
calculus [47,48]. JDM tasks are characterized by uncertainty and/or by a con-
cern for individual’s preferences and will therefore apply to central aspects of
human activities in iML [38, p. 140]. In detail, JDM research focuses on how
different factors (e.g., information visualization) affect decision quality and how
it can be improved [49,50]. In order to give any predictions about human judge-
ment, JDM usually presupposes a definition of rationality that makes certain
actions measurable. This instrumental view of rationality only accords with nor-
mative theory if keeping in line with it helps to attain satisfaction – measured
in subjective utility [51]. A basic approach of JDM is to compare actual judge-
ments to normative models and look for deviations. These so called biases are
2See also [44] for the chapter.
366 S. Robert et al.
the starting point for building models that explain and predict human decision
making behaviour. A fundamental outcome of early JDM research reveals that
the typical model of a “rational man” as presumed by most normative theories –
considering every possible action, every outcome in every possible state and cal-
culating the choice that would lead to the best outcome – is unrealistic and does
not exist [5]. Instead innumerous studies revealed that people cannot carry out
the complex and time-consuming calculations necessary to determine the ideal
choice out of possible actions [52, p. 7]. Instead people act as “satisficers” and
make decisions on the basis of limited information, cognitive limitations and
the time available. Simon’s concept of bounded rationality describes how people
actually reach a judgement or a decision and has become a widely used model
for human decision behaviour [5].
Building on Simon’s model, Tversky and Kahneman developed their heuris-
tics and biases program that fundamentally shaped our understanding of judg-
ment as we know it today [48]. According to their argumentation, coming to a
decision requires a process of information search. Information can be retrieved
from memory or other external sources. In any case, information has to be pre-
processed for the particular problem and a final conclusion has so be drawn.
Therefore, information processing is key for decision making and limited cog-
nitive abilities, as stated in the model of bounded rationality, might essentially
impact decision quality. The major reason for the huge impact of the heuristics
and biases program in research is, that it is able to explain a wide variety of
different decision situations without restricting it due to motivated irrationality
[52, p. 1].
Tversky and Kahneman assume, that decisions under uncertainty are based
on heuristics rather than complex algorithms [48]. Heuristics are defined as men-
tal short-cuts or rules of thumb and require only limited amount of information
and cognitive abilities. Generally, heuristics achieve results fast and depend on
low effort. To do so, they neglect relevant information, which can lead to system-
atic predictable deviations from rationality. There is a huge amount of evidence
that biases can lead to poor outcomes in important and novel decisions [42,53].
This, together with the fact that biases are systematic, emphasises the impor-
tance of incorporating heuristics in modelling.
In their pioneering work, Tversky and Kahneman described three funda-
mental heuristics [48] which are relevant in countless practical situations. The
representativeness heuristic is applied when people make judgements about the
probability of an unknown event. To come up with a judgement, people tend to
judge the probability of the unknown event by finding a comparable known event
and assume that the probabilities will be similar. For illustration, Tversky and
Kahneman developed the “Linda problem”, where they describe the fictitious
person Linda as “31 years old, single, outspoken, and very bright. She majored
in philosophy. As a student, she was deeply concerned with issues of discrimi-
nation and social justice, and also participated in anti-nuclear demonstrations”
[54, p. 297]. Thereupon they asked subjects which is more probable, (a) Linda
being a bank teller or (b) Linda being a bank teller and actively involved in
Reasoning Under Uncertainty 367
feminist movement. Results reveal, that in accordance with their hypothesis, a
vast majority (80–90%) of subjects chose the conjunction (b) to be more likely
than the single event (a). From a logical perspective, a conjunction of events (b)
can never be more likely than any of its constituents (a) and therefore indicates
a violation of rationality. Within the last decades, many different biases have
been linked to the representativeness heuristic (e.g., conjunction fallacy, base
rate neglect, insensitivity to sample size) [42].
The availability heuristic is the second of Tversky and Kahnemans heuristics
and states, that people rely upon knowledge that is easily available and comes
to mind rather than complete data [55]. By relying on the availability of a given
event in someone’s memory, the actual probability of the event can often be
predicted quite good. Nevertheless, sometimes the availability of an event is
influenced by other factors besides the probability or frequency of the occurrence
and in this case the availability heuristic will lead to systematic deviations from
rationality [55]. For example the chronological distance or conciseness are factors
that can influence the availability of an event. The cause of death “firearm”
is estimated as much higher compared to “tobacco”, which can be attributed
to the media coverage of violence [42,56]. Similar to this, subjects who were
asked to estimate “If a random word is taken from an English text, is it more
likely that the word starts with a K, or that K is the third letter?” [55,p.
1125]. Following Tversky and Kahneman’s hypothesis, people easier recall words
beginning with an K and therefore overestimate the number of words that begin
with the letter K. Although experimental results support this hypothesis, a text
typically contains twice as many words which have the letter K at the third,
rather than first letter.
The so-called anchoring and adjustment heuristic describes a widely explored
and robust phenomenon in human decision making [48]. The heuristic can be
very useful when primary values of information do hint to a correct answer and
are relevant to the underlying decision-problem – a situation found in many daily
tasks. The anchor effect – as the central result of the anchoring and adjustment
heuristic – can be found in situations, where a numerical starting point (the
anchor) is processed to form a final estimation. In case the final estimation is
biased towards the initial starting point, one talks about an anchoring effect.
In a well-known demonstration, Tversky and Kahneman asked subjects to esti-
mate the percentage of African countries that are in the United Nations (UN)
[48, p. 1128]. Prior to this, for every subject of the experiment, a random num-
ber between one and one hundred after spinning a wheel of fortune was chosen.
Subjects had to state if the random number is higher or lower compared to
the true value. It was found, that people who had a lower number estimated
fewer countries in the UN than people who had a higher number. Thereupon
numerous experiments validated the robustness of the anchoring effect in vary-
ing fields of application, e.g. general knowledge [57], probability estimates [44,58]
and negotiations [59,60]. Neither financial incentives nor explicit advices could
effectivly mitigate the anchoring effect [61,62]. Moreover, the numerical start-
ing point does not have to be relevant to the underlying decision-problem, even
368 S. Robert et al.
unconsciously perceived or irrelevant values can distort the judgement [61,p.
123]. In general, there are two different approaches to explain the occurrence of
the anchor effect. The original approach of Tversky and Kahneman states that
individuals tend to anchor onto a numerical value and then gradually adjust
away from that value until they reach a decision that seems reasonable [48]. This
anchoring and adjustment process is usually insufficient and therefore biased. In
contrast, the selective accessibility approach argues, that biased estimations are
rooted in an early phase of information processing [57,63,64]. Following the
approach, individuals, when given an anchor, will evaluate the hypothesis that
the anchor is a suitable answer (confirmatory hypothesis testing) and therefore
access all the relevant attributes of the anchor value. Thereon, the approach
assumes that anchoring effects are mediated by the selectively increased acces-
sibility of anchor-consistent knowledge and the final estimate is therefore biased
towards the anchor. Overall, none of the mentioned approaches can fully explain
empirical evidence and the origin of the anchoring effect is still highly debated
within the research community [42,65].
In addition to the three fundamental heuristics and their resulting biases,
there are further heuristics which try to explain decision making under specific
situations. Despite the tremendous success of the heuristics and biases program,
there are alternative approaches to explain actual decision making behaviour. For
example the fast-and-frugal-approach – mostly based on Gigerenzers works – is
also based on several simple heuristics, but in contrast to the classical heuristics,
they are precisely defined and can be directly validated [66,67]. Moreover, the
probabilistic mental model [68] and prospect theory [69] also build on limited
cognitive abilities and are used in different areas to predict decision making
behaviour.
3.3 Practical Aspects for the Integration of Interactive ML
and Decision Theory
The importance of the integration of interactive ML and decision theory is evi-
dent. Given the massive consequences that can result from suboptimal decision
making, it is critical to improve our knowledge about ways to yield better deci-
sion outcomes [46, p. 379]. In our knowledge-based economy, each decision is
likely to have vast implications and will affect subsequent decisions on their
own. Decision problems have to be analysed for their potential receptiveness
to decision biases and in what ways they are likely to benefit from automatic
processing.
On the one side, current technological and methodical advances enable us to
cope with more complex decision tasks. But on the other side, in many practi-
cal situations decision making in terms of the interaction between human and
computer input is still limited and does not tap the full potential. Moreover,
new decision situations in many fields of application are characterized by the
same underlying process and therefore share the common need for new ways of
interaction.
Reasoning Under Uncertainty 369
For example, there are innumerable applications in the field of medical deci-
sion making and cyber-physical systems (e.g. “Industry 4.0”) such as assistance
or recommender systems that are based on the same abstract decision problem,
combine similar approaches of computer algorithms with human input and there-
fore face similar challenges. For instance, the analysis of sensor data is pretty
similar in many practical applications. On the one hand, data may describe
body parameters such as temperature, heartbeat or blood plasma concentration
in a medical context. On the other hand, data may provide information about
the energy consumption of a power unit, the temperature of an engine or the
status of a relay in an industrial context. Although there are many algorithms
that can analyse the captured data in a purely unsupervised fashion, in order
to achieve excellent and instant results, an interactive data analysis backed by
human decision making skills can offer new possibilities and bring context infor-
mation into the process. The same applies to the area of image exploitation. In
many cases, it is about finding structural anomalies in data and learning from
previous examples. With up-to-date methods of image exploitation, algorithms
can detect, count and cluster different types of objects. These algorithms are in
many cases only partially automatic and require human input. In medical image
exploitation, doctors can help to provide diagnostic findings in the segmentation
of skin cancer images [70]. In the industrial context, image exploitation is for
example used to detect tool wear [71]. In both situations, wrong diagnoses and
decisions potentially bear extensive risk and therefore the optimal integration
of human and computer input is of great importance. A big issue is accordingly
the integration process, because exactly here setting up a system between the
expert and the algorithm requires a common ground between them and is crucial
for total imaging. This common ground has to exploit computational power and
integrate human intelligence to realise the best possible output.
4 Open Problems
The study of ML is primarily based on normative models. Most of these models
are the result of centuries of refection and analysis and are widely accepted as
the basis of logical reasoning. For the fact that human decision making skills
are in certain settings superior to computer algorithms – e.g. many ML-methods
perform very badly on extrapolation problems which would be very easy for
humans [32, p. 4] – and major assumptions of normative models cannot be
applied in reality, a conjoint approach of human and machine input could be
key to enhanced decision quality. Therefore, the answer is to put humans in the
loop [40]. However, using normative models to integrate human decision making
in centrals parts of machine learning could lead to faulty predictions since the
nature of actual decision making is of bounded rationality [5].
Based on the described approaches, today we know the specific ways in which
decision makers are likely to be biased and we can describe how people make
decisions with astonishing detail and reliability. In addition, with regards to
normative models, we have a clear vision of how much better decision making
370 S. Robert et al.
could be [46]. The most important step now is to integrate those two different
approaches, correct biases and improve decision making. The prescriptions for
such corrections are called prescriptive models [33, p. 19] and will decide about
the success of human-in-the-loop approaches in ML. Altogether, not only do we
need to know the nature of the specific problem, “but normative models must
be understood in terms of their role in looking for biases, understanding these
biases in terms of descriptive models and developing prescriptive models” [72,
p. 20].
In consideration of this fact, interactive ML approaches are a promising can-
didate for further enhancing the knowledge discovery process. One important
problem which we have to face in future research is which questions to pose to
humans and how to ask those questions [40]. At this point, human machine-
interaction could provide useful insights and offer guidelines for the design of
interfaces and visualisations. Moreover, research in this area, i.e. at the intersec-
tion of cognitive science and computational science is fruitful for further improv-
ing ML thus improve performance on a wide range of tasks, including settings
which are difficult for humans to process (e.g., big data and high dimensional
problems) [32]. According to Lee and Holzinger [73], there is a very common
misconception about high dimensionality, i.e. that ML would produce better
outcomes with higher dimensional data. Increasing amounts of input features
can build more accurate predictors as features are key to learning and under-
standing. However, such attempts need high computational power, and due to
limitations in human perception, understanding structures in high dimensional
spaces is practically impossible. Hence, the outcome must be shaped in a form
perceivable for humans, which is a very difficult problem. Here graph-based rep-
resentations in R2are very helpful in that respect and open up a lot of future
possibilities [74,75].
5 Future Challenges
The important role of iML for dealing with complexity is evident. However,
future research has to be done in various areas.
First of all, only a few research projects have dealt with ciML. The devel-
opment of new ciML approaches for different algorithms has to be expanded to
be able to develop generic human-algorithm interfaces. Research has to focus on
further algorithms beyond decision trees and ant colony algorithms that could
benefit from the new approach of ciML to analyze its full potential.
Secondly, from the knowledge today it cannot be said, which problems ciML
can address and which problems will not be addressable with ciML. Future
research has to focus on the classification of problems in terms of the different
aML, iML and ciML approaches. For some problems we do know that aML can
provide very efficient algorithms, some problems are known to be unsolvable in
polynomial time, but we currently do not have comprehensive knowledge about
the opportunities of ciML.
Thirdly, the iML algorithms proposed so far address very specific problems.
In general, the questions have been solved, how humans can be integrated into
Reasoning Under Uncertainty 371
the algorithm and understand both the underlying problem and the algorithm
with its parameters. Therefore, the past and ongoing research on HCI will play a
prominent role in the future of iML: It has to be further analyzed, how humans
(not only computer scientists) can be empowered to better understand the spe-
cific ML algorithms. This involves adequate visualization techniques of the input
data, as shown by past research projects as well as visualizations to support the
understandability of complex algorithms. In this respect, new interaction tech-
nologies might come in handy. Large displays [76], room-spanning projections
[77], gesture-based interactions and virtual and augmented reality (VR and AR)
[78,79] are new interaction concepts and technologies that have been applied
successfully in the medical [80] and industrial [8183] domains and might be
able to play a roll in the interaction with algorithms in the future.
6 Conclusion
In this paper, we presented the current state of research in two domains: JDM
and ML. We presented a new classification of ML emphasizing on iML and –
more specificly – on ciML. We bridged the two research domains and argued
that future research will have to take both research domains into account, when
dealing with highly complex problems. Both humans and computers have their
specific strengths and weaknesses and putting humans into the loop of ML algo-
rithms might be a very efficient way for solving specific problems. We identified
two application areas, which provide complex problems that might benefit from
the new approach of ciML: health informatics and cyber-physical systems. While
these two domains seem to be different on the first sight, their problems often
share the same characteristics: Often exceptional variances in data need to be
found, e.g. a specific diseases based on physiological data in medicine or malfunc-
tions of complex cyber-physical systems based on sensor data of machines. The
classical approach of aML focuses on finding these patterns based on previous
knowledge from data. However, aML struggles on function extrapolation prob-
lems which are trivial for human learners. Consequently, integrating a human-
into-the-loop (e.g., a human kernel [84]) could make use of human cognitive
abilities and will be a promising approach. While we outlined the potential of
ciML there are multiple open questions to be tackled in the research community.
The explorative development of new ciML approaches for different algorithms
will help to analyze the full potential of ciML. Existing complex problems need
to be classified and application areas for the different iML approaches need to
be identified. And last but not least, the questions on how to support humans
ideally when collaborating with algorithms and big data needs to be addressed.
In this area the experts from both ML and HCI will have to work hand in hand in
this new joint research endeavor that will greatly help in future problem solving.
Acknowledgements. We thank our colleague Henrik Mucha who provided insight
and expertise that greatly assisted this research. We also thank the anonymous review-
ers for their encouraging reviews.
372 S. Robert et al.
References
1. Holzinger, A.: Trends in interactive knowledge discovery for personalized medicine:
cognitive science meets machine learning. IEEE Intell. Inform. Bull. 15, 6–14 (2014)
2. Von Neumann, J., Morgenstern, O.: Theory of Games and Economic Behavior.
Princeton University Press, Princeton (1944)
3. Fox, J., Glasspool, D., Bury, J.: Quantitative and qualitative approaches to rea-
soning under uncertainty in medical decision making. In: Quaglini, S., Barahona,
P., Andreassen, S. (eds.) AIME 2001. LNCS, vol. 2101, pp. 272–282. Springer,
Heidelberg (2001). doi:10.1007/3-540-48229-6 39
4. Ma, W., Xiong, W., Luo, X.: A model for decision making with missing, imprecise,
and uncertain evaluations of multiple criteria. Int. J. Intell. Syst. 28, 152–184
(2013)
5. Simon, H.A.: A behavioral model of rational choice. Q. J. Econ. 69, 99–118 (1955)
6. Dequech, D.: Bounded rationality, institutions, and uncertainty. J. Econ. Issues
35, 911–929 (2001)
7. Holzinger, A.: Lecture 8 biomedical decision making: reasoning and decision sup-
port. In: Biomedical Informatics, pp. 345–377. Springer, Heidelberg (2014)
8. March, S.T., Hevner, A.R.: Integrated decision support systems: a data warehous-
ing perspective. Decis. Support Syst. 43, 1031–1043 (2007)
9. Hansson, S.O.: Decision theory: a brief introduction (2005)
10. Bell, D.E., Raiffa, H., Tversky, A.: Descriptive, normative, and prescriptive interac-
tions in decision making. Decis. Making Descriptive Normative Prescriptive Inter-
act. 1, 9–32 (1988)
11. Pearl, J.: Heuristics: Intelligent Search Strategies for Computer Problem Solving.
Addison-Wesley, Reading (1984)
12. Alan, D., Janet, F., Gregory, A., Russell, B.: Human-Computer Interaction.
Pearson Education Limited, Harlow (2004)
13. Kohavi, R., Provost, F.: Glossary of terms. Mach. Learn. 30, 271–274 (1998)
14. Ankerst, M., Elsen, C., Ester, M., Kriegel, H.P.: Visual classification: an interactive
approach to decision tree construction. In: Proceedings of the Fifth ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining, pp. 392. ACM
(1999)
15. Wakker, P., Deneffe, D.: Eliciting von neumann-morgenstern utilities when proba-
bilities are distorted or unknown. Manage. Sci. 42, 1131–1150 (1996)
16. Domingos, P.: A few useful things to know about machine learning. Commun.
ACM 55, 78–87 (2012)
17. Mitchell, T.M.: Machine Learning. McGraw-Hill, Boston (1997)
18. Martin, J.H., Jurafsky, D.: Speech and language processing. In: International 710th
edn. (2000)
19. Adomavicius, G., Tuzhilin, A.: Toward the next generation of recommender sys-
tems: a survey of the state-of-the-art and possible extensions. IEEE Trans. Knowl.
Data Eng. 17, 734–749 (2005)
20. Li, Q., Zheng, N., Cheng, H.: Springrobot: a prototype autonomous vehicle and its
algorithms for lane detection. IEEE Trans. Intell. Transp. Syst. 5, 300–308 (2004)
21. Marsland, S.: Machine Learning: An Algorithmic Perspective. CRC Press,
Boca Raton (2015)
22. Ankerst, M., Ester, M., Kriegel, H.P.: Towards an effective cooperation of the user
and the computer for classification. In: Proceedings of the Sixth ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining, pp. 179–188.
ACM (2000)
Reasoning Under Uncertainty 373
23. Ware, M., Frank, E., Holmes, G., Hall, M., Witten, I.H.: Interactive machine learn-
ing: letting users build classifiers. Int. J. Hum. Comput. Stud. 55, 281–292 (2001)
24. Fails, J.A., Olsen Jr., D.R.: Interactive machine learning. In: Proceedings of the
8th International Conference on Intelligent User Interfaces, pp. 39–45. ACM (2003)
25. Fiebrink, R., Cook, P.R., Trueman, D.: Human model evaluation in interactive
supervised learning. In: Proceedings of the SIGCHI Conference on Human Factors
in Computing Systems, CHI 2011, pp. 147–156. ACM, New York (2011)
26. Fogarty, J., Tan, D., Kapoor, A., Winder, S.: Cueflik: interactive concept learning
in image search. In: Proceedings of the SIGCHI Conference on Human Factors in
Computing Systems, CHI 2008, pp. 29–38. ACM, New York (2008)
27. Simard, P., Chickering, D., Lakshmiratan, A., Charles, D., Bottou, L., Suarez,
C.G.J., Grangier, D., Amershi, S., Verwey, J., Suh, J.: Ice: enabling non-experts
to build models interactively for large-scale lopsided problems. arXiv preprint
arXiv:1409.4814 (2014)
28. Amershi, S., Chickering, M., Drucker, S.M., Lee, B., Simard, P., Suh, J.: Mod-
eltracker: redesigning performance analysis tools for machine learning. In: Pro-
ceedings of the 33rd Annual ACM Conference on Human Factors in Computing
Systems, CHI 2015, pp. 337–346. ACM, New York (2015)
29. Talbot, J., Lee, B., Kapoor, A., Tan, D.S.: Ensemblematrix: interactive visual-
ization to support machine learning with multiple classifiers. In: Proceedings of
the SIGCHI Conference on Human Factors in Computing Systems, CHI 2009, pp.
1283–1292. ACM, New York (2009)
30. Ankerst, M., Ester, M., Kriegel, H.P.: Towards an effective cooperation of the user
and the computer for classification. In: Proceedings of the Sixth ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining, KDD 2000,
pp. 179–188. ACM, New York (2000)
31. Holzinger, A.: Interactive machine learning for health informatics: when do we need
the human-in-the-loop? Brain Inform. 3, 119–131 (2016)
32. Holzinger, A., Plass, M., Holzinger, K., Cri¸san, G.C., Pintea, C.-M., Palade, V.:
Towards interactive Machine Learning (iML): applying ant colony algorithms to
solve the traveling salesman problem with the human-in-the-loop approach. In:
Buccafurri, F., Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E. (eds.) CD-
ARES 2016. LNCS, vol. 9817, pp. 81–95. Springer, Heidelberg (2016). doi:10.1007/
978-3-319-45507-5 6
33. Baron, J.: Normative Models of Judgment and Decision Making. Wiley, New York
(2004)
34. Raiffa, H.: Applied statistical decision theory (1974)
35. Murphy, K.P.: Machine Learning: A Probabilistic Perspective. MIT Press,
Cambridge (2012)
36. Friedman, J., Hastie, T., Tibshirani, R.: The Elements of Statistical Learning.
Springer Series in Statistics, vol. 1. Springer, New York (2001)
37. Tulabandhula, T., Rudin, C.: Machine learning with operational costs. J. Mach.
Learn. Res. 14, 1989–2028 (2013)
38. Pitz, G.F., Sachs, N.J.: Judgment and decision: theory and application. Annu. Rev.
Psychol. 35, 139–164 (1984)
39. Fischhoff, B.: Judgment and decision making. Wiley Interdisc. Rev. Cogn. Sci. 1,
724–735 (2010)
40. Russakovsky, O., Li, L.J., Fei-Fei, L.: Best of both worlds: human-machine collabo-
ration for object annotation. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 2121–2131 (2015)
374 S. Robert et al.
41. Rapoport, A.: Decision Theory and Decision Behaviour: Normative and Descriptive
Approaches, vol. 15. Springer, Amsterdam (2013)
42. Bazerman, M.H., Moore, D.A.: Judgment in managerial decision making (2013)
43. Bonner, S.E.: Judgment and Decision Making in Accounting. Prentice Hall,
Upper Saddle River (2008)
44. Robert, S.: Informationsverarbeitung in Prognosen: Experimentelle Evidenz. dis-
sertation, University of Osnabrueck (2016)
45. Goldstein, W.M., Hogarth, R.M.: Research on Judgment and Decision Making:
Currents, Connections, and Controversies. Cambridge University Press, Cambridge
(1997)
46. Milkman, K.L., Chugh, D., Bazerman, M.H.: How can decision making be
improved? Perspect. Psychol. Sci. 4, 379–383 (2009)
47. Baron, J.: Thinking and Deciding. Cambridge University Press, Cambridge (2000)
48. Tversky, A., Kahneman, D.: Judgment under uncertainty: heuristics and biases.
In: Wendt, D., Vlek, C. (eds.) Utility, Probability, and Human Decision Making,
pp. 1124–1131. Springer, Amsterdam (1974)
49. Libby, R.: Accounting and Human Information Processing: Theory and Applica-
tions. Prentice Hall, Englewood Cliffs (1981)
50. Ashton, R.H.: Human Information Processing in Accounting. American Accounting
Association, Sarasota (1982)
51. Over, D.: Rationality and the normative/descriptive distinction. In: Blackwell
Handbook of Judgment and Decision Making, London, pp. 3–18 (2004)
52. Gilovich, T., Griffin, D., Kahneman, D.: Heuristics and Biases: The Psychology of
Intuitive Judgment. Cambridge University Press, New York (2002)
53. Newell, B.R.: Judgment under uncertainty (2013)
54. Tversky, A., Kahneman, D.: Extensional versus intuitive reasoning: the conjunction
fallacy in probability judgment. Psychol. Rev. 90, 293 (1983)
55. Tversky, A., Kahneman, D.: Availability: a heuristic for judging frequency and
probability. Cogn. Psychol. 5, 207–232 (1973)
56. Mokdad, A.H., Marks, J.S., Stroup, D.F., Gerberding, J.L.: Actual causes of death
in the United States, 2000. JAMA 291, 1238–1245 (2004)
57. Strack, F., Mussweiler, T.: Explaining the enigmatic anchoring effect: mechanisms
of selective accessibility. J. Pers. Soc. Psychol. 73, 437 (1997)
58. Plous, S.: Thinking the unthinkable: the effects of anchoring on likelihood estimates
of nuclear war1. J. Appl. Soc. Psychol. 19, 67–91 (1989)
59. Ritov, I.: Anchoring in simulated competitive market negotiation. Organ. Behav.
Hum. Decis. Process. 67, 16–25 (1996)
60. Galinsky, A.D., Mussweiler, T.: First offers as anchors: the role of perspective-
taking and negotiator focus. J. Pers. Soc. Psychol. 81, 657 (2001)
61. Chapman, G.B., Johnson, E.J.: Incorporating the irrelevant: anchors in judgments
of belief and value. In: The Psychology of Intuitive Judgment, Heuristics and
Biases, pp. 120–138 (2002)
62. Wilson, T.D., Houston, C.E., Etling, K.M., Brekke, N.: A new look at anchoring
effects: basic anchoring and its antecedents. J. Exp. Psychol. Gen. 125, 387 (1996)
63. Mussweiler, T., Strack, F.: Comparing is believing: a selective accessibility model
of judgmental anchoring. Eur. Rev. Soc. Psychol. 10, 135–167 (1999)
64. Chapman, G.B., Johnson, E.J.: Anchoring, activation, and the construction of
values. Organ. Behav. Hum. Decis. Process. 79, 115–153 (1999)
65. Furnham, A., Boo, H.C.: A literature review of the anchoring effect. J. Socio-Econ.
40, 35–42 (2011)
Reasoning Under Uncertainty 375
66. Gigerenzer, G.: Why the distinction between single-event probabilities and fre-
quencies is important for psychology (and vice versa). In: Subjective Probability,
pp. 129–161 (1994)
67. Gigerenzer, G., Czerlinski, J., Martignon, L.: How good are fast and frugal heuris-
tics? In: Shanteau, J., Mellers, B.A., Schum, D.A. (eds.) Decision Science and
Technology, pp. 81–103. Springer, New York (1999)
68. Gigerenzer, G., Hoffrage, U., Kleinb¨olting, H.: Probabilistic mental models: a
brunswikian theory of confidence. Psychol. Rev. 98, 506 (1991)
69. Kahneman, D., Tversky, A.: Prospect theory: an analysis of decision under risk.
Econometrica: J. Econometric Soc. 47(2), 263–291 (1979)
70. Xu, L., Jackowski, M., Goshtasby, A., Roseman, D., Bines, S., Yu, C., Dhawan, A.,
Huntley, A.: Segmentation of skin cancer images. Image Vis. Comput. 17, 65–74
(1999)
71. Kr´olczyk, G., Legutko, S., Raos, P.: Cutting wedge wear examination during turn-
ing of duplex stainless steel. Tehniˇcki Vjesnik-Technical Gazette 20, 413–418 (2013)
72. Baron, J.: Rationality and Intelligence. Cambridge University Press, New York
(2005)
73. Lee, S., Holzinger, A.: Knowledge discovery from complex high dimensional data.
In: Michaelis, S., Piatkowski, N., Stolpe, M. (eds.) Solving Large Scale Learning
Tasks. Challenges and Algorithms. LNCS (LNAI), vol. 9580, pp. 148–167. Springer,
Heidelberg (2016). doi:10.1007/978-3-319-41706-6 7
74. Holzinger, A., Malle, B., Giuliani, N.: On graph extraction from image data. In:
Slezak, D., Peters, J.F., Tan, A.H., Schwabe, L. (eds.) Brain Informatics and
Health, BIH 2014. LNAI, vol. 8609, pp. 552–563. Springer, Heidelberg (2014)
75. Valdez, A.C., Dehmer, M., Holzinger, A.: Application of graph entropy
for knowledge discovery and data mining in bibliometric data. In: Dehmer, M.,
Emmert-Streib, F., Chen, Z., Li, X., Shi, Y. (eds.) Mathematical Foundations and
Applications of Graph Entropy, pp. 259–272. Wiley, New York (2016)
76. Cao, X., Balakrishnan, R.: Visionwand: interaction techniques for large displays
using a passive wand tracked in 3d. In: Proceedings of the 16th Annual ACM
Symposium on User Interface Software and Technology, UIST 2003, pp. 173–182.
ACM, New York (2003)
77. Jones, B.R., Benko, H., Ofek, E., Wilson, A.D.: Illumiroom: peripheral projected
illusions for interactive experiences. In: Proceedings of the SIGCHI Conference on
Human Factors in Computing Systems, CHI 2013, pp. 869–878. ACM, New York
(2013)
78. Milgram, P., Takemura, H., Utsumi, A., Kishino, F.: Augmented reality: a class of
displays on the reality-virtuality continuum. In: Photonics for industrial applica-
tions, International Society for Optics and Photonics, pp. 282–292 (1995)
79. Azuma, R.T.: A survey of augmented reality. Presence: Teleoperators Virtual Env-
iron. 6, 355–385 (1997)
80. Fuchs, H., et al.: Augmented reality visualization for laparoscopic surgery. In: Wells,
W.M., Colchester, A., Delp, S. (eds.) MICCAI 1998. LNCS, vol. 1496, pp. 934–943.
Springer, Heidelberg (1998). doi:10.1007/BFb0056282
81. Paelke, V., R¨ocker,C.,Koch,N.,Flatt,H.,B¨uttner, S.: User interfaces for cyber-
physical systems. at-Automatisierungstechnik 63, 833–843 (2015)
82. B¨uttner, S., Sand, O., R¨ocker, C.: Extending the design space in industrial man-
ufacturing through mobile projection. In: Proceedings of the 17th International
Conference on Human-Computer Interaction with Mobile Devices and Services
Adjunct, MobileHCI 2015, pp. 1130–1133. ACM, New York (2015)
376 S. Robert et al.
83. B¨uttner, S., Funk, M., Sand, O., R¨ocker, C.: Using head-mounted displays and
in-situ projection for assistive systems - a comparison. In: Proceedings of the 9th
ACM International Conference on PErvasive Technologies Related to Assistive
Environments, vol. 8. ACM (2016)
84. Wilson, A.G., Dann, C., Lucas, C.G., Xing, E.P.: The human kernel. arXiv preprint
arXiv:1510.07389 (2015)
... IML has been shown to create better models than ML alone while at the same time giving people a better understanding of the process being modeled (Robert et al. 2016). Using IML could speed up the scenario process by integrating advances in automation with human participation to create more transparent quantitative scenario models without the workshops. ...
... Figure 23 shows a user training a model that will recognize different ways someone can ask a chatbot to book a flight and what the keywords are they need to work with such as where they are leaving from, where they are going to, and when. By adding a human to help train these algorithms models can be trained faster with shorter learning iterations, and the people that interact with the models have a better overall understanding of them (Robert et al. 2016). Because data science experts are not needed to do all the data analysis, modeling, programming, and deployment IML allows faster learning through more rapid and frequent iterations on the model with the user of the model. ...
... The practice of IML (Fails and Olsen Jr 2003;Amershi et al. 2014) has been shown to create better models than ML alone while increasing people's understanding of the model (Robert et al. 2016). This research furthers the field of IML by showing that SAAM, a system coordinated through IML, was able to create scenarios that are at least as good as those created with other methods that use automation and demonstrates a practical example of implementing an IML workflow to create scenarios through a socio- ...
Thesis
Full-text available
Scenario planning is used extensively in strategic planning because it helps leaders broaden their perspectives and make better decisions by presenting possible futures in story form. Some of the benefits of using scenarios include breaking away from groupthink, creating better products, acceleration of organization learning and reducing bias. Product development teams, particularly for digital products, are gaining more autonomy in organizations and tend to manage risk by undergoing very short development iterations on their products while leaning on their consumers for feedback – a process known as agile development. This method tends to limit the perspective of the team and foster groupthink, two side effects which could potentially be addressed using scenarios. However, the time-consuming and expensive processes used to create scenarios are inaccessible to agile product development teams, and even teams that use scenarios for strategic direction typically use them at the beginning of product development and do not keep them up to date over time, eventually making them irrelevant to decision making. This research explores automating the bottlenecks of the scenario process so they can be incorporated into autonomous agile teams by creating and rigorously tests an artifact that combines Natural Language Processing (NLP) to understand data, Interactive Machine Learning (IML) to combine automation with human expertise, Fuzzy Cognitive Maps (FCM) for quantitative scenario modeling, and Horizon Scanning (HS) to keep models up to date; a system I call Scenario Acceleration through Automated Modelling (SAAM). Using Design Science Research (DSR), I demonstrate how these technologies can be used together to speed up the scenario creation process while keeping people in the loop, and how they can be kept up to date over time. This research lays the foundation for product development teams to use scenarios in agile processes, with the goal of creating better products and avoiding disruption. This work makes several contributions: Firstly, it furthers the body of knowledge on scenario development by showing how to create scenarios with automation and how scenarios could be used by agile teams. Secondly, it demonstrates a novel method of creating FCM with NLP and human collaboration, and how to use Horizon Scanning to keep models up to date over time. Finally, I leave an artifact that can be used by other teams who want to continue this vein of research, or for product teams that want to utilize this method.
... The success of this collaboration relies on integrating the human element as an active decision-maker within the system. For this reason, from a user-centric point of view, the interest of IML lies in how this process can help humans to address how humans provide data, cope with inaccurate models, and what types or frequency of questions are appropriate to ask them [112]. To attain this goal, it is crucial to comprehensively examine the different dimensions of interaction, including communication, involvement, and adaptability, to gain deeper insight into their impact on user integration and engagement. ...
... With this collaboration, users are included in the control loop and iteratively help the system create personalized models adapted to their particularities on the fly. Hence, this joint action can improve both the model's performance and the user's confidence in the intelligent system [112], which is a relevant aspect of human-centric solutions designed for the workplace [120]. At the same time, an improvement of the user experience is required to achieve the actual settlement of IoT systems and intelligent solutions in our society [121]. ...
... They are designed to automatically learn from data and produce outcomes, resolv-ing problems without human intervention. In such scenarios, the initial strategy for building some collaboration was relegated to allowing humans to alter a predefined parameter of an aML algorithm to improve its performance [48]. As aML techniques are still extensively employed, they relegate human interaction to be extremely limited and concrete, preventing the integration of humans into the systems and long-lasting interaction and collaboration. ...
Article
Full-text available
Human-AI collaboration has evolved into a complex, multidimensional paradigm shaped by research in various domains. Key areas such as human-in-the-loop systems, Interactive Machine Learning (IML), Hybrid Intelligence, and Human-Agent Interaction have significantly contributed to this development. However, these fields often lack cohesion, underscoring the need for a cohesive perspective to advance. This work addresses this gap by integrating insights from diverse aspects of collaboration to present a holistic approach to fostering effective and adaptive interactions between humans and artificial agents. It emphasizes empowering end-users with greater control and involvement in decision-making processes, thereby enhancing both the levels of interactivity and adaptability within intelligent systems. Moving beyond a focus on AI training techniques, this paper presents a broader perspective on incorporating human input into AI decision-making and learning processes, highlighting the importance of flexibility in systems and user engagement. The manuscript proposes a framework encompassing five levels of human integration and examines their relationship with core collaboration aspects, including the system purpose, participant expertise, and system proactivity. By synthesizing current knowledge on human-AI collaboration and outlining essential design principles, this work aims to advance the field and foster interdisciplinary collaboration among researchers, practitioners, and designers.
... L'implication de l'humain dans le processus d'apprentissage peut prendre différentes formes (Robert et al., 2016) : soit en utilisant l'humain pour remplacer l'algorithme (i.e l'humain générant le modèle directement à partir des données), soit en plaçant l'humain dans la boucle d'entraînement-évaluation (i.e. l'humain donnant des feedbacks à l'algorithme), soit en faisant travailler l'humain et l'algorithme main dans la main pour créer le modèle. ...
Thesis
Full-text available
Dans le contexte de la transition énergétique et de l'augmentation des interconnexions entre les réseaux de transport d'électricité en Europe, les opérateurs du réseau français doivent désormais faire face à davantage de fluctuations et des dynamiques nouvelles sur le réseau. Pour garantir la sûreté de ce réseau, les opérateurs s'appuient sur des logiciels informatiques permettant de réaliser des simulations, ou de suivre l'évolution d'indicateurs créés manuellement par des experts grâce à leur connaissance du fonctionnement du réseau. Le gestionnaire de réseau de transport d'électricité français RTE (Réseau de Transport d'Electricité) s'intéresse notamment aux développements d'outils permettant d'assister les opérateurs dans leur tâche de surveillance des transits sur les lignes électriques. Les transits sont en effet des grandeurs particulièrement importantes pour maintenir le réseau dans un état de sécurité, garantissant la sûreté du matériel et des personnes. Cependant, les indicateurs utilisés ne sont pas faciles à mettre à jour du fait de l'expertise nécessaire pour les construire et les analyser. Pour répondre à la problématique énoncée, cette thèse a pour objet la construction d'indicateurs, sous la forme d'expressions symboliques, permettant d'estimer les transits sur les lignes électriques. Le problème est étudié sous l'angle de la Régression Symbolique et investigué à la fois par des approches génétiques d'Evolution Grammaticale et d'Apprentissage par Renforcement dans lesquelles la connaissance experte, explicite et implicite, est prise en compte. Les connaissances explicites sur la physique et l'expertise du domaine électrique sont représentées sous la forme d'une grammaire non-contextuelle délimitant l'espace fonctionnel à partir duquel l'expression est créée. Une première approche d'Evolution Grammaticale Interactive propose d’améliorer incrémentalement les expressions trouvées par la mise à jour d'une grammaire entre les apprentissages évolutionnaires. Les expressions obtenues sur des données réelles issues de l'historique du réseau sont validées par une évaluation de métriques d'apprentissages, complétée par une évaluation de leur interprétabilité. Dans un second temps, nous proposons une approche par renforcement pour chercher dans un espace délimité par une grammaire non-contextuelle afin de construire une expression symbolique pertinente pour des applications comportant des contraintes physiques. Cette méthode est validée sur des données de l'état de l'art de la régression symbolique, ainsi qu’un jeu de données comportant des contraintes physiques pour en évaluer l'interprétabilité. De plus, afin de tirer parti des complémentarités entre les capacités des algorithmes d'apprentissage automatique et de l'expertise des opérateurs du réseau, des algorithmes interactifs de Régression Symbolique sont proposés et intégrés dans des plateformes interactives. L'interactivité est employée à la fois pour mettre à jour la connaissance représentée sous forme grammaticale, analyser, interagir avec et commenter les solutions proposées par les différentes approches. Ces algorithmes et interfaces interactifs ont également pour but de prendre en compte de la connaissance implicite, plus difficile à formaliser, grâce à l'utilisation de mécanismes d'interactions basés sur des suggestions et des préférences de l’utilisateur.
... An information processing approach also suggests, that the design process and the cognitive functions behind the problem solving process could be modeled artificially [7], [8]. Through that Artificial Intelligence (AI) could assist during the design process and enhance human problem solving (see [9]). However, the introduction of AI to the design process implies new challenges that can make the design of systems even more complex. ...
Article
Full-text available
Artificial intelligence methods are increasingly used in the design of sociotechnical systems. The design of such systems can be described as an ill-defined problem. The problem space often changes while working on a design solution. Further, an anticipation of future system states is challenging, because of possible changes of organisational, technological or other environmental aspects during the lifecycle of a sociotechnical system. Embodied artificial intelligence can help to structure the problem space. It can also act as a feedback mechanism of the inner workings of a system and its surroundings for a designer. Therefore, this short article discusses the emphasis of embodied artificial intelligence for the theory of design in the context of sociotechnical systems and its implications.
... For the present study, ML in the loop (MLL) is investigated; this strategy depends on how the ML tool will be used. Researchers have analyzed the necessary workflows to improve results (44), but in medicine, where health professionals play an indispensable role, other investigators have studied the doctorin-the-loop in terms of system performance (45,46). Today, how ML models perform is no longer the sole concern; models' . ...
Article
Full-text available
The use of machine learning (ML) for diagnosis support has advanced in the field of health. In the present paper, the results of studying ML techniques in a tuberculosis diagnosis loop in a scenario of limited resources are presented. Data are analyzed using a tuberculosis (TB) therapy program at a health institution in a main city of a developing country using five ML models. Logistic regression, classification trees, random forest, support vector machines, and artificial neural networks are trained under physician supervision following physicians' typical daily work. The models are trained on seven main variables collected when patients arrive at the facility. Additionally, the variables applied to train the models are analyzed, and the models' advantages and limitations are discussed in the context of the automated ML techniques. The results show that artificial neural networks obtain the best results in terms of accuracy, sensitivity, and area under the receiver operating curve. These results represent an improvement over smear microscopy, which is commonly used techniques to detect TB for special cases. Findings demonstrate that ML in the TB diagnosis loop can be reinforced with available data to serve as an alternative diagnosis tool based on data processing in places where the health infrastructure is limited.
Chapter
Machine learning (ML) optimization techniques serve as essential for training models to achieve high performance in a diverse areas. This chapter offers a thorough summary of machine learning optimization techniques. This chapter analysis the development of optimization techniques over time. A number of common optimization constraints are also discussed. Developing a model that works effectively and provides accurate predictions in a certain set of instances is the main objective of ML. We require ML optimization to accomplish that. The practice of modifying hyper parameters with an optimization technique to minimize the cost function is called ML optimization. Because the cost function indicates the difference between the actual value of the estimated parameter and the value predicted by the model, it is crucial to reduce it. This chapter will provide a general explanation of the workings and drawbacks of the ML optimization strategies. Numerous advancements in machine learning optimization techniques have been put forth in this chapter.
Article
Full-text available
Over the decades, the spin dynamics of a large set of lanthanide complexes have been explored. Lanthanide-based molecular nanomagnets are bistable spin systems, generally conceptualised as classical bits, but many lanthanide complexes have also been presented as candidate quantum bits (qubits). Here, we offer a third alternative and model them as probabilistic bits (p-bits), where their stochastic behaviour constitutes a computational resource instead of a limitation. Employing an ad-hoc modelling tool for molecular spin p-bits and molecular nanomagnets, we simulate a minimal p-bit network under realistic conditions. Finally, we go back to a recently published dataset and screen the best lanthanide complexes for p-bit behaviour, lay out the performance of the different lanthanide ions and chemical families and offer some chemical design considerations.
Preprint
Full-text available
Over the decades, the spin dynamics of a large set of lanthanide complexes have been explored. Lanthanide-based molecular nanomagnets are bistable spin systems, generally conceptualized as classical bits, but many lanthanide complexes have also been presented as candidate quantum bits (qubits). Here we offer a third alternative and model them as probabilistic bits (p-bits), where their stochastic behavior constitutes a computational resource instead of a limitation. We present a modelling tool for molecular spin p-bits, we demonstrate its capability to simulate bulk magnetic relaxation data and ac experiments and to simulate a minimal p-bit network under realistic conditions. Finally, we go back to a recent systematic data gathering and screen the best lanthanide complexes for p-bit behavior, lay out the performance of the different lanthanide ions and chemical families and offer some chemical design considerations.
Chapter
This chapter demonstrates the crucial role that human-AI interfaces play in conveying the trustworthiness of AI solutions to their users. Explainability is a central component of such interfaces, particularly in high-stake domains where human oversight is essential: justice, finance, security, and medicine. To successfully build and communicate trustworthiness, a user-centered approach to the design and development of AI solutions and their human interfaces is essential. In this chapter, we explain how proven methods for stakeholder analysis and user testing from human-computer interaction (HCI) research can be adapted to human-AI interaction (HAII) in support of this goal. The practical implementation of a user-centric approach is described within the context of AI applications in computational pathology.
Article
Full-text available
In previous anchoring studies people were asked to consider an anchor as a possible answer to the target question or were given informative anchors. The authors predicted that basic anchoring effects can occur, whereby uninformative numerical anchors influence a judgment even when people are not asked to compare this number to the target value. Five studies supported these hypotheses: Basic anchoring occurs if people pay sufficient attention to the anchor value; knowledgeable people are less susceptible to basic anchoring effects; anchoring appears to operate unintentionally and nonconsciously in that it is difficult to avoid even when people are forewarned. The possible mechanisms of basic anchoring and the relation between these mechanisms and other processes of judgment and correction are discussed.
Article
Full-text available
Research on people's confidence in their general knowledge has to date produced two fairly stable effects, many inconsistent results, and no comprehensive theory. We propose such a comprehensive framework, the theory of probabilistic mental models (PMM theory). The theory (a) explains both the overconfidence effect (mean confidence is higher than percentage of answers correct) and the hard–easy effect (overconfidence increases with item difficulty) reported in the literature and (b) predicts conditions under which both effects appear, disappear, or invert. In addition, (c) it predicts a new phenomenon, the confidence–frequency effect, a systematic difference between a judgment of confidence in a single event (i.e., that any given answer is correct) and a judgment of the frequency of correct answers in the long run. Two experiments are reported that support PMM theory by confirming these predictions, and several apparent anomalies reported in the literature are explained and integrated into the present framework.
Conference Paper
Full-text available
Book
What is intelligence? Can it be increased by teaching? If so, how, and what difference would an increase make? Before we can answer these questions, we need to clarify them. Jonathan Baron argues that when we do so we find that intelligence has much to do with rational thinking, and that the skills involved in rational thinking are in fact teachable, at least to some extent. Rationality and Intelligence develops and justifies a prescriptive theory of rational thinking in terms of utility theory and the theory of rational life plans. The prescriptive theory, buttressed by other assumptions, suggests that people generally think too little and in a way that is insufficiently critical of the initial possibilities that occur to them. However these biases can be - and sometimes are - corrected by education.
Chapter
The analysis of decision making under uncertainty has again become a major focus of interest. This volume presents contributions from leading specialists in different fields and provides a summary and synthesis of work in this area. It is based on a conference held at the Harvard Business School. The book brings together the different approaches to decision making - normative, descriptive, and prescriptive - which largely correspond to different disciplinary interests. Mathematicians have concentrated on rational procedures for decision making - how people should make decisions. Psychologists have examined how poeple do make decisions, and how far their behaviour is compatible with any rational model. Operations researchers study the application of decision models to actual problems. Throughout, the aim is to present the current state of research and its application and also to show how the different disciplinary approaches can inform one another and thus lay the foundations for the integrated analysis of decision making. The book will be of interest to researchers, teachers - for use as background reading for a decision theory course - students, and consultants and others involved in the practical application of the analysis of decision making. It will be of interest to specialists and students in statistics, mathematics, economics, psychology and the behavioural sciences, operations research, and management science.