ArticlePDF Available

Abstract

The clever algorithms needed to process big data cannot (and will never) solve most of the critical risk analysis problems that we face. The problems are especially acute where we must assess and manage risk in areas where there is little or no direct historical data to draw upon; where relevant data are difficult to identify or are novel; or causal mechanisms or human intentions remain hidden. Such risks include terrorist attacks, ecological disasters and failures of novel systems and marketplaces. Here, the tendency has been to rely on the intuition of ‘experts’ for decision-making. However, there is an effective and proven alternative: the smart data approach that combines expert judgment (including understanding of underlying causal mechanisms) with relevant data. In particular Bayesian Networks (BNs) provide workable models for combining human and artificial sources of intelligence even when big data approaches to risk assessment are not possible.
The era of ‘big data’ offers
enormous opportunities for
societal improvements. There is
an expectation – and even excitement
– that, by simply applying sophisti-
cated machine learning algorithms to
‘big data’ sets, we may automatically
find solutions to problems that were
previously either unsolvable or would
incur prohibitive economic costs.
Yet, the clever algorithms needed to
process big data cannot (and will
never) solve most of the critical risk
analysis problems that we face. Big
data, even when carefully collected is
typically unstructured and noisy; even
the ‘biggest data’ typically lack crucial,
often hidden, information about key
causal or explanatory variables that
generate or influence the data we
observe. For example, the world’s
leading economists failed to predict
the 2008–2010 international financial
crisis because they relied on models
based on historical statistical data that
could not adapt to new circumstances,
even when those circumstances were
foreseeable by contrarian experts. In
short, analysts often depend on
models that are inadequate represen-
tations of reality – good for predicting
the past but poor at predicting the
future.
These fundamental problems are
especially acute where we must assess
and manage risk in areas where there
is little or no direct historical data to
draw upon; where relevant data are
difficult to identify or are novel; or
causal mechanisms or human inten-
tions remain hidden. Such risks include
terrorist attacks, ecological disasters
and failures of novel systems and mar-
ketplaces. Here, the tendency has been
to rely on the intuition of ‘experts’ for
decision-making. However, there is an
effective and proven alternative: the
smart data approach that combines
expert judgment (including an under-
standing of underlying causal mecha-
nisms) with relevant data. In particular,
Bayesian Networks (BNs) provide
workable models for combining
human and artificial sources of intelli-
gence even when big data approaches
to risk assessment are not possible
BNs describe networks of causes and
effects, using a graphical framework
that provides rigorous quantification
of risks and clear communication of
results. Quantitative probability
assignments accompany the graphical
specification of a BN and can be
derived from historical data or expert
judgment. A BN then serves as a basis
for answering probabilistic queries
given knowledge about the world.
Computations are based on a theo-
rem by the Reverend Thomas Bayes
dating back to 1763 and, to date, pro-
vides the only rational and consistent
way to update a belief in some uncer-
tain event (such as a decline in share
price) when we observe new evidence
related to that event (such as better
than expected earnings).
The problem of correctly updating
beliefs in the light of new evidence is
central to all disciplines that involve
any form of reasoning (law, medicine
and engineering as well as finance
and indeed AI). Thus, a BN provides a
Norman Fenton and Martin Neil ask what next after ‘big data’, focussing on how
Bayesian Networks are pioneering the ‘smart data’ revolution
How Bayesian Networks are pioneering
the ‘smart data’ revolution
22
PROFILE
general approach to reasoning, with
explainable models of reality, in con-
trast to big data approaches, where
the emphasis is on prediction, rather
than explanation and on association
rather than causal connection.
BNs are now widely recognised as a
powerful technology for handling risk,
uncertainty and decision making.
Since 1995, researchers have incorpo-
rated BN techniques into software
products, which in turn have helped
develop decision support systems in
many scientific and industrial applica-
tions, including: medical diagnostics,
operational and financial risk, cyberse-
curity, safety and quality assessment,
sports prediction, the law, forensics
and equipment fault diagnosis.
A major challenge of reasoning causally
is that people lacked the methods and
tools to do so productively and effec-
tively. Fortunately, there has been a
quiet revolution in both areas. Work by
Pearl (Turing award winner for AI), has
provided the necessary philosophical
and practical instruction on how to
elicit, articulate and manipulate causal
models. Likewise, our work on causal
idioms and influence diagrams has
been applied in many application areas
to make model building and validation
faster, more accurate and ultimately
more productive.
Also, there are now software products,
containing sophisticated algorithms,
that help us to easily design the BN
models needed to represent complex
problems and present insightful
results to decision makers. Compared
to previous generations of software
these are more powerful and easier to
use – so much so that they are becom-
ing as familiar and accessible as
spreadsheets became in the 1980s.
Indeed, this big leap forward is helping
decision makers think both graphically,
about relationships and numerically,
about the strength of these relation-
ships, when modelling complex
problems, in a way impossible to do
previously.
Recent research has now made it easy
to accurately incorporate numeric
variables in the analysis, an obvious
practical requirement, but one that
the past generation of BN algorithms
could not satisfy. There are now BN
products that implement the latest and
most accurate inference algorithms, as
well as:
Provide ‘smartlearning of relation-
ships from data – with or without
missing values – incorporating as
much or as little expert judgement
as required.
Automatically identify and select a
decision strategy to maximise overall
utility or minimise overall risk, using
hybrid influence diagrams.
Compute the ‘value of Information’
of uncertain variables in terms of
how much should be paid to find
more information out about them.
Our recent and ongoing research pro-
jects are providing ever more efficient
algorithms both for building and
deploying BNs (such as in patient-held
medical devices and energy smart
meters), including efficient cloud-
based services for applications like
cybersecurity risk analysis.
Many are asking what comes after ‘big
data’? Surprisingly, the ideas of Thomas
Bayes, despite being pioneered over
250 years ago, may provide the answer
in the form of smarter decisions from
data and causal, uncertain knowledge.
Norman Fenton
Professor of Risk and Information
Management
Queen Mary University of London
Tel: +44 (0)20 7882 7860
n.fenton@qmul.ac.uk
www.eecs.qmul.ac.uk/~norman/
www.twitter.com/profnfenton
23
PROFILE
Our projects
Much of the recent and ongoing BN research described here
is from projects:
www.bayes-knowledge.com
www.pambayesian.org
www.causaldynamics.org
Software
Much of the new BN functionality described here has been
incorporated into version 10 of the AgenaRisk software
( www.agenarisk.com )
Book
“Risk Assessment and Decision Analysis with Bayesian
Networks2012, CRC Press by Fenton and Neil provides a
thorough overview of BNs that is accessible to non-mathe-
matical readers. Second edition available August 2018
(see https://www.crcpress.com/9781138035119 )
... In this instance, progress toward improving project performance will be stymied (Ghasemaghaei and Calic, 2020). Accordingly, Fenton and Neil (2018) maintain that "big data, even when carefully collected, is typically unstructured and noisy; even the 'biggest data' typically lack crucial, often hidden, information about key causal or explanatory variables that generate or influence the data we observe". (p. 1). ...
... A smart data approach is driven by what data are required for prediction rather than what is available (Constantinou and Fenton, 2017;Fenton and Neil, 2018). Such data is processed and turned into actionable information, empowering an organisation. ...
Article
Full-text available
Within construction, we have become increasingly accustomed to relying on the benefits of digital technologies, such as Building Information Modelling, to improve the performance and productivity of projects. We have, however, overlooked the problems that technology is unable to redress. One such problem is rework, which has become so embedded in practice that technology adoption alone can not resolve the issue without fundamental changes in how information is managed for decision-making. Hence, the motivation of this paper is to bring to the fore the challenges of classifying and creating an ontology for rework that can be used to understand its patterns of occurrence and risks and provide a much-needed structure for decision-making in transport mega-projects. Using an exploratory case study approach, we examine ‘how’ rework information is currently being managed by an alliance that contributes significantly to delivering a multi-billion dollar mega-transport project. We reveal the challenges around location, format, structure, granularity and redundancy hindering the alliance’s ability to classify and manage rework data. We use the generative machine learning technique of Correlation Explanation to illustrate how we can make headway toward classifying and then creating an ontology for rework. We propose a theoretical framework utilising a smart data approach to generate an ontology that can effectively use business analytics (i.e., descriptive, predictive and prescriptive) to manage rework risks.
Article
Rework can be a problem in construction projects. While research has contributed to the accumulation of knowledge about rework, there remains limited understanding about its causal mechanisms. The result is an inability to manage and control the risks and uncertainties that surround rework causation in construction projects. This paper suggests that possessing knowledge and having an understanding of rework causation and its context is pivotal for anticipating its likely occurrence and putting in place risk mitigation strategies. The paper concludes by suggesting there is a need to develop a knowledge engineering system to capture rework in projects so that artificial intelligence can be used to support uncertainty management.
Article
Full-text available
Rework ca be a significant and costly problem during the construction of large-scale transport (>$500 million) projects. There is, however, limited understanding and knowledge about the underlying dynamics and causal mechanisms of rework. Both the public and private sector organisations tend to ignore the costs of rework and thus have been unable to contain and manage its risks. This paper takes a look at the wicked and inter-organisational problem of rework and invites the public and private sectors to work in unison to thwart this risk. The paper makes a twofold contribution: (1) it calls on the public and private sectors to consider the likelihood of rework as a part of their risk management strategy; and (2) suggests that there is need use a smart data approach to anticipate what might go wrong in terms of rework so as to deliver large-scale transport projects successfully.
ResearchGate has not been able to resolve any references for this publication.