ArticlePDF AvailableLiterature Review

Abstract and Figures

The emergence of artificial intelligence (AI) and its progressively wider impact on many sectors requires an assessment of its effect on the achievement of the Sustainable Development Goals. Using a consensus-based expert elicitation process, we find that AI can enable the accomplishment of 134 targets across all the goals, but it may also inhibit 59 targets. However, current research foci overlook important aspects. The fast development of AI needs to be supported by the necessary regulatory insight and oversight for AI-based technologies to enable sustainable development. Failure to do so could result in gaps in transparency, safety, and ethical standards. Artificial intelligence (AI) is becoming more and more common in people’s lives. Here, the authors use an expert elicitation method to understand how AI may affect the achievement of the Sustainable Development Goals.
Content may be subject to copyright.
PERSPECTIVE
The role of articial intelligence in achieving the
Sustainable Development Goals
Ricardo Vinuesa 1*, Hossein Azizpour 2, Iolanda Leite2, Madeline Balaam3,
Virginia Dignum4,SamiDomisch 5, Anna Felländer6, Simone Daniela Langhans7,8,
Max Tegmark9& Francesco Fuso Nerini 10*
The emergence of articial intelligence (AI) and its progressively wider impact on many
sectors requires an assessment of its effect on the achievement of the Sustainable Devel-
opment Goals. Using a consensus-based expert elicitation process, we nd that AI can enable
the accomplishment of 134 targets across all the goals, but it may also inhibit 59 targets.
However, current research foci overlook important aspects. The fast development of AI needs
to be supported by the necessary regulatory insight and oversight for AI-based technologies
to enable sustainable development. Failure to do so could result in gaps in transparency,
safety, and ethical standards.
The emergence of articial intelligence (AI) is shaping an increasing range of sectors. For
instance, AI is expected to affect global productivity1, equality and inclusion2, environ-
mental outcomes3, and several other areas, both in the short and long term4. Reported
potential impacts of AI indicate both positive5and negative6impacts on sustainable develop-
ment. However, to date, there is no published study systematically assessing the extent to which
AI might impact all aspects of sustainable developmentdened in this study as the 17 Sus-
tainable Development Goals (SDGs) and 169 targets internationally agreed in the 2030 Agenda
for Sustainable Development7. This is a critical research gap, as we nd that AI may inuence the
ability to meet all SDGs.
Here we present and discuss implications of how AI can either enable or inhibit the delivery of
all 17 goals and 169 targets recognized in the 2030 Agenda for Sustainable Development.
Relationships were characterized by the methods reported at the end of this study, which can be
summarized as a consensus-based expert elicitation process, informed by previous studies aimed
at mapping SDGs interlinkages810. A summary of the results is given in Fig. 1and the Sup-
plementary Data 1 provides a complete list of all the SDGs and targets, together with the detailed
results from this work. Although there is no internationally agreed denition of AI, for this study
we considered as AI any software technology with at least one of the following capabilities:
perceptionincluding audio, visual, textual, and tactile (e.g., face recognition), decision-making
(e.g., medical diagnosis systems), prediction (e.g., weather forecast), automatic knowledge
https://doi.org/10.1038/s41467-019-14108-y OPEN
1Linné FLOW Centre, KTH Mechanics, SE-100 44 Stockholm, Sweden. 2Division of Robotics, Perception, and Learning, School of EECS, KTH Royal Institute
Of Technology, Stockholm, Sweden. 3Division of Media Technology and Interaction Design, KTH Royal Institute of Technology, Lindstedtsvägen 3,
Stockholm, Sweden. 4Responsible AI Group, Department of Computing Sciences, Umeå University, SE-90358 Umeå, Sweden. 5Leibniz-Institute of
Freshwater Ecology and Inland Fisheries, Müggelseedamm 310, 12587 Berlin, Germany. 6AI Sustainability Center, SE-114 34 Stockholm, Sweden. 7Basque
Centre for Climate Change (BC3), 48940 Leioa, Spain. 8Department of Zoology, University of Otago, 340 Great King Street, 9016 Dunedin, New Zealand.
9Center for Brains, Minds and Machines, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139, USA. 10 Unit of Energy Systems Analysis
(dESA), KTH Royal Institute of Technology, Brinellvagen, 68SE-1004 Stockholm, Sweden. *email: rvinuesa@mech.kth.se;francesco.fusonerini@energy.kth.se
NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications 1
1234567890():,;
extraction and pattern recognition from data (e.g., discovery of
fake news circles in social media), interactive communication
(e.g., social robots or chat bots), and logical reasoning (e.g., theory
development from premises). This view encompasses a large
variety of subelds, including machine learning.
Documented connections between AI and the SDGs
Our review of relevant evidence shows that AI may act as an
enabler on 134 targets (79%) across all SDGs, generally through a
technological improvement, which may allow to overcome certain
present limitations. However, 59 targets (35%, also across all
SDGs) may experience a negative impact from the development
of AI. For the purpose of this study, we divide the SDGs into
three categories, according to the three pillars of sustainable
development, namely Society, Economy, and Environment11,12
(see the Methods section). This classication allows us to provide
an overview of the general areas of inuence of AI. In Fig. 1,we
also provide the results obtained when weighting how appropriate
is the evidence presented in each reference to assess an inter-
linkage to the percentage of targets assessed, as discussed in the
Methods section and below. A detailed assessment of the Society,
Economy, and Environment groups, together with illustrative
examples, are discussed next.
AI and societal outcomes. Sixty-seven targets (82%) within the
Society group could potentially benet from AI-based technolo-
gies (Fig. 2). For instance, in SDG 1 on no poverty, SDG 4 on
quality education, SDG 6 on clean water and sanitation, SDG 7 on
affordable and clean energy, and SDG 11 on sustainable cities, AI
may act as an enabler for all the targets by supporting the pro-
vision of food, health, water, and energy services to the popula-
tion. It can also underpin low-carbon systems, for instance, by
supporting the creation of circular economies and smart cities
that efciently use their resources13,14. For example, AI can
enable smart and low-carbon cities encompassing a range of
interconnected technologies such as electrical autonomous vehi-
cles and smart appliances that can enable demand response in the
electricity sector13,14 (with benets across SDGs 7, 11, and 13 on
climate action). AI can also help to integrate variable renewables
by enabling smart grids that partially match electrical demand to
times when the sun is shining and the wind is blowing13. Fewer
targets in the Society group can be impacted negatively by AI (31
targets, 38%) than the ones with positive impact. However, their
consideration is crucial. Many of these relate to how the tech-
nological improvements enabled by AI may be implemented in
countries with different cultural values and wealth. Advanced AI
technology, research, and product design may require massive
computational resources only available through large computing
centers. These facilities have a very high energy requirement and
carbon footprint15. For instance, cryptocurrency applications
such as Bitcoin are globally using as much electricity as some
nationselectrical demand16, compromising outcomes in the SDG
7 sphere, but also on SDG 13 on Climate Action. Some estimates
suggest that the total electricity demand of information and
communications technologies (ICTs) could require up to 20% of
the global electricity demand by 2030, from around 1% today15.
Green growth of ICT technology is therefore essential17. More
efcient cooling systems for data centers, broader energy ef-
ciency, and renewable-energy usage in ICTs will all play a role in
containing the electricity demand growth15. In addition to more
efcient and renewable-energy-based data centers, it is essential
to embed human knowledge in the development of AI models.
Besides the fact that the human brain consumes much less energy
than what is used to train AI models, the available knowledge
introduced in the model (see, for instance, physics-informed deep
learning18) does not need to be learnt through data-intensive
training, a fact that may signicantly reduce the associated energy
consumption. Although AI-enabled technology can act as a cat-
alyst to achieve the 2030 Agenda, it may also trigger inequalities
100%
(100%) 75%
(69%) 69%
(69%)
100%
(93%)
56%
(44%)
100%
(100%)
100%
(100%)
100%
(90%)
58%
(52%)
92%
(77%)
100%
(91%)
90%
(75%)
82%
(59%)
26%
(15%)
80%
(70%)
90%
(90%)
100%
(88%)
12
3
4
5
6
7
11
16
8
9
10
12
17
13
14
15
86%
(43%) 25%
(13%)
8%
(8%)
70%
(60%)
33%
(31%)
63%
(28%)
40%
(40%)
20%
(10%)
25%
(15%)
33%
(25%)
50%
(34%)
70%
(55%)
27%
(16%)
11%
(5%)
20%
(20%)
30%
(13%)
33%
(8%)
12
3
4
5
6
7
11
16
8
9
10
12
17
13
14
15
Positive impacts of AI: 79% (71%) Negative impacts of AI: 35% (23%)
ab
E
n
v
i
r
o
n
m
e
n
t
:
9
3
%
(
8
5
%
)
E
n
v
i
r
o
n
m
e
n
t
:
S
o
c
i
e
t
y
:
3
8
%
(
2
5
%
)
S
o
c
i
e
t
y
:
8
2
%
(
7
7
%
)
3
0
%
(
1
2
%
)
E
c
o
n
o
m
y
:
7
0
%
(
5
5
%
)
E
c
o
n
o
m
y
:
3
3
%
(
2
3
%
)
Fig. 1 Summary of positive and negative impact of AI on the various SDGs. Documented evidence of the potential of AI acting as (a) an enabler or (b)an
inhibitor on each of the SDGs. The numbers inside the colored squares represent each of the SDGs (see the Supplementary Data 1). The percentages on
the top indicate the proportion of all targets potentially affected by AI and the ones in the inner circle of the gure correspond to proportions within each
SDG. The results corresponding to the three main groups, namely Society, Economy, and Environment, are also shown in the outer circle of the gure. The
results obtained when the type of evidence is taken into account are shown by the inner shaded area and the values in brackets.
PERSPECTIVE NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y
2NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications
that may act as inhibitors on SDGs 1, 4, and 5. This duality is
reected in target 1.1, as AI can help to identify areas of poverty
and foster international action using satellite images5. On the
other hand, it may also lead to additional qualication require-
ments for any job, consequently increasing the inherent
inequalities19 and acting as an inhibitor towards the achievement
of this target.
Another important drawback of AI-based developments is that
they are traditionally based on the needs and values of nations in
which AI is being developed. If AI technology and big data are
used in regions where ethical scrutiny, transparency, and
democratic control are lacking, AI might enable nationalism,
hate towards minorities, and bias election outcomes20. The term
big nudginghas emerged to represent using big data and AI to
exploit psychological weaknesses to steer decisionscreating
problems such as damaging social cohesion, democratic princi-
ples, and even human rights21. AI has been recently utilized to
develop citizen scores, which are used to control social
behavior22. This type of score is a clear example of threat to
human rights due to AI misuse and one of its biggest problems is
the lack of information received by the citizens on the type of
analyzed data and the consequences this may have on their lives.
It is also important to note that AI technology is unevenly
distributed: for instance, complex AI-enhanced agricultural
equipment may not be accessible to small farmers and thus
produce an increased gap with respect to larger producers in
more developed economies23, consequently inhibiting the
achievement of some targets of SDG 2 on zero hunger. There is
another important shortcoming of AI in the context of SDG 5 on
gender equality: there is insufcient research assessing the
potential impact of technologies such as smart algorithms, image
recognition, or reinforced learning on discrimination against
women and minorities. For instance, machine-learning algo-
rithms uncritically trained on regular news articles will
inadvertently learn and reproduce the societal biases against
women and girls, which are embedded in current languages.
Word embeddings, a popular technique in natural language
processing, have been found to exacerbate existing gender
stereotypes2. In addition to the lack of diversity in datasets,
another main issue is the lack of gender, racial, and ethnic
diversity in the AI workforce24. Diversity is one of the main
principles supporting innovation and societal resilience, which
will become essential in a society exposed to changes associated to
AI development25. Societal resilience is also promoted by
decentralization, i.e., by the implementation of AI technologies
adapted to the cultural background and the particular needs of
different regions.
AI and economic outcomes. The technological advantages pro-
vided by AI may also have a positive impact on the achievement
of a number of SDGs within the Economy group. We have
identied benets from AI on 42 targets (70%) from these SDGs,
whereas negative impacts are reported in 20 targets (33%), as
shown in Fig. 1. Although Acemoglu and Restrepo1report a net
positive impact of AI-enabled technologies associated to
increased productivity, the literature also reects potential nega-
tive impacts mainly related to increased inequalities2629. In the
context of the Economy group of SDGs, if future markets rely
heavily on data analysis and these resources are not equally
available in low- and middle- income countries, the economical
gap may be signicantly increased due to the newly introduced
inequalities30,31 signicantly impacting SDGs 8 (decent work and
economic growth), 9 (industry, innovation and infrastructure),
and 10 (reduced inequalities). Brynjolfsson and McAfee31 argue
that AI can exacerbate inequality also within nations. By repla-
cing old jobs with ones requiring more skills, technology dis-
proportionately rewards the educated: since the mid 1970s, the
salaries in the United States (US) salaries rose about 25% for
those with graduate degrees, while the average high-school
dropout took a 30% pay cut. Moreover, automation shifts cor-
porate income to those who own companies from those who
work there. Such transfer of revenue from workers to investors
helps explain why, even though the combined revenues of
Society
Enabler
Inhibitor
1
2
3
4
5
A
B
1
2
3
4
5
A
B
1
2
3
4
5
B
C
1
2
3
4
5
B
C
1
2
3
4
5
6
7
8
9
A
B
C
D
1
2
3
4
5
6
7
8
9
A
B
C
D
1
2
3
4
5
6
7
A
B
C
1
2
3
4
5
6
7
A
B
C
1
2
3
4
5
6
A
B
C
123456ABC
1
2
3456AB
1
2
3
4
5
6
A
B
1
2
3
A
B
1
2
3
A
B
1
2
3
4
5
6
7
A
B
C
1
2
3
4
5
6
7
A
B
C
1
2
3
4
5
6
7
8
9
1
0
A
B
1234567891
0AB
Fig. 2 Detailed assessment of the impact of AI on the SDGs within the Society group. Documented evidence of positive or negative impact of AI on the
achievement of each of the targets from SDGs 1, 2, 3, 4, 5, 6, 7, 11, and 16 (https://www.un.org/sustainabledevelopment/). Each block in the diagram
represents a target (see the Supplementary Data 1 for additional details on the targets). For targets highlighted in green or orange, we found published
evidence that AI could potentially enable or inhibit such target, respectively. The absence of highlighting indicates the absence of identied evidence. It is
noteworthy that this does not necessarily imply the absence of a relationship. (The content of of this gure has not been reviewed by the United Nations
and does not reect its views).
NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y PERSPECTIVE
NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications 3
Detroit's Big 3(GM, Ford, and Chrysler) in 1990 were almost
identical to those of Silicon Valley's Big 3(Google, Apple, and
Facebook) in 2014, the latter had 9 times fewer employees and
were worth 30 times more on the stock market32. Figure 3shows
an assessment of the documented positive and negative effects on
the various targets within the SDGs in the Economy group.
Although the identied linkages in the Economy group are
mainly positive, trade-offs cannot be neglected. For instance, AI
can have a negative effect on social media usage, by showing users
content specically suited to their preconceived ideas. This may
lead to political polarization33 and affect social cohesion21 with
consequences in the context of SDG 10 on reduced inequalities.
On the other hand, AI can help identify sources of inequality and
conict34,35, and therewith potentially reduce inequalities, for
instance, by using simulations to assess how virtual societies may
respond to changes. However, there is an underlying risk when
using AI to evaluate and predict human behavior, which is the
inherent bias in the data. It has been reported that a number of
discriminatory challenges are faced in the automated targeting of
online job advertising using AI35, essentially related to the
previous biases in selection processes conducted by human
recruiters. The work by Dalenberg35 highlights the need of
modifying the data preparation process and explicitly adapting
the AI-based algorithms used for selection processes to avoid
such biases.
AI and environmental outcomes. The last group of SDGs, i.e.,
the one related to Environment, is analyzed in Fig. 4. The three
SDGs in this group are related to climate action, life below water
and life on land (SDGs 13, 14, and 15). For the Environment
group, we identied 25 targets (93%) for which AI could act as an
enabler. Benets from AI could be derived by the possibility of
analyzing large-scale interconnected databases to develop joint
actions aimed at preserving the environment. Looking at SDG 13
on climate action, there is evidence that AI advances will support
the understanding of climate change and the modeling of its
possible impacts. Furthermore, AI will support low-carbon
energy systems with high integration of renewable energy and
energy efciency, which are all needed to address climate
change13,36,37. AI can also be used to help improve the health of
ecosystems. The achievement of target 14.1, calling to prevent and
signicantly reduce marine pollution of all kinds, can benet
from AI through algorithms for automatic identication of pos-
sible oil spills38. Another example is target 15.3, which calls for
combating desertication and restoring degraded land and soil.
According to Mohamadi et al.39, neural networks and objective-
oriented techniques can be used to improve the classication of
vegetation cover types based on satellite images, with the possi-
bility of processing large amounts of images in a relatively short
time. These AI techniques can help to identify desertication
trends over large areas, information that is relevant for environ-
mental planning, decision-making, and management to avoid
further desertication, or help reverse trends by identifying the
major drivers. However, as pointed out above, efforts to achieve
SDG 13 on climate action could be undermined by the high-
energy needs for AI applications, especially if non carbon-neutral
energy sources are used. Furthermore, despite the many examples
of how AI is increasingly applied to improve biodiversity mon-
itoring and conservation40, it can be conjectured that an increased
access to AI-related information of ecosystems may drive over-
exploitation of resources, although such misuse has so far not
Enabler
Economy
Inhibitor
1
2
3
4
5
6
7
8
9
1
0
A
B
1
2
3
4
5
6
7
8
9
1
0
A
B
1
2
3
4
5
A
B
C
1
2
3
4
5
A
B
C
1
2
3
4
5
6
7
A
B
C
1234567ABC
1
2
345678AB
C
1
2
3
4
5
6
7
8
A
B
C
1
2
3
4
5
6
7
8
9
1
0
1
1
1
2
1
3
1
4
1
5
1
6
1
7
1
8
1
9
1234567891
0
1
1
1
2
1
3
1
4
1
5
1
6
1
7
1
8
1
9
Fig. 3 Detailed assessment of the impact of AI on the SDGs within the Economy group. Documented evidence of positive or negative impact of AI on the
achievement of each of the targets from SDGs 8, 9, 10, 12, and 17 (https://www.un.org/sustainabledevelopment/). The interpretation of the blocks and
colors is as in Fig. 2. (The content of of this gure has not been reviewed by the United Nations and does not reect its views).
Environment
Enabler
Inhibitor
1
2
3
A
B
1
2
3
A
B
1
2
3
4
5
6
7
A
B
C
1234567ABC
1
2
3456789AB
C
123456789ABC
Fig. 4 Detailed assessment of the impact of AI on the SDGs within the Environment group. Documented evidence of positive or negative impact of AI on
the achievement of each of the targets from SDGs 13, 14, and 15 (https://www.un.org/sustainabledevelopment/). The interpretation of the blocks and
colors is as in Fig. 2. (The content of of this gure has not been reviewed by the United Nations and does not reect its views).
PERSPECTIVE NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y
4NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications
been sufciently documented. This aspect is further dis-
cussed below, where currently identied gaps in AI research are
considered.
An assessment of the collected evidence on the interlinkages.A
deeper analysis of the gathered evidence was undertaken as
shown in Fig. 1(and explained in the Methods section). In
practice, each interlinkage was weighted based on the applic-
ability and appropriateness of each of the references to assess a
specic interlinkageand possibly identify research gaps.
Although accounting for the type of evidence has a relatively
small effect on the positive impacts (we see a reduction of posi-
tively affected targets from 79% to 71%), we observe a more
signicant reduction (from 35% to 23%) in the targets with
negative impact of AI. This can be partly due the fact that AI
research typically involves quantitative methods that would bias
the results towards the positive effects. However, there are some
differences across the Society, Economy and Environment
spheres. In the Society sphere, when weighting the appropriate-
ness of evidence, positively affected targets diminish by 5 per-
centage points (p.p.) and negatively affected targets by 13 p.p. In
particular, weighting the appropriateness of evidence on negative
impacts on SDG 1 (on no poverty) and SDG 6 (on clean water
and sanitation) reduces the fraction of affected targets by 43 p.p.
and 35 p.p., respectively. In the Economy group instead, positive
impacts are reduced more (15 p.p.) than negative ones (10 p.p.)
when taking into account the appropriateness of the found evi-
dence to speak of the issues. This can be related to the extensive
study in literature assessing the displacement of jobs due to AI
(because of clear policy and societal concerns), but overall the
longer-term benets of AI on the economy are perhaps not so
extensively characterized by currently available methods. Finally,
although the weighting of evidence decreases the positive impacts
of AI on the Environment group only by 8 p.p., the negative
impacts see the largest average reduction (18 p.p.). This is
explained by the fact that, although there are some indications of
the potential negative impact of AI on this SDG, there is no
strong evidence (in any of the targets) supporting this claim, and
therefore this is a relevant area for future research.
In general, the fact that the evidence on interlinkages between
AI and the large majority of targets is not based on tailored
analyses and tools to refer to that particular issue provides a
strong rationale to address a number of research gaps, which are
identied and listed in the section below.
Research gaps on the role of AI in sustainable development
The more we enable SDGs by deploying AI applications, from
autonomous vehicles41 to AI-powered healthcare solutions42 and
smart electrical grids13, the more important it becomes to invest
in the AI safety research needed to keep these systems robust and
benecial, so as to prevent them from malfunctioning, or from
getting hacked43. A crucial research venue for a safe integration of
AI is understanding catastrophes, which can be enabled by a
systemic fault in AI technology. For instance, a recent World
Economic Forum (WEF) report raises such a concern due to the
integration of AI in the nancial sector44. It is therefore very
important to raise awareness on the risks associated to possible
failures of AI systems in a society progressively more dependent
on this technology. Furthermore, although we were able to nd
numerous studies suggesting that AI can potentially serve as an
enabler for many SDG targets and indicators, a signicant frac-
tion of these studies have been conducted in controlled laboratory
environments, based on limited datasets or using prototypes4547.
Hence, extrapolating this information to evaluate the real-world
effects often remains a challenge. This is particularly true when
measuring the impact of AI across broader scales, both tempo-
rally and spatially. We acknowledge that conducting controlled
experimental trials for evaluating real-world impacts of AI can
result in depicting a snapshot situation, where AI tools are tai-
lored towards that specic environment. However, as society is
constantly changing (also due to factors including non-AI-based
technological advances), the requirements set for AI are changing
as well, resulting in a feedback loop with interactions between
society and AI. Another underemphasized aspect in existing lit-
erature is the resilience of the society towards AI-enabled chan-
ges. Therefore, novel methodologies are required to ensure that
the impact of new technologies are assessed from the points of
view of efciency, ethics, and sustainability, prior to launching
large-scale AI deployments. In this sense, research aimed at
obtaining insight on the reasons for failure of AI systems,
introducing combined humanmachine analysis tools48, are an
essential step towards accountable AI technology, given the large
risk associated to such a failure.
Although we found more published evidence of AI serving as
an enabler than as an inhibitor on the SDGs, there are at least two
important aspects that should be considered. First, self-interest
can be expected to bias the AI research community and industry
towards publishing positive results. Second, discovering detri-
mental aspects of AI may require longer-term studies and, as
mentioned above, there are not many established evaluation
methodologies available to do so. Bias towards publishing positive
results is particularly apparent in the SDGs corresponding to the
Environment group. A good example of this bias is target 14.5 on
conserving coastal and marine areas, where machine-learning
algorithms can provide optimum solutions given a wide range of
parameters regarding the best choice of areas to include in con-
servation networks49. However, even if the solutions are optimal
from a mathematical point of view (given a certain range of
selected parameters), additional research would be needed to
assess the long-term impact of such algorithms on equity and
fairness6, precisely because of the unknown factors that may
come into play. Regarding the second point stated above, it is
likely that the AI projects with the highest potential of max-
imizing prot will get funded. Without control, research on AI is
expected to be directed towards AI applications where funding
and commercial interests are. This may result in increased
inequality50. Consequently, there is the risk that AI-based tech-
nologies with potential to achieve certain SDGs may not be
prioritized, if their expected economic impact is not high. Fur-
thermore, it is essential to promote the development of initiatives
to assess the societal, ethical, legal, and environmental implica-
tions of new AI technologies.
Substantive research and application of AI technologies to
SDGs is concerned with the development of better data-mining
and machine-learning techniques for the prediction of certain
events. This is the case of applications such as forecasting
extreme weather conditions or predicting recidivist offender
behavior. The expectation with this research is to allow the
preparation and response for a wide range of events. However,
there is a research gap in real-world applications of such systems,
e.g., by governments (as discussed above). Institutions have a
number of barriers to the adoption AI systems as part of their
decision-making process, including the need of setting up mea-
sures for cybersecurity and the need to protect the privacy of
citizens and their data. Both aspects have implications on human
rights regarding the issues of surveillance, tracking, commu-
nication, and data storage, as well as automation of processes
without rigorous ethical standards21. Targeting these gaps would
be essential to ensure the usability and practicality of AI tech-
nologies for governments. This would also be a prerequisite for
understanding long-term impacts of AI regarding its potential,
NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y PERSPECTIVE
NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications 5
while regulating its use to reduce the possible bias that can be
inherent to AI6.
Furthermore, our research suggests that AI applications are
currently biased towards SDG issues that are mainly relevant to
those nations where most AI researchers live and work. For
instance, many systems applying AI technologies to agriculture,
e.g., to automate harvesting or optimize its timing, are located
within wealthy nations. Our literature search resulted in only a
handful of examples where AI technologies are applied to SDG-
related issues in nations without strong AI research. Moreover, if
AI technologies are designed and developed for technologically
advanced environments, they have the potential to exacerbate
problems in less wealthy nations (e.g., when it comes to food
production). This nding leads to a substantial concern that
developments in AI technologies could increase inequalities both
between and within countries, in ways which counteract the
overall purpose of the SDGs. We encourage researchers and
funders to focus more on designing and developing AI solutions,
which respond to localized problems in less wealthy nations and
regions. Projects undertaking such work should ensure that
solutions are not simply transferred from technology-intensive
nations. Instead, they should be developed based on a deep
understanding of the respective region or culture to increase the
likelihood of adoption and success.
Towards sustainable AI
The great wealth that AI-powered technology has the potential
to create may go mainly to those already well-off and educated,
while job displacement leaves others worse off. Globally, the
growing economic importance of AI may result in increased
inequalities due to the unevenly distributed educational and
computing resources throughout the world. Furthermore, the
existing biases in the data used to train AI algorithms may
result in the exacerbation of those biases, eventually leading to
increased discrimination. Another related problem is the usage
of AI to produce computational (commercial, political) pro-
paganda based on big data (also dened as big nudging),
which is spread through social media by independent AI agents
with the goals of manipulating public opinion and producing
political polarization51. Despite the fact that current scientic
evidence refutes technological determinism of such fake news51,
long-term impacts of AI are possible (although unstudied) due
to a lack of robust research methods. A change of paradigm is
therefore needed to promote cooperation and to limit the
possibilities for control of citizen behavior through AI. The
concept of Finance 4.0 has been proposed52 as a multi-currency
nancial system promoting a circular economy, which is
aligned with societal goals and values. Informational self-
determination (in which the individual takes an active role in
how their data are handled by AI systems) would be an essential
aspect of such a paradigm52. The data intensiveness of AI
applications creates another problem: the need for more and
more detailed information to improve AI algorithms, which is
in conict with the need of more transparent handling and
protection of personal data53. One area where this conict is
particularly important is healthcare: Panch et al.54 argue that
although the vast amount of personal healthcare data could lead
to the development of very powerful tools for diagnosis and
treatment, the numerous problems associated to data owner-
ship and privacy call for careful policy intervention. This is also
an area where more research is needed to assess the possible
long-term negative consequences. All the challenges mentioned
above culminate in the academic discourse about legal per-
sonality of robots55, which may lead to alarming narratives of
technological totalitarianism.
Many of these aspects result from the interplay between
technological developments on one side and requests from indi-
viduals, response from governments, as well as environmental
resources and dynamics on the other. Figure 5shows a schematic
representation of these dynamics, with emphasis on the role of
technology. Based on the evidence discussed above, these inter-
actions are not currently balanced and the advent of AI has
exacerbated the process. A wide range of new technologies are
being developed very fast, signicantly affecting the way indivi-
duals live as well as the impacts on the environment, requiring
new piloting procedures from governments. The problem is that
neither individuals nor governments seem to be able to follow the
pace of these technological developments. This fact is illustrated
by the lack of appropriate legislation to ensure the long-term
viability of these new technologies. We argue that it is essential to
reverse this trend. A rst step in this direction is to establish
adequate policy and legislation frameworks, to help direct the vast
potential of AI towards the highest benet for individuals and the
environment, as well as towards the achievement of the SDGs.
Regulatory oversight should be preceded by regulatory
insight, where policymakers have sufcient understanding of AI
challenges to be able to formulate sound policy. Developing such
insight is even more urgent than oversight, as policy formulated
without understanding is likely to be ineffective at best and
counterproductive at worst.
Although strong and connected institutions (covered by SDG
16) are needed to regulate the future of AI, we nd that there is
limited understanding of the potential impact of AI on institu-
tions. Examples of the positive impacts include AI algorithms
aimed at improving fraud detection56,57 or assessing the possible
effects of certain legislation58,59. Another concern is that data-
driven approaches for policing may hinder equal access to justice
because of algorithm bias, particularly towards minorities60.
Consequently, we believe that it is imperative to develop legisla-
tion regarding transparency and accountability of AI, as well as to
decide the ethical standards to which AI-based technology should
be subjected to. This debate is being pushed forward by initiatives
such as the IEEE (Institute of Electrical and Electronics Engi-
neers) ethical aligned design60 and the new EU (European
Union) ethical guidelines for trustworthy AI61. It is noteworthy
that despite the importance of an ethical, responsible, and
trustworthy approach to AI development and use, in a sense, this
issue is independent of the aims of the article. In other words, one
can envision AI applications that improve SDG outcomes while
not being fully aligned with AI ethics guidelines. We therefore
recommend that AI applications that target SDGs are open and
explicit about guiding ethical principles, also by indicating
explicitly how they align with the existing guidelines. On the
other hand, the lack of interpretability of AI, which is currently
one of the challenges of AI research, adds an additional com-
plication to the enforcement of such regulatory actions62. Note
that this implies that AI algorithms (which are trained with data
consisting of previous regulations and decisions) may act as a
mirrorreecting biases and unfair policy. This presents an
opportunity to possibly identify and correct certain errors in the
existing procedures. The friction between the uptake of data-
driven AI applications and the need of protecting the privacy and
security of the individuals is stark. When not properly regulated,
the vast amount of data produced by citizens might potentially be
used to inuence consumer opinion towards a certain product or
political cause51.
AI applications that have positive societal welfare implica-
tions may not always benet each individual separately41.This
inherent dilemma of collective vs. individual benet is relevant
in the scope of AI applications but is not one that should be
solved by the application of AI itself. This has always been an
PERSPECTIVE NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y
6NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications
issue affecting humankind and it cannot be solved in a simple
way, since such a solution requires participation of all involved
stakeholders. The dynamicity of context and the level of
abstraction at which human values are described imply that
there is not a single ethical theory that holds all the time in all
situations63. Consequently, a single set of utilitarian ethical
principles with AI would not be recommendable due to the high
complexity of our societies52. It is also essential to be aware of
the potential complexity in the interaction between human
and AI agents, and of the increasing need for ethics-driven
legislation and certication mechanisms for AI systems. This is
true for all AI applications, but especially those that, if they
became uncontrolled, could have even catastrophic effects on
humanity, such as autonomous weapons. Regarding the latter,
associations of AI and robotics experts are already getting
together to call for legislation and limitations of their use64.
Furthermore, associations such as the Future of Life Institute
are reviewing and collecting policy actions and shared princi-
ples around the world to monitor progress towards sustainable-
development-friendly AI65. To deal with the ethical dilemmas
raised above, it is important that all applications provide
openness about the choices and decisions made during design,
development, and use, including information about the pro-
venance and governance of the data used for training algo-
rithms, and about whether and how they align with existing AI
guidelines. It is therefore important to adopt decentralized AI
approaches for a more equitable development of AI66.
We are at a critical turning point for the future of AI. A global
and science-driven debate to develop shared principles and leg-
islation among nations and cultures is necessary to shape a future
in which AI positively contributes to the achievement of all the
SDGs. The current choices to develop a sustainable-development-
Technology
Planetary boundaries
Individuals Government
Political requests
Policy and legislation
+/– impacts
Resources
Environment
Resources, well-being
Resources
Needs, +/– impacts, behaviors
New needs
Technology advancements
New technology for legislation
Legislation and standards
Needs, +/– impacts, policy
Fig. 5 Interaction of AI and society. Schematic representation showing the identied agents and their roles towards the development of AI. Thicker arrows
indicate faster change. In this representation, technology affects individuals through technical developments, which change the way people work and
interact with each other and with the environment, whereas individuals would interact with technology through new needs to be satised. Technology
(including technology itself and its developers) affects governments through new developments that need appropriate piloting and testing. Also,
technology developers affect government through lobbying and inuencing decision makers. Governments provide legislation and standards to technology.
The governments affect individuals through policy and legislation, and individuals would require new legislation consistent with the changing circumstances
from the governments. The environment interacts with technology by providing the resources needed for technological development and is affected by the
environmental impact of technology. Furthermore, the environment is affected either negatively or positively by the needs, impacts, and choices of
individuals and governments, which in turn require environmental resources. Finally, the environment is also an underlying layer that provides the
planetary boundariesto the mentioned interactions.
NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y PERSPECTIVE
NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications 7
friendly AI by 2030 have the potential to unlock benets that
could go far-beyond the SDGs within our century. All actors in all
nations should be represented in this dialogue, to ensure that no
one is left behind. On the other hand, postponing or not having
such a conversation could result in an unequal and unsustainable
AI-fueled future.
Methods
In this section we describe the process employed to obtain the results described in
the present study and shown in the Supplementary Data 1. The goal was to answer
the question Is there published evidence of AI acting as an enabler or an inhibitor
for this particular target?for each of the 169 targets within the 17 SDGs. To this
end, we conducted a consensus-based expert elicitation process, informed by
previous studies on mapping SDGs interlinkages8,9and following Butler et al.67
and Morgan68. The authors of this study are academics spanning a wide range of
disciplines, including engineering, natural and social sciences, and acted as experts
for the elicitation process. The authors performed an expert-driven literature
search to support the identied connections between AI and the various targets,
where the following sources of information were considered as acceptable evidence:
published work on real-world applications (given the quality variation depending
on the venue, we ensured that the publications considered in the analysis were of
sufcient quality); published evidence on controlled/laboratory scenarios (given the
quality variation depending on the venue, we ensured that the publications con-
sidered in the analysis were of sufcient quality); reports from accredited organi-
zations (for instance: UN or government bodies); and documented commercial-
stage applications. On the other hand, the following sources of information were
not considered as acceptable evidence: educated conjectures, real-world applica-
tions without peer-reviewed research; media, public beliefs or other sources of
information.
The expert elicitation process was conducted as follows: each of the SDGs was
assigned to one or more main contributors, and in some cases to several additional
contributors as summarized in the Supplementary Data 1 (here the initials cor-
respond to the author names). The main contributors carried out a rst literature
search for that SDG and then the additional contributors completed the main
analysis. One published study on a synergy or a trade-off between a target and AI
was considered enough for mapping the interlinkage. However, for nearly all
targets several references are provided. After the analysis of a certain SDG was
concluded by the contributors, a reviewer was assigned to evaluate the connections
and reasoning presented by the contributors. The reviewer was not part of the rst
analysis and we tried to assign the roles of the main contributor and reviewer to
experts with complementary competences for each of the SDGs. The role of the
reviewer was to bring up additional points of view and considerations, while cri-
tically assessing the analysis. Then, the main contributors and reviewers iteratively
discussed to improve the results presented for each of the SDGs until the analysis
for all the SDGs was sufciently rened.
After reaching consensus regarding the assessment shown in the Supplementary
Data 1, we analyzed the results by evaluating the number of targets for which AI
may act as an enabler or an inhibitor, and calculated the percentage of targets with
positive and negative impact of AI for each of the 17 goals, as shown in Fig. 1.In
addition, we divided the SDGs into the three following categories: Society, Econ-
omy, and Environment, consistent with the classication discussed by Refs. 11,12.
The SDGs assigned to each of the categories are shown in Fig. 6and the individual
results from each of these groups can be observed in Figs. 24. These gures
indicate, for each target within each SDG, whether any published evidence of
positive or negative impact was found.
Taking into account the types of evidence. In the methodology described above,
a connection between AI and a certain target is established if at least one reference
documenting such a link was found. As the analyzed studies rely on very different
types of evidence, it is important to classify the references based on the methods
employed to support their conclusions. Therefore, all the references in the Sup-
plementary Data 1 include a classication from (A) to (D) according to the fol-
lowing criteria:
References using sophisticated tools and data to refer to this particular issue
and with the possibility to be generalized are of type (A).
Studies based on data to refer to this particular issue, but with limited
generalizability, are of type (B).
Anecdotal qualitative studies and methods are of type (C) .
Purely theoretical or speculative references are of type (D).
The various classes were assigned following the same expert elicitation process
described above. Then, the contribution of these references towards the linkages is
weighted and categories (A), (B), (C), and (D) are assigned relative weights of 1,
0.75, 0.5, and 0.25, respectively. It is noteworthy that, given the vast range of studies
on all the SDG areas, the literature search was not exhaustive and, therefore, certain
targets are related to more references than others in our study. To avoid any bias
associated to the different amounts of references in the various targets, we
considered the largest positive and negative weight to establish the connection with
each target. Let us consider the following example: for a certain target, one
reference of type (B) documents a positive connection and two references of types
(A) and (D) document a negative connection with AI. In this case, the potential
positive impact of AI on that target will be assessed with 0.75, while the potential
negative impact is 1.
Limitations of the research. The presented analysis represents the perspective of
the authors. Some literature on how AI might affect certain SDGs could have been
missed by the authors or there might not be published evidence yet on such
interlinkage. Nevertheless, the employed methods tried to minimize the subjectivity
of the assessment. How AI might affect the delivery of each SDG was assessed and
reviewed by several authors and a number of studies were reviewed for each
interlinkage. Furthermore, as discussed in the Methods section, each interlinkage
was discussed among a subset of authors until consensus was reached on its nature.
Finally, this study relies on the analysis of the SDGs. The SDGs provide a
powerful lens for looking at internationally agreed goals on sustainable
development and present a leap forward compared with the Millenium
Development Goals in the representation of all spheres of sustainable development,
encompassing human rights69, social sustainability, environmental outcomes, and
economic development. However, the SDGs are a political compromise and might
be limited in the representation of some of the complex dynamics and cross-
interactions among targets. Therefore, the SDGs have to be considered in
conjunction with previous and current, and other international agreements9. For
instance, as pointed out in a recent work by UN Human Rights69, human rights
considerations are highly embedded in the SDGs. Nevertheless, the SDGs should be
considered as a complement, rather than a replacement, of the United Nations
Universal Human Rights Charter70.
Data availability
The authors declare that all the data supporting the ndings of this study are available
within the paper and its Supplementary Data 1 le.
Received: 3 May 2019; Accepted: 16 December 2019;
Society
Economy
Environment
Fig. 6 Categorization of the SDGs (https://www.un.org/sustainabledevelopment/) into the Society, Economy, and Environment groups. (The content
of this gure has not been reviewed by the United Nations and does not reect its views).
PERSPECTIVE NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y
8NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications
References
1. Acemoglu, D. & Restrepo, P. Articial Intelligence, Automation, and Work.
NBER Working Paper No. 24196 (National Bereau of Economic Research,
2018).
2. Bolukbasi, T., Chang, K.-W., Zou, J., Saligrama, V. & Kalai, A. Man is to
computer programmer as woman is to homemaker? Debiasing word
embeddings. Adv. Neural Inf. Process. Syst. 29, 43494357 (2016).
3. Norouzzadeh, M. S. et al. Automatically identifying, counting, and describing
wild animals in camera-trap images with deep learning. Proc. Natl Acad. Sci.
USA 115, E5716E5725 (2018).
4. Tegmark, M. Life 3.0: Being Human in the Age of Articial Intelligence
(Random House Audio Publishing Group, 2017).
5. Jean, N. et al. Combining satellite imagery and machine learning to predict
poverty. Science (80-.) 353, 790794 (2016).
6. Courtland, R. Bias detectives: the researchers striving to make algorithms fair.
Nature 558, 357360 (2018).
7. UN General Assembly (UNGA). A/RES/70/1Transforming our world: the
2030 Agenda for Sustainable Development. Resolut 25,135 (2015).
8. Fuso Nerini, F. et al. Mapping synergies and trade-offs between energy and the
Sustainable Development Goals. Nat. Energy 3,1015 https://doi.org/10.1038/
s41560-017-0036-5 (2017).
9. Fuso Nerini, F. et al. Connecting climate action with other Sustainable
Development Goals. Nat. Sustain.1, 674680 (2019). https://doi.org/10.1038/
s41893-019-0334-y
10. Fuso Nerini, F. et al. Use SDGs to guide climate action. Nature 557,https://
doi.org/10.1038/d41586-018-05007-1 (2018).
11. United Nations Economic and Social Council. Sustainable Development
(United Nations Economic and Social Council, 2019).
12. Stockholm Resilience Centres (SRC) contribution to the 2016 Swedish 2030
Agenda HLPF report (Stockholm University, 2017).
13. International Energy Agency. Digitalization & Energy (International Energy
Agency, 2017).
14. Fuso Nerini, F. et al. A research and innovation agenda for zero-emission
European cities. Sustainability 11, 1692 https://doi.org/10.3390/su11061692
(2019).
15. Jones, N. How to stop data centres from gobbling up the worlds electricity.
Nature 561, 163166 (2018).
16. Truby, J. Decarbonizing Bitcoin: law and policy choices for reducing the
energy consumption of Blockchain technologies and digital currencies. Energy
Res. Soc. Sci. 44, 399410 (2018).
17. Ahmad Karnama, Ehsan Bitaraf Haghighi, Ricardo Vinuesa, (2019) Organic
data centers: A sustainable solution for computing facilities. Results in
Engineering 4:100063
18. Raissi, M., Perdikaris, P. & Karniadakis, G. E. Physics informed deep learning
(part I): data-driven solutions of nonlinear partial differential equations.
arXiv:1711.10561 (2017).
19. Nagano, A. Economic growth and automation risks in developing countries
due to the transition toward digital modernity. Proc. 11th International
Conference on Theory and Practice of Electronic GovernanceICEGOV 18
(2018). https://doi.org/10.1145/3209415.3209442
20. Helbing, D. & Pournaras, E. Society: build digital democracy. Nature 527,
3334 (2015).
21. Helbing, D. et al. in Towards Digital Enlightenment 7398 (Springer
International Publishing, 2019). https://doi.org/10.1007/978-3-319-90869-4_7
22. Nagler, J., van den Hoven, J. & Helbing, D. in Towards Digital Enlightenment
4146 (Springer International Publishing, 2019). https://doi.org/10.1007/978-
3-319-90869-4_5
23. Wegren, S. K. The left behind: smallholders in contemporary Russian
agriculture. J. Agrar. Chang. 18, 913925 (2018).
24. NSF - National Science Foundation. Women and Minorities in the S&E
Workforce (NSF - National Science Foundation, 2018).
25. Helbing, D. The automation of society is next how to survive the digital
revolution; version 1.0 (Createspace, 2015).
26. Cockburn, I., Henderson, R. & Stern, S. The Impact of Articial Intelligence on
Innovation (NBER, 2018). https://doi.org/10.3386/w24449
27. Seo, Y., Kim, S., Kisi, O. & Singh, V. P. Daily water level forecasting using
wavelet decomposition and articial intelligence techniques. J. Hydrol. 520,
224243 (2015).
28. Adeli, H. & Jiang, X. Intelligent Infrastructure: Neural Networks, Wavelets, and
Chaos Theory for Intelligent Transportation Systems and Smart Structures
(CRC Press, 2008).
29. Nunes, I. & Jannach, D. A systematic review and taxonomy of explanations in
decision support and recommender systems. Use. Model Use. Adapt Interact.
27, 393444 (2017).
30. Bissio, R. Vector of hope, source of fear. Spotlight Sustain. Dev.7786 (2018).
31. Brynjolfsson, E. & McAfee, A. The Second Machine Age: Work, Progress, and
Prosperity in a Time of Brilliant Technologies (W. W. Norton & Company,
2014).
32. Dobbs, R. et al. Poorer Than Their Parents? Flat or Falling Incomes in
Advanced Economies (McKinsey Global Institute, 2016).
33. Francescato, D. Globalization, articial intelligence, social networks and
political polarization: new challenges for community psychologists. Commun.
Psychol. Glob. Perspect. 4,2041 (2018).
34. Saam, N. J. & Harrer, A. Simulating norms, social inequality, and functional
change in articial societies. J. Articial Soc.Social Simul.2(1999).
35. Dalenberg, D. J. Preventing discrimination in the automated targeting of job
advertisements. Comput. Law Secur. Rev. 34, 615627 (2018).
36. World Economic Forum (WEF). Fourth Industrial Revolution for the Earth
Series Harnessing Articial Intelligence for the Earth (World Economic Forum,
2018).
37. Vinuesa, R., Fdez. De Arévalo, L., Luna, M. & Cachafeiro, H. Simulations and
experiments of heat loss from a parabolic trough absorber tube over a range of
pressures and gas compositions in the vacuum chamber. J. Renew. Sustain.
Energy 8(2016).
38. Keramitsoglou, I., Cartalis, C. & Kiranoudis, C. T. Automatic identication of
oil spills on satellite images. Environ. Model. Softw. 21, 640652 (2006).
39. Mohamadi, A., Heidarizadi, Z. & Nourollahi, H. Assessing the desertication
trend using neural network classication and object-oriented techniques. J.
Fac. Istanb. Univ. 66, 683690 (2016).
40. Kwok, R. AI empowers conservation biology. Nature 567, 133134 (2019).
41. Bonnefon, J.-F., Shariff, A. & Rahwan, I. The social dilemma of autonomous
vehicles. Science 352, 15731576 (2016).
42. De Fauw, J. et al. Clinically applicable deep learning for diagnosis and referral
in retinal disease. Nat. Med 24, 13421350 (2018).
43. Russell, S., Dewey, D. & Tegmark, M. Research priorities for robust and
benecial articial intelligence. AI Mag. 34, 105114 (2015).
44. World Economic Forum (WEF). The New Physics of Financial Services How
Articial Intelligence is Transforming the Financial Ecosystem (World
Economic Forum, 2018).
45. Gandhi, N., Armstrong, L. J. & Nandawadekar, M. Application of data mining
techniques for predicting rice crop yield in semi-arid climatic zone of India.
2017 IEEE Technological Innovations in ICT for Agriculture and Rural
Development (TIAR) (2017). https://doi.org/10.1109/tiar.2017.8273697
46. Esteva, A. et al. Corrigendum: dermatologist-level classication of skin cancer
with deep neural networks. Nature 546, 686 (2017).
47. Cao, Y., Li, Y., Coleman, S., Belatreche, A. & McGinnity, T. M. Detecting price
manipulation in the nancial market. 2014 IEEE Conference on Computational
Intelligence for Financial Engineering &Economics (CIFEr) (2014). https://doi.
org/10.1109/cifer.2014.6924057
48. Nushi, B., Kamar, E. & Horvitz, E. Towards accountable AI: hybrid human-
machine analyses for characterizing system failure. arXiv:1809.07424 (2018).
49. Beyer, H. L., Dujardin, Y., Watts, M. E. & Possingham, H. P. Solving
conservation planning problems with integer linear programming. Ecol.
Model. 328,1422 (2016).
50. Whittaker, M. et al. AI Now Report 2018 (AI Now Institute, 2018).
51. Petit, M. Towards a critique of algorithmic reason. A state-of-the-art review of
articial intelligence, its inuence on politics and its regulation. Quad. del
CAC 44 (2018).
52. Scholz, R. et al. Unintended side effects of the digital transition: European
scientistsmessages from a proposition-based expert round table.
Sustainability 10, 2001 (2018).
53. Ramirez, E., Brill, J., Maureen, K., Wright, J. D. & McSweeny, T. Data Brokers:
A Call for Transparency and Accountability (Federal Trade Commission,
2014).
54. Panch, T., Mattie, H. & Celi, L. A. The inconvenient truthabout AI in
healthcare. npj Digit. Med 2, 77 (2019).
55. Solaiman, S. M. Legal personality of robots, corporations, idols and
chimpanzees: a quest for legitimacy. Artif. Intell. Law 25, 155179 (2017).
56. West, J. & Bhattacharya, M. Intelligent nancial fraud detection: a
comprehensive review. Comput. Secur 57,4766 (2016).
57. Hajek, P. & Henriques, R. Mining corporate annual reports for intelligent
detection of nancial statement fraud A comparative study of machine
learning methods. Knowl.-Based Syst. 128, 139152 (2017).
58. Perry, W. L., McInnis, B., Price, C. C., Smith, S. C. & Hollywood, J. S.
Predictive Policing: The Role of Crime Forecasting in Law Enforcement
Operations (RAND Corporation, 2013).
59. Gorr, W. & Neill, D. B. Detecting and preventing emerging epidemics of
crime. Adv. Dis. Surveillance 4, 13 (2007).
60. IEEE. Ethically Aligned Design - Version II overview (2018). https://doi.org/
10.1109/MCS.2018.2810458
61. European Commission. Draft Ethics Guidelines for Trustworthy AI (Digital
Single Market, 2018).
62. Lipton, Z. C. The mythos of model interpretability. Commun. ACM 61,3643
(2018).
63. Dignum, V. Responsible Articial Intelligence (Springer International
Publishing, 2019).
NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y PERSPECTIVE
NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications 9
64. Future of Life Institute. Open Letter on Autonomous Weapons (Future of Life
Institute, 2015).
65. Future of Life Institute. Annual Report 2018. https://futureoife.org/wp-
content/uploads/2019/02/2018-Annual-Report.pdf?x51579
66. Montes, G. A. & Goertzel, B. Distributed, decentralized, and democratized
articial intelligence. Technol. Forecast. Soc. Change 141, 354358 (2019).
67. Butler, A. J., Thomas, M. K. & Pintar, K. D. M. Systematic review of expert
elicitation methods as a tool for source attribution of enteric illness. Foodborne
Pathog. Dis. 12, 367382 (2015).
68. Morgan, M. G. Use (and abuse) of expert elicitation in support of decision
making for public policy. Proc. Natl Acad. Sci. USA 111, 71767184 (2014).
69. United Nations Human Rights. Sustainable Development Goals Related
Human Rights (United Nations Human Rights, 2016).
70. Draft Committee. Universal Declaration of Human Rights (United Nations,
1948).
Acknowledgements
R.V. acknowledges funding provided by KTH Sustainability Ofce. I.L. acknowledges the
Swedish Research Council (registration number 2017-05189) and funding through an
Early Career Research Fellowship granted by the Jacobs Foundation. M.B. acknowledges
Implicit SSF: Swedish Foundation for Strategic Research project RIT15-0046. V.D.
acknowledges the support of the Wallenberg AI, Autonomous Systems, and Software
Program (WASP) program funded by the Knut and Alice Wallenberg Foundation. S.D.
acknowledges funding from the Leibniz Competition (J45/2018). S.L. acknowledges
funding from the European Unions Horizon 2020 Research and Innovation Programme
under the Marie SkłodowskaCurie grant agreement number 748625. M.T. was sup-
ported by the Ethics and Governance of AI Fund. F.F.N. acknowledges funding from the
Formas grant number 2018-01253.
Author contributions
R.V. and F.F.N. ideated, designed, and wrote the paper; they also coordinated inputs
from the other authors, and assessed and reviewed SDG evaluations as for the Supple-
mentary Data 1. H.A. and I.L. supported the design, wrote, and reviewed sections of the
paper; they also assessed and reviewed SDG evaluations as for the Supplementary Data 1.
M.B., V.D., S.D., A.F. and S.L. wrote and reviewed sections of the paper; they also
assessed and reviewed SDG evaluations as for the Supplementary Data 1. M.T. reviewed
the paper and acted as nal editor.
Competing interests
The authors declare no competing interests.
Additional information
Supplementary information is available for this paper at https://doi.org/10.1038/s41467-
019-14108-y.
Correspondence and requests for materials should be addressed to R.V. or F.F.N.
Peer review information Nature Communications thanks Dirk Helbing and the other,
anonymous, reviewer(s) for their contribution to the peer review of this work.
Reprints and permission information is available at http://www.nature.com/reprints
Publishers note Springer Nature remains neutral with regard to jurisdictional claims in
published maps and institutional afliations.
Open Access This article is licensed under a Creative Commons
Attribution 4.0 International License, which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give
appropriate credit to the original author(s) and the source, provide a link to the Creative
Commons license, and indicate if changes were made. The images or other third party
material in this article are included in the articles Creative Commons license, unless
indicated otherwise in a credit line to the material. If material is not included in the
articles Creative Commons license and your intended use is not permitted by statutory
regulation or exceeds the permitted use, you will need to obtain permission directly from
the copyright holder. To view a copy of this license, visit http://creativecommons.org/
licenses/by/4.0/.
© The Author(s) 2020
PERSPECTIVE NATURE COMMUNICATIONS | https://doi.org/10.1038/s41467-019-14108-y
10 NATURE COMMUNICATIONS | (2020) 11:233 | https://doi.org/10.1038/s41467-019-14108-y | www.nature.com/naturecommunications
... "Unsupervised learning does not involve labelling data but involves identification of hidden patterns in the data by a machine. Reinforcement learning involves machine learning by trial and error to achieve an objective for which the machine is either rewarded or penalized, depending on whether its inferences reach or hinder achievement of an objective (Vinuesa et al., 2020). Deep learning, also known as deep structural learning, is a family of machine learning based on the use of multi-layered models to progressively extract features from data. ...
... Deep learning can be supervised, unsupervised or semi-supervised. Deep learning generally requires large amounts of data to be fed into the model" (Vinuesa et al 2020;Flynn, 2020). ...
Article
Full-text available
Introduction: Clinical research is a key area in which the use of AI in healthcare data seen a significant increase, even though met with great ethical, legal and regulatory challenges. Artificial Intelligence (AI) concerns the ability of algorithms encoded in technology to learn from data, to be able to perform automated tasks without every step in the process being explicitly to be programmed by a human. AI development relies on big data collected from clinical trials to train algorithms, that requires careful consideration of consent, data origin and ethical standards. When data is acquired from third-party sources, transparency about collection methods, geographic origin and anonymization standards becomes critical. While consent forms used in clinical trials can offer clearer terms for data use, ambiguity remains about how this data can be reused for AI purposes after the trial ends. There are very few or no laws on the use of AI especially in developing countries. Also, there are a lot of misconceptions on the global use of AI. Statement of Objectives: Artificial intelligence as an innovative technology has contributed to a shift in paradigm in conducting clinical research. Unfortunately, AI faces ethical, and regulatory challenges especially in limited resource countries where the technology is still to be consolidated. One of the main concerns of AI involves data re-identification, in which anonymized data can potentially be traced back to individuals, especially when linked with other datasets. Data ownership is also a complex and often controversial area within the healthcare sector. AI developers needs to clearly explain the value of data collection to hospitals and cybersecurity teams to ensure that they understand how the data will be secured and used ethically Methodology: The World Health Organization (WHO) recognizes that AI holds great promise for clinical health research and in the practice of medicine, biomedical and pharmaceutical sciences. WHO also recognizes that, to fully maximize the contribution of AI, there is the need to address the ethical, legal and regulatory challenges for the health care systems, practitioners and beneficiaries of medical and public health services. In this study we have pulled data from accessible websites, peered reviewed open-access publications that deal with the ethical and regulatory concerns of AI, that we have discussed in this writeup. We have attempted to place our focus on the development of AI and applications with particular bias in the ethical and regulatory concerns. We have discussed and given an insight on whether AI can advance the interests of patients and communities within the framework of collective effort to design and implement ethically defensible laws and policies and ethically designed AI technologies. Finally, we have investigated the potential serious negative consequences of ethical principles and human rights obligations if they are not prioritized by those who fund, design, regulate or use AI technologies for health research. Results: From our data mining and access to multiple documentations, vital information has been pooled together by a systematic online search to show that AI is contributing significantly in the growth of global clinical research and advancement of medicine. However, we observed many ethical and regulatory challenges that has impacted health research in developing economies. Ethical challenges include AI and human rights, patient’s privacy, safety and liability, informed consent and data ownership, bias and fairness. For the legal and regulatory challenges, we observed issues with data security compliance, data monitoring and maintenance, transparency and accountability, data collection, data storage and use. The role of third-party vendors in AI healthcare solutions and finally AI development and integration into the health systems has also been reviewed. Conclusion: The advancement of AI, coupled with the innovative digital health technology has made a significant contribution to address some challenges in clinical research, within the domain of medicine, biomedical and pharmaceutical products development. Despite the challenging ethical and regulatory challenges AI has impacted significant innovation and technology in clinical research, especially within the domain of drug discovery and development, and clinical trials studies.
... Persepsi etis masyarakat terhadap penggunaan AI sangat bergantung pada tingkat pemahaman mereka terhadap teknologi tersebut. (Vinuesa et al., 2020) menekankan bahwa literasi digital yang kuat akan membantu individu mengidentifikasi potensi risiko dan manfaat dari teknologi AI, termasuk dalam penggunaan AI Image Generator. Dalam konteks mahasiswa dan masyarakat umum, persepsi etis dipengaruhi oleh faktor pendidikan, nilai moral, norma sosial, serta paparan terhadap teknologi itu sendiri. ...
... Dalam konteks mahasiswa dan masyarakat umum, persepsi etis dipengaruhi oleh faktor pendidikan, nilai moral, norma sosial, serta paparan terhadap teknologi itu sendiri. Oleh karena itu, penelitian ini penting untuk memahami bagaimana persepsi tersebut terbentuk dan berkembang di tengah masyarakat digital saat ini (Vinuesa et al., 2020 ...
Article
Full-text available
Penggunaan AI Image Generator semakin meluas dalam kehidupan sehari-hari, terutama untuk kepentingan pribadi seperti kreasi konten digital, hiburan, dan representasi diri di media sosial. Kemajuan teknologi ini memunculkan berbagai dilema etika, termasuk isu keaslian karya, potensi penyalahgunaan identitas visual, serta pelanggaran hak cipta dan privasi. Penelitian ini bertujuan untuk mengkaji persepsi etis terhadap penggunaan AI Image Generator di kalangan mahasiswa dari berbagai jurusan-termasuk Sistem Informasi dan Desain Komunikasi Visual-serta masyarakat umum. Dengan menggunakan pendekatan kuantitatif melalui metode survei, penelitian ini menghimpun pandangan responden terhadap aspek moral, sosial, dan hukum dari penggunaan teknologi tersebut. Data dikumpulkan secara anonim guna memastikan kejujuran dan kenyamanan partisipan dalam menyampaikan opini mereka. Hasil akhir penelitian menunjukkan bahwa persepsi etis terhadap AI Image Generator dipengaruhi oleh tingkat literasi digital, pemahaman hukum, dan kesadaran etika pengguna. Mayoritas responden menyadari potensi risiko teknologi ini, terutama terkait penyalahgunaan dan pelanggaran hak cipta, meskipun penerimaan sosialnya masih moderat. Oleh karena itu, diperlukan peningkatan edukasi, regulasi yang lebih jelas, serta kampanye literasi etika digital agar teknologi ini dapat dimanfaatkan secara bertanggung jawab.
... The sharing economy connects peers with each other to facilitate direct exchange without traditional intermediaries of offered goods and services (Vinuesa et al., 2020). The model taps into available resources and assets that would otherwise sit idle or go to waste to offer efficient and, often, sustainable alternatives to traditional consumption. ...
... AI can accelerate progress toward the Sustainable Development Goals (SDGS), particularly climate action. Partnerships among governments, academia, and industry are key to developing AI-driven educational tools, like personalised learning systems that cater to learners from diverse backgrounds (Vinuesa et al., 2020). As technology evolves, its role in education extends beyond mere facilitation. ...
Article
Climate crisis is an urgent transnational issue that demands global efforts from various stakeholders. The United Nations and its allied agencies, together with non-governmental organisations, are actively involved in fighting the climate crisis. Technology, particularly artificial intelligence, has long been sought as a solution to mitigate the climate crisis. This paper explores the role of Artificial Intelligence-Powered Technology-Enhanced Learning (AIPTEL) in addressing this global climate crisis. We have proposed the AIPTEL framework that serves as a mechanism to raise environmental awareness and bring about systemic changes in educational infrastructure to reduce its carbon footprint. This dual-role framework aligns with Sustainable Development Goals 4 (Quality Education) and 13 (Climate Action).
Article
The capabilities of 6G networks surpass those of existing networks, aiming to enable seamless connectivity between all entities and users at any given time. A critical aspect of achieving enhanced and ubiquitous mobile broadband, as promised by 6G networks, is merging satellite networks with land-based networks, which offers significant potential in terms of coverage area. Advanced routing techniques in next-generation network technologies, particularly when incorporating terrestrial and non-terrestrial networks, are essential for optimizing network efficiency and delivering promised services. However, the dynamic nature of the network, the heterogeneity and complexity of next-generation networks, and the relative distance and mobility of satellite networks all present challenges that traditional routing protocols struggle to address. This paper provides an in-depth analysis of 6G networks, addressing key enablers, technologies, commitments, satellite networks, and routing techniques in the context of 6G and satellite network integration. To ensure 6G fulfills its promises, the paper emphasizes necessary scenarios and investigates potential bottlenecks in routing techniques. Additionally, it explores satellite networks and identifies routing challenges within these systems. The paper highlights routing issues that may arise in the integration of 6G and satellite networks and offers a comprehensive examination of essential approaches, technologies, and visions required for future advancements in this area. 6G and satellite networks are associated with technical terms such as AI/ML, quantum computing, THz communication, beamforming, MIMO technology, ultra-wide band and multi-band antennas, hybrid channel models, and quantum encryption methods. These technologies will be utilized to enhance the performance, security, and sustainability of future networks.
Article
A comprehensive analysis of two significant artificial intelligence systems, Salesforce's Einstein Reasoning Engine and OpenAI's ChatGPT, is presented. While both systems leverage advanced AI technologies, they serve fundamentally different purposes and operate under distinct paradigms. Their architectural differences, use cases, limitations, and effectiveness in their respective domains are examined. Through empirical analysis and case studies, it is demonstrated that these systems complement rather than compete with each other in the enterprise AI ecosystem.
Article
Im Beitrag werden die Wechselwirkungen zwischen Künstlicher ­ Intelligenz (KI), Klima und Bildung im Kontext nachhaltiger Entwicklung untersucht. Angesichts der globalen Klimakrise und der wachsenden Bedeutung digitaler Technologien werden die ökologischen und sozialen Folgen von KI-Systemen kritisch reflektiert. Im Fokus stehen Bildungskonzepte, die ökologische und ethische Verantwortung fördern und transformative Handlungsmöglichkeiten aufzeigen.
Preprint
Full-text available
The rapid advancement of digital technologies has fundamentally transformed urban development patterns, creating unprecedented opportunities for ecological innovation and sustainable development. This study investigates the impact mechanisms of digital transformation on urban sustainable development through ecological innovation, using panel data from 96 Chinese cities spanning 2015-2023. We construct comprehensive indices for digital transformation, ecological innovation, and sustainable development using principal component analysis, and employ multiple regression models to examine their relationships. Our findings reveal that digital transformation significantly promotes ecological innovation (β=0.973,β=0.973, p<0.001,p<0.001), which in turn enhances urban sustainable development (β=0.921,β=0.921, p<0.001p<0.001). Mediation analysis demonstrates that ecological innovation serves as a crucial mediator, accounting for 87.4% of the total effect of digital transformation on sustainable development. The study also identifies significant regional heterogeneity, with stronger effects observed in eastern regions and larger cities. These results provide empirical evidence for the "digital-ecological-sustainable" development paradigm and offer important policy implications for promoting urban sustainability through digital-ecological synergy. The research contributes to the literature by integrating digital transformation theory, ecological modernization theory, and sustainable development theory into a unified analytical framework, while providing practical guidance for policymakers seeking to leverage digital technologies for environmental and sustainable development goals.
Article
Rapid advancements in artificial intelligence (AI) have led to a substantial increase in energy consumption, particularly during the training phase of AI models. As AI adoption continues to grow, its environmental impact presents a significant challenge to the achievement of the United Nations’ Sustainable Development Goals (SDGs). This study examines how three key training configuration parameters—early-stopping epochs, training data size, and batch size—can be optimized to balance model accuracy and energy efficiency. Through a series of experimental simulations, we analyze the impact of each parameter on both energy consumption and model performance, offering insights that contribute to the development of environmental policies that are aligned with the SDGs. The results demonstrate strong potential for reducing energy usage without compromising model reliability. The results highlight three lessons: promoting early-stopping epochs as an energy-efficient practice, limiting training data size to enhance energy efficiency, and developing standardized guidelines for batch size optimization. The practical applicability of these three lessons is illustrated through the implementation of a smart building attendance system using facial recognition technology within an Ecocampus environment. This real-world application highlights how energy-conscious AI training configurations support sustainable urban innovation and contribute to climate action and environmentally responsible AI development.
Article
Full-text available
In the present perspective article we provide an overview of on-going work in the literature and possible future development of organic data centers (ODC). These are defined as the combined operation of a data center and a greenhouse, and given their compatible thermal and operation requirements, ODCs have the potential to provide an excellent solution in terms of sustainability. In particular, we identify possible positive impacts of ODCs on at least 5 of the 17 United Nations (UN) Sustainable Development Goals (SDGs), including SDGs 2 and 13 on zero hunger and climate change, respectively.
Article
Full-text available
As the UK sits in painful deadlock over Brexit, it is important to remember that governments are regularly faced with crises, and their responses can create enduring benefit for future generations. Back in 1858, for example, the UK parliament was dealing with another messy crisis: “the great stink.” In a world before sanitation, the river Thames had become an open latrine, and as summer blossomed parliament was engulfed in a pestilential stench. £2.5 million (about £300 million in today’s money) was hastily approved to build a network of sewers throughout the capital.1 This particular model of sanitation, developed by Bazalgette, was adopted by other cities around the world and the rest, as they say, is history. It is now unthinkable that a developed nation would not have sanitation infrastructure. However, back in 1858 the debate was whether sanitation infrastructure was worthy of investment and whether it was a public or private good. A similar debate has been simmering for some time regarding health data infrastructure, defined as the hardware and software to securely aggregate, store, process and transmit healthcare data. Is data infrastructure necessary for healthcare organizations and if so, is it the responsibility of individual healthcare organizations, of local health systems, or is it a public good? [First paragraph]
Article
Full-text available
The international community has committed to combat climate change and achieve 17 Sustainable Development Goals (SDGs). Here we explore (dis)connections in evidence and governance between these commitments. Our structured evidence review suggests that climate change can undermine 16 SDGs, while combatting climate change can reinforce all 17 SDGs but undermine efforts to achieve 12. Understanding these relationships requires wider and deeper interdisciplinary collaboration. Climate change and sustainable development governance should be better connected to maximize the effectiveness of action in both domains. The emergence around the world of new coordinating institutions and sustainable development planning represents promising progress.
Article
Full-text available
The Paris Agreement and SDG13 on Climate Action require a global drop in Green House Gases (GHG) emissions to stay within a “well below 2 degrees” climate change trajectory. Cities will play a key role in achieving this, being responsible for 60 to 80% of the global GHG emissions depending on the estimate. This paper describes how Research and Innovation (R&I) can play a key role in decarbonizing European cities, and the role that research and education institutions can play in that regard. The paper highlights critical R&I actions in cities based on three pillars: (1) innovative technology and integration, (2) governance innovation, and (3) social innovation. Further, the research needed to harmonize climate mitigation and adaptation in cities are investigated.
Article
As machine learning systems move from computer-science laboratories into the open world, their accountability becomes a high priority problem. Accountability requires deep understanding of system behavior and its failures. Current evaluation methods such as single-score error metrics and confusion matrices provide aggregate views of system performance that hide important shortcomings. Understanding details about failures is important for identifying pathways for refinement, communicating the reliability of systems in different settings, and for specifying appropriate human oversight and engagement. Characterization of failures and shortcomings is particularly complex for systems composed of multiple machine learned components. For such systems, existing evaluation methods have limited expressiveness in describing and explaining the relationship among input content, the internal states of system components, and final output quality. We present Pandora, a set of hybrid human-machine methods and tools for describing and explaining system failures. Pandora leverages both human and system-generated observations to summarize conditions of system malfunction with respect to the input content and system architecture. We share results of a case study with a machine learning pipeline for image captioning that show how detailed performance views can be beneficial for analysis and debugging.
Book
In this book, the author examines the ethical implications of Artificial Intelligence systems as they integrate and replace traditional social structures in new sociocognitive-technological environments. She discusses issues related to the integrity of researchers, technologists, and manufacturers as they design, construct, use, and manage artificially intelligent systems; formalisms for reasoning about moral decisions as part of the behavior of artificial autonomous systems such as agents and robots; and design methodologies for social agents based on societal, moral, and legal values. Throughout the book the author discusses related work, conscious of both classical, philosophical treatments of ethical issues and the implications in modern, algorithmic systems, and she combines regular references and footnotes with suggestions for further reading. This short overview is suitable for undergraduate students, in both technical and non-technical courses, and for interested and concerned researchers, practitioners, and citizens.
Article
Faced with mountains of image and audio data, researchers are turning to artificial intelligence to answer pressing ecological questions. Faced with mountains of image and audio data, researchers are turning to artificial intelligence to answer pressing ecological questions.
Article
The accelerating investment in artificial intelligence has vast implications for economic and cognitive development globally. However, AI is currently dominated by an oligopoly of centralized mega-corporations, who focus on the interests of their stakeholders. There is a now universal need for AI services by businesses who lack access to capital to develop their own AI services, and independent AI developers lack visibility and a source of revenue. This uneven playing field has a high potential to lead to inequitable circumstances with negative implications for humanity. Furthermore, the potential of AI is hindered by the lack of interoperability standards. The authors herein propose an alternative path for the development of AI: a distributed, decentralized, and democratized market for AIs run on distributed ledger technology. We describe the features and ethical advantages of such a system using SingularityNET, a watershed project being developed by Ben Goertzel and colleagues, as a case study. We argue that decentralizing AI opens the doors for a more equitable development of AI and AGI. It will also create the infrastructure for coordinated action between AIs that will significantly facilitate the evolution of AI into true AGI that is both highly capable and beneficial for humanity and beyond.