PreprintPDF Available

"Weak AI" is Likely to Never Become "Strong AI", So What is its Greatest Value for us?

Authors:
  • E-surfing Digital Life Tech Co., Ltd.
Preprints and early-stage research may not have been peer reviewed yet.

Abstract

AI has surpassed humans across a variety of tasks such as image classification, playing games (e.g., go, "Starcraft" and poker), and protein structure prediction. However, at the same time, AI is also bearing serious controversies. Many researchers argue that little substantial progress has been made for AI in recent decades. In this paper, the author (1) explains why controversies about AI exist; (2) discriminates two paradigms of AI research, termed "weak AI" and "strong AI" (a.k.a. artificial general intelligence); (3) clarifies how to judge which paradigm a research work should be classified into; (4) discusses what is the greatest value of "weak AI" if it has no chance to develop into "strong AI".
arXiv:2103.15294v1 [cs.AI] 29 Mar 2021
1
“Weak AI” is Likely to Never Become “Strong AI”,
So What is its Greatest Value for us?
Bin Liu
First posted March 30th, 2021
Abstract
AI has surpassed humans across a variety of tasks such as image classification, playing games (e.g., go,
“Starcraft” and poker), and protein structure prediction. However, at the same time, AI is also bearing serious
controversies. Many researchers argue that little substantial progress has been made for AI in recent decades. In
this paper, the author (1) explains why controversies about AI exist; (2) discriminates two paradigms of AI research,
termed “weak AI” and “strong AI” (a.k.a. artificial general intelligence); (3) clarifies how to judge which paradigm
a research work should be classified into; (4) discusses what is the greatest value of “weak AI” if it has no chance
to develop into “strong AI”.
Index Terms
Artificial intelligence, artificial general intelligence, deep learning, weak AI, strong AI
I. INT RO DUC TION
The last decade has seen impressive applications of AI represented mostly by deep neural networks,
i.e., deep learning [1]. The striking point lies in that the computing agent has reached and even surpassed
humans in many tasks, e.g., image classification [2], speech recognition [3, 4], games [5–7], protein
structure prediction [8]. Even ten years ago, it was hard to imagine that AI would achieve so many
amazing breakthroughs.
On the other side, AI is also bearing serious controversies during the same period. Among the critics,
Judea Pearl, a pioneer for probabilistic reasoning in AI and a winner of the Turing award, argues that
“... all the impressive achievements of deep learning amount to just curve fitting,” and a necessary ability
to be supplemented for AI is causal reasoning [9, 10]. Gary Marcus, a professor of cognitive science,
B. Liu is with Zhejiang Lab, Hangzhou, China. e-mail: bins@ieee.org or liubin@zhejianglab.com.
2
summarizes ten limitations of deep learning [11], namely, ... it is data-hungry, ... it has limited capacity
for transfer, ... it has no natural way to deal with hierarchical structure, ... it struggles with open-ended
inference, ... it is not sufficiently transparent, ... it has not been well integrated with prior knowledge,
... it cannot inherently distinguish causation from correlation, ... it presumes a largely stable world, in
ways that may be problematic, ... it works well as an approximation, but its answers often cannot be fully
trusted, ... it is difficult to engineer with”. In a recent issue of the journal Frontiers in Psychology, another
cognitive scientist J. Mark Bishop argues that AI is stupid and causal reasoning will not fix it [12].
In this paper, I attempt to concisely respond to current controversies about AI. Specifically, I emphasize
discrimination between two paradigms of AI research, namely “weak AI” and “strong AI” (Section II);
provide a conceptual guide to judge which paradigm a research work should be classified into (Section
II-A), explain why controversies about AI last (Section III), present major views on whether “weak AI”
will grow into “strong AI” (Section IV) and discuss what is the greatest value of “weak AI” if it has no
chance to become “strong AI” (Section V).
II. WH AT DO WE AK AI AND ST RONG AI ME AN?
“Weak AI” and “Strong AI” are two terms coined by John Searle in the “Chinese room argument”
(CRA) [13]. CRA is a thought experiment as follows: Searle imagines himself alone in a room following
a computer program for responding to Chinese characters slipped under the door. Searle understands
nothing of Chinese, and yet, by following the program for manipulating symbols and numerals just as
a computer does, he sends appropriate strings of Chinese characters back out under the door, and this
leads those outside to mistakenly suppose there is a Chinese speaker in the room [14]. The term “strong
AI” entails that, ... the computer is not merely a tool in the study of the mind; rather, the appropriately
programmed computer really is a mind, in the sense that computers given the right programs can be
literally said to understand and have other cognitive states. In contrast, the term “weak AI” implies that
... the principal value of the computer in the study of the mind is that it gives us a very powerful tool. J.
Mark Bishop summarizes that ‘weak AI focuses on epistemic issues relating to engineering a simulation
of human intelligent behavior, whereas strong AI, in seeking to engineer a computational system with all
the causal power of a mind, focuses on the ontological [12].
I borrow the terms “weak AI” and “strong AI” here without an intent to discuss CRA. See related
discussions in e.g., [15–18].
3
Simply put, “weak AI” represents computational systems that exhibit as if they own human intelligence,
but they do not. In contrast, “strong AI” represents computational systems that have human intelligence.
Correspondingly, all AI research can be categorized into two paradigms: one is targeted for realizing
“strong AI”; and the other produces advanced “weak AI” systems to meet a variety of practical needs.
A. How to Judge a Research Work Belongs to Which Paradigm?
The biggest motivation for realizing “strong AI” is to answer the question: what are the generation
mechanisms of humans’ intelligence and how to implement these mechanisms with a machine. Therefore,
given a research work, it is easy to judge whether it belongs to the “strong AI” paradigm. If this work
provides any new and useful clue for us to answer the above question, it falls within the “strong AI”
paradigm; otherwise, it belongs to the “weak AI” paradigm.
Based on the above method, part of the (especially early) works on neural networks that deepen our
understanding of the working mechanism of biological neural systems, surely belongs to the “strong AI”
paradigm. On the other hand, most research works that involve artificial neural networks and deep learning,
even if they are proposed under the inspiration of research on neuroscience, cognitive science, behavior
psychology, they belong to the “weak AI” paradigm as long as they do not give us any new insight on
the generation mechanisms of humans’ intelligence or on how to better implement mechanisms that have
already been found.
III. WHY CO NTROVE RSI E S ABOUT AI LAST?
In controversies about AI, party A believes that AI has made substantial progress in the past decade;
party B doubts or even negates the development of AI.
I argue that controversies arise mainly because these two parties mix two different concepts, “weak
AI” and “strong AI”, together, when they talk about AI. The fact is that “weak AI” has made substantial
progress in the past decade, while “strong AI” has not. Party A thinks that “weak AI” is an important
member of the AI family; progress gained from “weak AI” also belongs to this AI family. In contrast,
in the mind of Party B, there always exists one ideal form of AI, namely a realized “strong AI”, and
the “distance” between current AI and this ideal AI is treated as a criterion for evaluating current AI.
Compared with decades ago, current AI still lacks basic human-level abilities such as causal reasoning
[9], robust decision making [19], commonsense utilization [20], and knowledge transfer, which implies
4
that the “distance” between the realized AI and the ideal “strong AI” has not been remarkably shortened.
Therefore, it is reasonable for party B to doubt or even negate the development of AI.
A natural question arises: how breakthroughs of “weak AI” have come out in the past decade? Judea
Pearl argues that “... all the impressive achievements of deep learning amount to just curve fitting”.
However, the point is that, different from previous fitting methods, deep learning permits to do an
extraordinary fitting - fitting multi-modal big data in an end-to-end way. This deep learning type of
fitting requires a big consumption of both computing and storage resources but avoids labor-intensive
feature engineering. Big data, big computing, and big storage are three requisites that make deep learning
surpass humans in playing Go, image classification, speech recognition, and so on. The luck for deep
learning is that the past decade happens to witness great improvements in sensing technologies, wireless
mobile phones, cloud computing, computing devices, computer storage, and databases, which give birth
to big data, big computing, and big storage required by deep learning.
IV. WILL “WEAK AI GROW INT O STRONG AI”?
A metaphor is often used to reply to this question: the relationship between “weak AI” and “strong AI”
is like that between flying machines and birds. Flying machines are not developed by accurately mimicking
birds’ flying. Birds perform much better in maneuvering than the most advanced flying machine today.
Birds can flexibly re-purpose their behaviors while flying machines cannot. But the appearance of flying
machines has met demands of speedy transportation and others. People may think that, since it is unlikely
and not necessary for flying machines to develop into birds, then similarly, “weak AI” is unlikely and not
necessary to grow into “strong AI”.
To formally consider whether “weak AI” will grow into “strong AI”, let recall the Turing test [21] and the
CRA (mentioned in Section II). An example statement of the Turing test is as follows [22]: Originally
known as the Imitation Game, the test evaluates if a machine’s behavior can be distinguished from a
human. In this test, there is a person known as the “interrogator” who seeks to identify a difference between
computer-generated output and human-generated ones through a series of questions. If the interrogator
cannot reliably discern the machines from human subjects, the machine passes the test. However, if
the evaluator can identify the human responses correctly, then this eliminates the machine from being
categorized as intelligent. Through the lens of CRA, Searle argues that the Turing test has serious flaws,
as passing the test does not indicate that the machine has consciousness or understanding. The absence
of an effective evaluation method hampers the development of “strong AI”.
5
Besides, philosophers and cognitive scientists often use G¨
odel’s first incompleteness theorem [23] to
argue that a machine cannot generate humans’ consciousness or understanding. See related discussions in
e.g., [12].
V. WH AT IS T HE GREATE ST VALU E OF WEAK AI FO R US?
In his most recent paper, Geoffrey Hinton states that The difference between science and philosophy is
that experiments can show that extremely plausible ideas are just wrong and extremely implausible ones,
like learning an entire complicated system by end-to-end gradient decent, are just right [24]. In [25], Judea
Pearl argues that Modern connectionism has in fact been viewed as a Triumph of Radical Empiricism
over its rationalistic rivals. Indeed, the ability to emulate knowledge acquisition processes on digital
machines offer enormously flexible testing grounds in which philosophical theories about the balance
between empiricism and innateness can be submitted to experimental evaluation on digital machines.
Combining their arguments, one can see that they both attribute recent deep learning’s success as a success
of empiricism which is data-driven, other than driven by philosophical theory or intuition.
A very important lesson that can be learned from the fast-pacing development and applications of AI
in the past decade is that deep learning running on big enough data can produce unexpected shortcuts
to solve extremely difficult problems. For example, by combining deep learning, reinforcement learning
[26], and Monte Carlo tree search [27], a computer program AlphaGO [28] can win the human champion
without having to understand any of the Go-playing strategies that have been accumulated by humans
for more than four thousand years. The Generative Pre-trained Transformer 3 (GPT-3) [29] can generate
human-like texts through deep learning without having to understand any syntax or semantics underlying
the texts. It is shown that the greatest value of “weak AI” represented by deep learning lies in that it
provides scalable, less-labor-involved, accurate, and generalizable tools for distilling, representing and
then exploiting patterns hidden from big data. Although such “weak AI” has no real intelligence, to a
large extent it meets urgent needs for scalable, efficient, and accurate processing of big data.
In a foreseeable future, “weak AI” is likely to become more robustly (with e.g., portfolio [19] or
dynamic portfolio methods [30–35]), while a big challenge is how to model “unknown unknowns”; it will
perform more automatically through e.g., auto machine learning [36], but it can not become completely
automatic provided that “strong AI” is realized [37]; it may perform as if it owns abilities of cognition
and understanding, but it does not.
6
VI. CO NCL USI O NS
AI has made great progress in the past decade. It has influenced almost all facets of human society
by providing more efficient algorithmic solutions to representation, management, analysis of multi-modal
big data. Controversies about AI last mainly because “weak AI” becomes so strong while “strong AI” is
almost as weak as it was decades ago. Almost all breakthroughs of AI that have attracted the public’s
attention in the past decade are within the “weak AI” paradigm. “Weak AI” is developing much faster
than expected. Even ten years ago, one could not imagine that a computer program would beat the human
champion soon in playing Go. In contrast, the “fruits” people have got from the “strong AI” paradigm are
not so striking as from “weak AI”. I suggest, when talking about AI in the future, one should better make
a statement in advance whether this talk is about “weak AI” or “strong AI”. In this way, more focused
and constructive discussions can be expected.
In a foreseeable future, “weak AI” cannot develop into “strong AI” (see why in Section IV), but it
provides a channel to synthesize advances obtained from related disciplines such as cloud computing,
computer storage, high-speed wireless mobile communications. Through this synthesis of technologies,
more advanced algorithmic tools will be developed in the “weak AI” paradigm, then “weak AI” will
continue to influence human society more profoundly, through big data. The man-computer symbiosis
world that Licklider predicted more than sixty years ago [38] is becoming a reality.
REFER ENC E S
[1] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015.
[2] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks, Advances in neural
information processing systems, vol. 25, pp. 1097–1105, 2012.
[3] W. Xiong, L. Wu, F. Alleva, J. Droppo, X. Huang, and A. Stolcke, “The microsoft 2017 conversational speech recognition system,” in
IEEE int’l conf. on acoustics, speech and signal processing (ICASSP). IEEE, 2018, pp. 5934–5938.
[4] G. Saon, G. Kurata, T. Sercu et al., “English conversational telephone speech recognition by humans and machines,” arXiv preprint
arXiv:1703.02136, 2017.
[5] D. Silver, T. Hubert, J. Schrittwieser et al., A general reinforcement learning algorithm that masters chess, shogi, and go through
self-play, Science, vol. 362, no. 6419, pp. 1140–1144, 2018.
[6] N. Brown and T. Sandholm, “Superhuman ai for heads-up no-limit poker: Libratus beats top professionals,” Science, vol. 359, no.
6374, pp. 418–424, 2018.
[7] O. Vinyals, I. Babuschkin, W. M. Czarnecki et al., “Grandmaster level in starcraft ii using multi-agent reinforcement learning,” Nature,
vol. 575, no. 7782, pp. 350–354, 2019.
[8] A. W. Senior, R. Evans, J. Jumper et al., “Improved protein structure prediction using potentials from deep learning,” Nature, vol. 577,
no. 7792, pp. 706–710, 2020.
7
[9] J. Pearl, “The limitations of opaque learning machines,” Possible minds: twenty-five ways of looking at AI, pp. 13–19, 2019.
[10] J. Pearl and D. Mackenzie, “Ai can’t reason why,” Wall Street Journal, 2018.
[11] G. Marcus, “Deep learning: A critical appraisal,” arXiv preprint arXiv:1801.00631, 2018.
[12] J. M. Bishop, “Artificial intelligence is stupid and causal reasoning will not fix it,” Frontiers in Psychology, vol. 11, pp. 1–18, 2021.
[13] S. John, “Minds, brains, and programs,” Behavioral and Brain Sciences, vol. 3, no. 3, pp. 417–457, 1980.
[14] D. Cole, “The chinese room argument,” https://plato.stanford.edu/entries/chinese-room/.
[15] G. Rey, “What’s really going on in Searle’s “Chinese room”,” Philosophical Studies, vol. 50, no. 2, pp. 169–85, 1986.
[16] M. J. Shaffer, A logical hole in the chinese room, Minds and Machines, vol. 19, no. 2, pp. 229–235, 2009.
[17] A. Sloman and M. Croucher, “How to turn an information processor into an understander,” Behavioral and Brain Sciences, vol. 3,
no. 3, pp. 447–448, 1980.
[18] M. A. Boden, Computer models of mind: Computational approaches in theoretical psychology. Cambridge University Press, 1988.
[19] T. G. Dietterich, “Steps toward robust artificial intelligence,” AI Magazine, vol. 38, no. 3, pp. 3–24, 2017.
[20] G. Marcus, “The next decade in AI: four steps towards robust artificial intelligence,” arXiv preprint arXiv:2002.06177, 2020.
[21] A. M. Turing, “Computing machinery and intelligence,” in Parsing the turing test. Springer, 2009, pp. 23–65.
[22] IBM Cloud Education, “Strong AI,” https://www.ibm.com/cloud/learn/strong-ai.
[23] P. Raatikainen, “G¨
odel’s incompleteness theorems, https://plato.stanford.edu/entries/goedel-incompleteness/.
[24] G. Hinton, “How to represent part-whole hierarchies in a neural network,” arXiv preprint arXiv:2102.12627, 2021.
[25] J. Pearl, “Radical empiricism and machine learning research,” Causal Analysis in Theory and Practice (Blog), vol. 26, 2020.
[26] R. S. Sutton and A. G. Barto, Reinforcement learning: An introduction. MIT press, 2018.
[27] S. Gelly and D. Silver, “Monte-carlo tree search and rapid action value estimation in computer go, Artificial Intelligence, vol. 175,
no. 11, pp. 1856–1875, 2011.
[28] D. Silver, A. Huang, C. J. Maddison et al., “Mastering the game of go with deep neural networks and tree search,” Nature, vol. 529,
no. 7587, pp. 484–489, 2016.
[29] T. B. Brown, B. Mann, N. Ryder et al., “Language models are few-shot learners,” arXiv preprint arXiv:2005.14165, 2020.
[30] B. Liu, Y. Qi, and K. Chen, “Sequential online prediction in the presence of outliers and change points: an instant temporal structure
learning approach, Neurocomputing, vol. 413, pp. 240–258, 2020.
[31] Y. Qi, B. Liu, Y. Wang, and G. Pan, “Dynamic ensemble modeling approach to nonstationary neural decoding in brain-computer
interfaces,” Advances in neural information processing systems, pp. 6087–6096, 2019.
[32] B. Liu, “Robust particle filter by dynamic averaging of multiple noise models,” in IEEE Int’l Conf. on Acoustics, Speech and Signal
Processing (ICASSP). IEEE, 2017, pp. 4034–4038.
[33] Y. Dai and B. Liu, “Robust video object tracking via bayesian model averaging-based feature fusion, Optical Engineering, vol. 55,
no. 8, pp. 1–11, 2016.
[34] B. Liu, “Data-driven model set design for model averaged particle filter, in IEEE Int’l Conf. on Acoustics, Speech and Signal Processing
(ICASSP). IEEE, 2020, pp. 5835–5839.
[35] ——, “Instantaneous frequency tracking under model uncertainty via dynamic model averaging and particle filtering,” IEEE Trans. on
Wireless Communications, vol. 10, no. 6, pp. 1810–1819, 2011.
[36] F. Hutter, L. Kotthoff, and J. Vanschoren, Automated machine learning: methods, systems, challenges. Springer Nature, 2019.
[37] B. Liu, “A very brief and critical discussion on automl,” arXiv preprint arXiv:1811.03822, 2018.
[38] J. Licklider, “Man-computer symbiosis, IRE Transactions on human factors in electronics, no. 1, pp. 4–11, 1960.
... Prior studies have focused either on subsets of trauma patients (geriatric trauma and motor vehicle crashes), or the complete prehospital patient group, which could be argued to lower the prediction performance due to poor generalization of performance across target domains [25]. Studies on adult trauma have focused on one particular model, a small set of predictors and/or listwise deletion of missing data, which may benefit the performance of simpler models. ...
Article
Full-text available
Background Providing optimal care for trauma, the leading cause of death for young adults, remains a challenge e.g., due to field triage limitations in assessing a patient’s condition and deciding on transport destination. Data-driven On Scene Injury Severity Prediction (OSISP) models for motor vehicle crashes have shown potential for providing real-time decision support. The objective of this study is therefore to evaluate if an Artificial Intelligence (AI) based clinical decision support system can identify severely injured trauma patients in the prehospital setting. Methods The Swedish Trauma Registry was used to train and validate five models – Logistic Regression, Random Forest, XGBoost, Support Vector Machine and Artificial Neural Network – in a stratified 10-fold cross validation setting and hold-out analysis. The models performed binary classification of the New Injury Severity Score and were evaluated using accuracy metrics, area under the receiver operating characteristic curve (AUC) and Precision-Recall curve (AUCPR), and under- and overtriage rates. Results There were 75,602 registrations between 2013–2020 and 47,357 (62.6%) remained after eligibility criteria were applied. Models were based on 21 predictors, including injury location. From the clinical outcome, about 40% of patients were undertriaged and 46% were overtriaged. Models demonstrated potential for improved triaging and yielded AUC between 0.80–0.89 and AUCPR between 0.43–0.62. Conclusions AI based OSISP models have potential to provide support during assessment of injury severity. The findings may be used for developing tools to complement field triage protocols, with potential to improve prehospital trauma care and thereby reduce morbidity and mortality for a large patient population.
Article
Full-text available
This article does not describe a working system. Instead, it presents a single idea about representation that allows advances made by several different groups to be combined into an imaginary system called GLOM.¹ The advances include transformers, neural fields, contrastive representation learning, distillation, and capsules. GLOM answers the question: How can a neural network with a fixed architecture parse an image into a part-whole hierarchy that has a different structure for each image? The idea is simply to use islands of identical vectors to represent the nodes in the parse tree. If GLOM can be made to work, it should significantly improve the interpretability of the representations produced by transformer-like systems when applied to vision or language.
Article
Full-text available
I contrast the “data fitting” vs “data interpreting” approaches to data science along three dimensions: Expediency, Transparency, and Explainability. “Data fitting” is driven by the faith that the secret to rational decisions lies in the data itself. In contrast, the data-interpreting school views data, not as a sole source of knowledge but as an auxiliary means for interpreting reality, and “reality” stands for the processes that generate the data. I argue for restoring balance to data science through a task-dependent symbiosis of fitting and interpreting, guided by the Logic of Causation.
Article
Full-text available
Artificial Neural Networks have reached “grandmaster” and even “super-human” performance across a variety of games, from those involving perfect information, such as Go, to those involving imperfect information, such as “Starcraft”. Such technological developments from artificial intelligence (AI) labs have ushered concomitant applications across the world of business, where an “AI” brand-tag is quickly becoming ubiquitous. A corollary of such widespread commercial deployment is that when AI gets things wrong—an autonomous vehicle crashes, a chatbot exhibits “racist” behavior, automated credit-scoring processes “discriminate” on gender, etc.—there are often significant financial, legal, and brand consequences, and the incident becomes major news. As Judea Pearl sees it, the underlying reason for such mistakes is that “... all the impressive achievements of deep learning amount to just curve fitting.” The key, as Pearl suggests, is to replace “reasoning by association” with “causal reasoning” —the ability to infer causes from observed phenomena. It is a point that was echoed by Gary Marcus and Ernest Davis in a recent piece for the New York Times: “we need to stop building computer systems that merely get better and better at detecting statistical patterns in data sets—often using an approach known as ‘Deep Learning’—and start building computer systems that from the moment of their assembly innately grasp three basic concepts: time, space, and causality.” In this paper, foregrounding what in 1949 Gilbert Ryle termed “a category mistake”, I will offer an alternative explanation for AI errors; it is not so much that AI machinery cannot “grasp” causality, but that AI machinery (qua computation) cannot understand anything at all.
Article
Full-text available
In this paper, we consider sequential online prediction (SOP) for streaming data in the presence of outliers and change points. We propose an INstant TEmporal structure Learning (INTEL) algorithm to address this problem. Our INTEL algorithm is developed based on a full consideration of the duality between online prediction and anomaly detection. We first employ a mixture of weighted Gaussian process models (WGPs) to cover the expected possible temporal structures of the data. Then, based on the rich modeling capacity of this WGP mixture, we develop an efficient technique to instantly learn (capture) the temporal structure of the data that follows a regime shift. This instant learning is achieved only by adjusting one hyper-parameter value of the mixture model. A weighted generalization of the product of experts (POE) model is used for fusing predictions yielded from multiple GP models. An outlier is declared once a real observation seriously deviates from the fused prediction. If a certain number of outliers are consecutively declared, then a change point is declared. Extensive experiments are performed using a diverse of real datasets. Results show that the proposed algorithm is significantly better than benchmark methods for SOP in the presence of outliers and change points.
Article
Full-text available
Protein structure prediction can be used to determine the three-dimensional shape of a protein from its amino acid sequence¹. This problem is of fundamental importance as the structure of a protein largely determines its function²; however, protein structures can be difficult to determine experimentally. Considerable progress has recently been made by leveraging genetic information. It is possible to infer which amino acid residues are in contact by analysing covariation in homologous sequences, which aids in the prediction of protein structures³. Here we show that we can train a neural network to make accurate predictions of the distances between pairs of residues, which convey more information about the structure than contact predictions. Using this information, we construct a potential of mean force⁴ that can accurately describe the shape of a protein. We find that the resulting potential can be optimized by a simple gradient descent algorithm to generate structures without complex sampling procedures. The resulting system, named AlphaFold, achieves high accuracy, even for sequences with fewer homologous sequences. In the recent Critical Assessment of Protein Structure Prediction⁵ (CASP13)—a blind assessment of the state of the field—AlphaFold created high-accuracy structures (with template modelling (TM) scores⁶ of 0.7 or higher) for 24 out of 43 free modelling domains, whereas the next best method, which used sampling and contact information, achieved such accuracy for only 14 out of 43 domains. AlphaFold represents a considerable advance in protein-structure prediction. We expect this increased accuracy to enable insights into the function and malfunction of proteins, especially in cases for which no structures for homologous proteins have been experimentally determined⁷.
Article
Full-text available
Many real-world applications require artificial agents to compete and coordinate with other agents in complex environments. As a stepping stone to this goal, the domain of StarCraft has emerged as an important challenge for artificial intelligence research, owing to its iconic and enduring status among the most difficult professional esports and its relevance to the real world in terms of its raw complexity and multi-agent challenges. Over the course of a decade and numerous competitions1–3, the strongest agents have simplified important aspects of the game, utilized superhuman capabilities, or employed hand-crafted sub-systems⁴. Despite these advantages, no previous agent has come close to matching the overall skill of top StarCraft players. We chose to address the challenge of StarCraft using general-purpose learning methods that are in principle applicable to other complex domains: a multi-agent reinforcement learning algorithm that uses data from both human and agent games within a diverse league of continually adapting strategies and counter-strategies, each represented by deep neural networks5,6. We evaluated our agent, AlphaStar, in the full game of StarCraft II, through a series of online games against human players. AlphaStar was rated at Grandmaster level for all three StarCraft races and above 99.8% of officially ranked human players.
Book
This open access book presents the first comprehensive overview of general methods in Automated Machine Learning (AutoML), collects descriptions of existing systems based on these methods, and discusses the first series of international challenges of AutoML systems. The recent success of commercial ML applications and the rapid growth of the field has created a high demand for off-the-shelf ML methods that can be used easily and without expert knowledge. However, many of the recent machine learning successes crucially rely on human experts, who manually select appropriate ML architectures (deep learning architectures or more traditional ML workflows) and their hyperparameters. To overcome this problem, the field of AutoML targets a progressive automation of machine learning, based on principles from optimization and machine learning itself. This book serves as a point of entry into this quickly-developing field for researchers and advanced students alike, as well as providing a reference for practitioners aiming to use AutoML in their work.
Article
One program to rule them all Computers can beat humans at increasingly complex games, including chess and Go. However, these programs are typically constructed for a particular game, exploiting its properties, such as the symmetries of the board on which it is played. Silver et al. developed a program called AlphaZero, which taught itself to play Go, chess, and shogi (a Japanese version of chess) (see the Editorial, and the Perspective by Campbell). AlphaZero managed to beat state-of-the-art programs specializing in these three games. The ability of AlphaZero to adapt to various game rules is a notable step toward achieving a general game-playing system. Science , this issue p. 1140 ; see also pp. 1087 and 1118
Article
Libratus versus humans Pitting artificial intelligence (AI) against top human players demonstrates just how far AI has come. Brown and Sandholm built a poker-playing AI called Libratus that decisively beat four leading human professionals in the two-player variant of poker called heads-up no-limit Texas hold'em (HUNL). Over nearly 3 weeks, Libratus played 120,000 hands of HUNL against the human professionals, using a three-pronged approach that included precomputing an overall strategy, adapting the strategy to actual gameplay, and learning from its opponent. Science , this issue p. 418