ArticlePDF Available

Introduction to ‘Cognitive artificial intelligence’

The Royal Society
Philosophical Transactions A
Authors:
royalsocietypublishing.org/journal/rsta
Introduction
Cite this article: Bundy A, Chater N,
Muggleton S. 2023 Introduction to ‘Cognitive
articial intelligence’. Phil.Trans.R.Soc.A381:
20220051.
https://doi.org/10.1098/rsta.2022.0051
Received: 5 April 2023
Accepted: 5 April 2023
One contribution of 11 to a discussion meeting
issue ‘Cognitive articial intelligence’.
Subject Areas:
articial intelligence
Author for correspondence:
Alan Bundy
e-mail: bundy@ed.ac.uk
Introduction to ‘Cognitive
articial intelligence’
Alan Bundy1,NickChater
2and Stephen Muggleton3
1The University of Edinburgh, Edinburgh, Edinburgh, UK
2Warwick Business School, University of Warwick, Coventry, West
Midlands, UK
3Computational Bioinformatics Laboratory, Imperial College
London, UK
AB, 0000-0002-0578-6474;NC,0000-0002-9745-0686;
SM, 0000-0001-6061-6104
1. Introduction
There is an increasing excitement concerning the
potential of artificial intelligence to both transform
human society and to understand cognition in humans
and other animals. This meeting addressed the leading
edge of research intersection of artificial intelligence and
cognitive science, an area we are calling cognitive artificial
intelligence. Topics covered include:
Improving the interaction between humans and
machines, including how machines can explain
their reasoning to humans, might be more
socially aware and understand a human’s beliefs
and intentions.
Contrasting how machines and humans learn,
and showing how machines might emulate
humans in learning from only a few examples
and how machines can aid the teaching of
humans.
How reasoning and learning interact, including
how failures of reasoning trigger the evolution
of models of the environment.
The contrast between symbolic and subsymbolic
reasoning, especially the role of large language
models, such as ChatGPT, in generating natural
language and serving as a model of human
cognition.
This special issue is the proceedings of a Royal Society
Hooke Meeting on Cognitive artificial intelligence.
2023 The Authors. Published by the Royal Society under the terms of the
Creative Commons Attribution License http://creativecommons.org/licenses/
by/4.0/, which permits unrestricted use, provided the original author and
source are credited.
2
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
The meeting was proposed by the EPSRC Network+on Human-Like Computing (HLC). According
to the influential US funding agency DARPA (originator of the Internet and Self-Driving Cars),
this new area represents the third wave of artificial intelligence (3AI, 2020s–2030s), and is being
actively investigated in the USA, Europe and China. The HLC Network was one of the first
internationally to initiate and support research specifically in this area. Starting activities in 2018,
the Network represents around 60 leading UK groups of artificial intelligence and cognitive
scientists involved in the development of the inter-disciplinary area of HLC. The research of
network groups aims to address key unsolved problems at the interface between Psychology
and Computer Science.
2. Objectives
The key fields brought together at this meeting are artificial intelligence and cognitive science.
This meeting helped forge better understanding and interactions in a joint area which we refer
to as cognitive artificial intelligence. Artificial intelligence and machine learning are becoming
centrally relevant to a variety of sciences in supporting the construction of complex models from
data. Furthermore, within society at large artificial intelligence is viewed as having both immense
potential for enabling human societies, while simultaneously presenting dangers for weakening
the social fabric. It is clear that advances in understanding of how to build automated learning
systems which are compatible with human understanding, planning and reasoning has immense
potential for beneficial effects in many areas. However, interactions of cognitive scientists with
leading edge artificial intelligence research requires many advances and new experimental work
in psychology to further understand the cognitive and social constraints of human beings when
interacting with machines. The meeting allowed presentations on the latest results from leading
laboratories in this area, as well as encouraging discussion on key topics for joint research between
artificial intelligence and cognitive science groups.
3. Signicance
Both artificial intelligence and cognitive science have a variety of large-scale annual conferences.
However, researchers within each of these areas typically have limited understanding of advances
in each other’s fields. The meeting helped bring together leading scientists from artificial
intelligence and cognitive science to inform each other of key open questions that joint work
could help address.
4. Social implications
In recent years, there has been increasing public concern about the application of artificial
intelligence. Such concerns were documented within the House of Commons and Lords Select
Committee Reports and the Royal Society Report on Machine Learning. Key issues raised
included those of the need for (a) transparent decision making, (b) accountability and its related
legal implications, (c) safety of automated systems in control tasks and (d) the threat to jobs.
Research in the new area of cognitive artificial intelligence will aim to advance fundamental
understanding for the key artificial intelligence technologies being developed. This meeting has
the potential to promote understanding of the advances required to go beyond the development
of simple black-box decision makers to allow the development of systems which take account
of our understanding human modes of perception and social interaction. Such advances have
potential for wide-ranging social benefit.
5. Points of view
Since both artificial intelligence and cognitive science are both well-established fields, there is
considerable diversity of viewpoints within each field. Within artificial intelligence, this tends to
3
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
be related to the choice of representation used for representing knowledge, while in Cognitive
Science, there are a wide variety of differences in methodology and experimental techniques. The
meeting was devised to include representative from these diverse communities and viewpoints,
with the aim of encouraging wide-ranging discussion.
6. Overview of the contributions
This issue of Cognitive artificial intelligence consists of 11 substantive papers, drawn roughly
equally from the artificial intelligence and cognitive science communities, but each drawing on
and having relevance to both.
We begin with Muggleton’s [1] paper ‘Hypothesizing an algorithm from one example: the role
of specificity’ which argues that while the dominant methods of Statistical Machine Learning
achieve high accuracy, they require large numbers of examples to do so. By contrast, humans
typically learn new concepts from as few as one example. However, the high data efficiency
of human learning cannot be explained by existing standard formal frameworks for machine
learning. Muggleton shows that this disparity can be resolved by introducing a revised Bayesian
framework for expected error, and shows that highly specific concepts, as typified by computer
algorithms, can be learned within this theoretical framework with high accuracy from a single
example, by using a preference for specificity combined with minimality. Experiments with
Muggleton’s implementation of this approach, called DeepLog, indicate that such an approach
can be used in practice to efficiently construct relatively complex logic programs from a single
randomly selected example.
Wahlster’s [2] paper ‘Understanding computational dialogue understanding’ first explains
why human-like dialogue understanding is so difficult for AI. It discusses various methods
for testing the understanding capabilities of dialogue systems. It reviews the development of
dialogue systems over five decades, focusing on the transition from closed-domain to open-
domain systems and their extension to multimodal, multiparty and multilingual dialogues. From
being somewhat of a niche topic in artificial intelligence research for the first 40 years, it has made
newspaper headlines in recent years and is now being discussed by political leaders at events such
as the World Economic Forum in Davos. It asks whether large language models are super-parrots
or a milestone towards human-like dialogue understanding and how they relate to what we know
about language processing in the human brain. Using ChatGPT as an example, it presents some
limitations of this approach to dialogue systems. Finally, it presents some lessons learnt from
40 years of research in this field about system architecture principles: symmetric multimodality,
no presentation without representation and anticipation feedback loops. It concludes with a
discussion of grand challenges such as satisfying conversational maxims and the European
Language Equality Act through massive digital multilinguality—perhaps enabled by interactive
machine learning with human trainers.
The next article, ‘Symbols and grounding in large language models’ by Pavlick [3], considers
the practical and theoretical significance of large language models, which have in the last few
years been shown to carry out a large number of open-ended natural language tasks with often
close to human levels of performance on some measures. These models consist of very large
deep neural networks trained on a substantial fraction of the entire contents of the World Wide
Web. Within the cognitive science community, many have argued that models trained purely
on a large amount of language data, however impressive their performance, are inevitably
restricted in their relevance to human cognition. Pavlick takes up two specific charges against
the cognitive relevance of such models and argues for a verdict of ‘not proven’ in both cases. The
first issue she addresses is that large language models are not endowed with structured symbolic
representations, which are widely presumed to underpin perception, thought and language in
humans. But she notes, drawing on her research, that the internal representations learned by
large language models may actually have a distinctly symbolic character, and that the nature of
such representations can only be determined by sophisticated analysis of how large language
models work. The second issue is that large language models are sometimes presumed not to
4
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
be ‘grounded’ through the perceptuo-motor interaction with the world (although links between
large language models and models of visual processing are relatively advanced). But in any case,
Pavlick argues that modern philosophy of language assumes that the grounding of linguistic
symbols is a collective achievement, at the level of the entire language community, rather than
operating individual by individual. Thus, the lack of direct grounding may raise no special
difficulties for large language models, even if these are not integrated with modules for perception
and action. They may thus inherit grounded symbols from the human language on which they
are trained.
We next move from language to mathematical and scientific cognition. In ‘DreamCoder:
growing generalizable, interpretable knowledge with wake-sleep Bayesian program learning’,
Ellis et al.[4] introduce a system that learns to solve a wide range of representationally
challenging problems by learning to write programs. Combining symbolic and neural network
methods, it learns programming languages for capturing concepts relevant to the target domain,
and uses neural networks to guide the search process to create appropriate programs using
these languages. They apply a ‘wake-sleep’ algorithm, which interleaves the extension of
the programming language with new symbolic abstractions and training the neural network
on imagined and past problems. Dreamcoder can be applied successfully to a wide range
of problems, from drawing pictures and building scenes, to rediscovering the fundamentals
of functional programming, vector algebra and classical physics, including Newton’s and
Coulomb’s Laws. Learning operates by successively creating new abstractions from previous
abstractions, creating rich systems of representation which transfer effectively across task
domains.
Goodman & Poesia [5] continue the theme of how machines can learn to engage in rich,
structured representation and reasoning, now focusing on mathematics, in their paper ‘Peano:
learning formal mathematical reasoning’. They note that while mathematics is created slowly,
involving a huge collective intellectual effort over many centuries, it can relatively rapidly
be taught afresh to each generation of students, who can learn to apply it successfully also
from a very limited set of training examples. Goodman & Poesia argue that fundamental to
mathematical discovery and learning is the ability to create and reason over representations
at ever-increasing levels of abstraction. The computational model, Peano, is a theorem proving
environment which has the power to represent a wide range of aspects of mathematics. They
show that attempting to learn mathematical regularities using traditional reinforcement learning
methods is unsuccessful; but adding the ability to learn reusable abstractions (which they call
‘tactics’) from past problem-solving attempts, allows the agent to make cumulative progress.
The way in which these abstractions are generated sheds light on the ‘natural’ order in which
such abstractions should most helpfully be presented human learners—and this order agrees to a
substantial degree with the order in which ideas are introduced in learning curricula for human
learners, such as that used in the Kahn Academy. Their work raises the possibility that deeper
understanding of learning mathematics using automated methods may shed substantial might
both on the process by which humans learn mathematical concepts, and the optimal design of
mathematical curricula.
Bundy & Li’s [6] paper ‘Representational change is integral to reasoning’ proposes a
mechanism by which language evolves in response to reasoning failures. For instance, concepts
may be split (mother into birth mother and step mother) or merged (Morning Star and Evening
Star into Venus) when current theories either predict things observed to be false or fail to predict
things observed to be true. They start by illustrating that concept evolution occurs even in
mathematics. An examination of Imre Lakatos’s classic rational reconstruction of the history of
Euler’s Theorem (V+FE=2) about polyhedra shows that the initial concept of polyhedron was
not fully defined and that potential counter-examples can be included or excluded depending
on how this initial definition is refined. They then discuss their ABC system that evolves logical
theories by a combination of abduction, belief revision and conceptual change which reconcile an
initial theory’s predictions with conflicting observations of the environment, leading to a revised
theory.
5
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
The next article, ‘Argument and explanation’ by Hahn & Tesic [7] considers the relationship
between the concepts of argumentation and explanation in the context of philosophy of science
and common-sense reasoning. While the philosopher Carl Hempel saw scientific explanation as
a type of argument, Hahn and Tesic stress that arguments typically play a role in dialogue, in
trying to convince others (and perhaps also oneself) of the truth or rightness of some contested
matter. Here, arguments are in the service of the broader goal of persuasion, and factors beyond
the argument itself (such as who was its source) are crucial. But explanations can often apply
when there is no issue of doubt about the point to be explained: thus, facts ranging from, say, the
blueness of the sky or that a piece of kitchen cheese has been nibbled, may not be in doubt, but still
may stand in need of explanation. A crucial issue here is what makes an explanation satisfying—
what distinguishes chains of reasoning to a particular conclusion that provides a sense of insight
and understanding. Hahn & Tesic provide a review of the state-of-the-art in psychological and
atificial intelligence approaches to both argument and explanation, and point the way for future
research.
Gweon et al.[8] in their paper ‘Beyond imitation: machines that understand and are
understood by humans’ focus on a particular, and especially fundamental, aspect of explanation:
the human ability to infer and reason about the mental states of others from observing their
behaviour. Such inferences may be crucial when attempting to learn from another person; and
equally is crucial from the point of view of the teacher, attempting to infer what the learner
already knows and which actions or words will best help them learn successfully. This type
of social intelligence develops early in humans, but seems difficult to replicate in machines:
artificial intelligence systems currently have a limited ability to understand, or be understood
by, humans with which they interact. Gweon et al. argue that a central goal of AI should be the
creation of genuinely socially intelligent machines, that model and consider the minds of people
they interact with, rather than more superficial social niceties, such as mimicking human facial
expressions, gestures, or patterns of speech. They survey work on the development of human
social intelligence, and human–machine interaction, and argue that this could provide crucial
clues for how to create a new generation of machines that can engage in rich social interactions
with people. Indeed, they argue that integrating cognitive science and artificial intelligence
approaches to understanding social intelligence is likely to advance both our understanding of
ourselves, and the creation of socially intelligent machines which can interact naturally with
people.
A particularly critical aspect of the challenge of building computational models of other
minds—inferring emotional states—is taken up by Houlihan et al.[9] in their paper ‘Emotion
prediction as inference over a generative theory of mind’. They describe a computational model
of emotion prediction, the Inferred Appraisals model, that uses inverse planning to infer mental
states, which can include individual objectives but also ‘social preferences’ such as preference for
equity or the desire to maintain a good reputation in the eyes of others. They show how it is
possible to learn a mapping between these appraisals and 20 labels for emotions (including joy,
relief, guilt and envy), so that the model can quantitatively match human predictions concerning
these predictions in high-stakes game-like interactions. The model shows how social preferences
turn out to be important in predicting almost every emotion; and also captures the flexibility
of human emotion attribution. This work provides a starting point for computational models of
social interaction, crucial both for understanding the nature of human social behaviour and for
building socially sensitive artificial systems.
Continuing the theme of social interaction, in the final contribution to this issue, Chater [10]
asks ‘How could we make a social robot?’ He argues that human intelligence is inherently social
and that the spectacular achievements of our species arise through our ability to cooperate,
collaborate and cumulatively creating languages, social norms, organizational and political
structure, legal and financial systems, and the mathematics science and technology. A genuinely
social robot therefore would be an artificial agent able to ‘join in’ fluently with human projects
and activities, learning, collaboration and contributing alongside us. Chater argues that the
cognitive foundation of this process is a style of reasoning known as ‘virtual bargaining’ according
6
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
to which a pair of intelligent agents is able to coordinate their thoughts and actions by each
asking not merely ‘what should Ithink or do?’ but ‘what should we think or do?’ Chater argues
that answering this question successfully involves simulating what the agent will agree were
they able to engage in prior communication—each party needs to successfully simulate the
outcome of hypothetical bargaining process. Chater illustrates the approach by drawing on prior
experimental work in which people are shown to be astonishingly successful, and highly flexible,
in the use of novel and very restricted communicative signals. Here, the challenge of virtual
bargaining is to agree what a novel signal would most naturally be interpreted to mean. He
argues that this process of virtual bargaining underpins communication ‘in the moment’, and
that distinctive human ability to engage in virtual bargaining underpins the gradual creation of
natural language and complex systems of societal conventions. Successive communicative and
collaborative improvizations, each of which provides useful precedents for the next, explains
the gradual emergence of increasingly systematic patterns in language and behaviour, through
processes of spontaneous order. Thus, the complex machinery underpinning human society
arises, to paraphrase the words of Scottish Enlightenment philosopher Adam Ferguson, through
human action but not by human design. Chater suggests that by apparently skipping over
the subtle process of virtual bargaining that underpins human communication, large language
models (as discussed by Pavlick) may currently be missing out what may be a crucial step in
understanding human social behaviour, and how it may be replicated in machines.
Overall, the contribution to this special issue on Cognitive artificial intelligence highlights
convergent and overlapping research in both cognitive science and in artificial intelligence which
is likely to be crucial to building both the next generation of increasingly human-like artificial
systems and also providing a deeper understanding of the human mind.
Data accessibility. This article has no additional data.
Authors’ contributions. A.B.: writing—original draft, writing—review and editing; N.C.: writing—original draft,
writing—review and editing; S.M.: writing—original draft, writing—review and editing.
All authors gave final approval for publication and agreed to be held accountable for the work performed
therein.
Conict of interest declaration. This theme issue was put together by the Guest Editor team under supervision
from the journal’s Editorial staff, following the Royal Society’s ethical codes and best-practice guidelines.
The Guest Editor team invited contributions and handled the review process. Individual Guest Editors were
not involved in assessing papers where they had a personal, professional or financial conflict of interest with
the authors or the research described. Independent reviewers assessed all papers. Invitation to contribute did
not guarantee inclusion.
Funding. We received no funding for this study.
References
1. Muggleton FREng SH. 2023 Hypothesizing an algorithm from one example: the role of
specificity. Phil. Trans. R. Soc. A 381, 20220046. (doi:10.1098/rsta.2022.0046)
2. Wahlster W. 2023 Understanding computational dialogue understanding. Phil. Trans. R. Soc.
A381, 20220049. (doi:10.1098/rsta.2022.0049)
3. Pavlick E. 2023 Symbols and grounding in large language models. Phil. Trans. R. Soc. A 381,
20220041. (doi:10.1098/rsta.2022.0041)
4. Ellis K, Wong L, Nye M, Sablé-Meyer M, Cary L, Anaya Pozo L, Hewitt L, Solar-Lezama
A, Tenenbaum JB. 2023 DreamCoder: growing generalizable, interpretable knowledge with
wake–sleep Bayesian program learning. Phil. Trans. R. Soc. A 381, 20220050. (doi:10.1098/
rsta.2022.0050)
5. Poesia G, Goodman ND. 2023 Peano: learning formal mathematical reasoning. Phil. Trans. R.
Soc. A 381, 20220044. (doi:10.1098/rsta.2022.0044)
6. Bundy A, Li X. 2023 Representational change is integral to reasoning. Phil. Trans. R. Soc. A 381,
20220052. (doi:10.1098/rsta.2022.0052)
7. Hahn U, Teši´c M. 2023 Argument and explanation. Phil. Trans. R. Soc. A 381, 20220043.
(doi:10.1098/rsta.2022.0043)
7
royalsocietypublishing.org/journal/rsta Phil.Trans.R.Soc.A381: 20220051
...............................................................
8. Gweon H, Fan J, Kim B. 2023 Socially intelligent machines that learn from humans and help
humans learn. Phil. Trans. R. Soc. A 381, 20220048. (doi:10.1098/rsta.2022.0048)
9. Houlihan SD, Kleiman-Weiner M, Hewitt LB, Tenenbaum JB, Saxe R. 2023 Emotion prediction
as computation over a generative theory of mind. Phil. Trans. R. Soc. A 381, 20220047.
(doi:10.1098/rsta.2022.0047)
10. Chater N. 2023 How could we make a social robot? A virtual bargaining approach. Phil. Trans.
R. Soc. A 381, 20220040. (doi:10.1098/rsta.2022.0040)
... It is designed to replicate cognitive processes like perception, language processing, and decision-making, these systems draw from a diverse range of disciplines, including computer science, mathematics, psychology, and linguistics. AI technology has penetrated all aspects of human activities (12)(13)(14)(15)). In the cancer research field, AI is characterized by the use of machine learning and deep learning algorithms (16), which are important in processing and analyzing large-scale datasets (17)(18)(19). ...
Article
Full-text available
Liver cancer remains one of the most formidable challenges in modern medicine, characterized by its high incidence and mortality rate. Emerging evidence underscores the critical roles of the immune microenvironment in tumor initiation, development, prognosis, and therapeutic responsiveness. However, the composition of the immune microenvironment of liver cancer (LC-IME) and its association with clinicopathological significance remain unelucidated. In this review, we present the recent developments related to the use of artificial intelligence (AI) for studying the immune microenvironment of liver cancer, focusing on the deciphering of complex high-throughput data. Additionally, we discussed the current challenges of data harmonization and algorithm interpretability for studying LC-IME.
Chapter
Digital technologies, with the forerunner of artificial intelligence (AI), affect many aspects of modern society. With the integration of AI, the education sector has offered innovative educational experiences for both students and teachers. This work aims to give an overview of the implementation of artificial intelligence in the education landscape, considering two different aspects of AI. The first approach considers education using AI technology during teaching and learning, known as AI in education, while the second one is education for AI concentrated on AI literacy. AI transforms the role of teachers, shifting their focus to the more strategic and creative aspects of the educational process. The techniques and technologies implemented in the educational AI-powered systems have also been communicated. AI not only imparts privileges to the education systems but also triggers some concerns. Ethical issues, including data privacy, biases, algorithmic robustness, and the potential misuse of AI-generated content, have been studied as well.
Preprint
Full-text available
Digital Twinning (DT) has become a main instrument for Industry 4.0 and the digital transformation of manufacturing and industrial processes. In this statement paper, we elaborate on the potential of Digital Twinning as a valuable tool in support of the management of intelligent infrastructures throughout all stages of their life-cycle. We highlight the associated needs, opportunities and challenges and discuss the needs from both the research and applied perspectives. We elucidate the transformative impact of digital twin applications for strategic decision making, discussing its potential for situation awareness, as well as enhancement of system resilience, with a particular focus on applications that necessitate efficient, and often real-time, or near real-time, diagnostic and prognostic processes. In doing so we elaborate on the separate classes of DT, ranging from simple images of a system, all the way to interactive replicas that are continually updated to reflect a monitored system at hand. We root our approach in the adoption of hybrid modelling as a seminal tool for facilitating twinning applications. Hybrid modelling refers to the synergistic use of data with models that carry engineering or empirical intuition on the system's behaviour. We postulate that modern infrastructures can be viewed as cyber-physical systems comprising, on the one hand, an array of heterogeneous data of diversified granularity and, on the other, a model (analytical, numerical, or other) that carries information on the system's behaviour. We thus propose Hybrid Digital Twins (HDT) as the main enabler of smart and resilient infrastructures.
Article
Full-text available
Expert problem-solving is driven by powerful languages for thinking about problems and their solutions. Acquiring expertise means learning these languages—systems of concepts, alongside the skills to use them. We present DreamCoder, a system that learns to solve problems by writing programs. It builds expertise by creating domain-specific programming languages for expressing domain concepts, together with neural networks to guide the search for programs within these languages. A ‘wake–sleep’ learning algorithm alternately extends the language with new symbolic abstractions and trains the neural network on imagined and replayed problems. DreamCoder solves both classic inductive programming tasks and creative tasks such as drawing pictures and building scenes. It rediscovers the basics of modern functional programming, vector algebra and classical physics, including Newton’s and Coulomb’s laws. Concepts are built compositionally from those learned earlier, yielding multilayered symbolic representations that are interpretable and transferrable to new tasks, while still growing scalably and flexibly with experience. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
In this paper, we bring together two closely related, but distinct, notions: argument and explanation. We clarify their relationship. We then provide an integrative review of relevant research on these notions, drawn both from the cognitive science and the artificial intelligence (AI) literatures. We then use this material to identify key directions for future research, indicating areas where bringing together cognitive science and AI perspectives would be mutually beneficial. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
From sparse descriptions of events, observers can make systematic and nuanced predictions of what emotions the people involved will experience. We propose a formal model of emotion prediction in the context of a public high-stakes social dilemma. This model uses inverse planning to infer a person’s beliefs and preferences, including social preferences for equity and for maintaining a good reputation. The model then combines these inferred mental contents with the event to compute ‘appraisals’: whether the situation conformed to the expectations and fulfilled the preferences. We learn functions mapping computed appraisals to emotion labels, allowing the model to match human observers’ quantitative predictions of 20 emotions, including joy, relief, guilt and envy. Model comparison indicates that inferred monetary preferences are not sufficient to explain observers’ emotion predictions; inferred social preferences are factored into predictions for nearly every emotion. Human observers and the model both use minimal individualizing information to adjust predictions of how different people will respond to the same event. Thus, our framework integrates inverse planning, event appraisals and emotion concepts in a single computational model to reverse-engineer people’s intuitive theory of emotions. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
Reasoning is the derivation of new knowledge from old. The reasoner must represent both the old and new knowledge. This representation will change as reasoning proceeds. This change will not just be the addition of the new knowledge. We claim that the representation of the old knowledge will also often change as a side effect of the reasoning process. For instance, the old knowledge may contain errors, be insufficiently detailed or require new concepts to be introduced. Representational change triggered by reasoning is a common feature of human reasoning but it has been neglected both in Cognitive Science and Artificial Intelligence. We aim to put that right. We exemplify this claim by analysing Imre Lakatos’s rational reconstruction of the evolution of mathematical methodology. We then describe the abduction, belief revision and conceptual change (ABC) theory repair system, which can automate such representational change. We further claim that the ABC system has a diverse range of applications to successfully repair faulty representations. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
What is required to allow an artificial agent to engage in rich, human-like interactions with people? I argue that this will require capturing the process by which humans continually create and renegotiate ‘bargains’ with each other. These hidden negotiations will concern topics including who should do what in a particular interaction, which actions are allowed and which are forbidden, and the momentary conventions governing communication, including language. Such bargains are far too numerous, and social interactions too rapid, for negotiation to be conducted explicitly. Moreover, the very process of communication presupposes innumerable momentary agreements concerning the meaning of communicative signals, thus raising the threat of circularity. Thus, the improvised ‘social contracts’ that govern our interactions must be implicit. I draw on the recent theory of virtual bargaining, according to which social partners mentally simulate a process of negotiation, to outline how these implicit agreements can be made, and note that this viewpoint raises substantial theoretical and computational challenges. Nonetheless, I suggest that these challenges must be met if we are ever to create AI systems that can work collaboratively alongside people, rather than serving primarily as valuable special-purpose computational tools. This article is part of a discussion meeting issue 'Cognitive artificial intelligence'.
Article
Full-text available
In this paper, we first explain why human-like dialogue understanding is so difficult for artificial intelligence (AI). We discuss various methods for testing the understanding capabilities of dialogue systems. Our review of the development of dialogue systems over five decades focuses on the transition from closed-domain to open-domain systems and their extension to multi-modal, multi-party and multi-lingual dialogues. From being somewhat of a niche topic of AI research for the first 40 years, it has made newspaper headlines in recent years and is now being discussed by political leaders at events such as the World Economic Forum in Davos. We ask whether large language models are super-parrots or a milestone towards human-like dialogue understanding and how they relate to what we know about language processing in the human brain. Using ChatGPT as an example, we present some limitations of this approach to dialogue systems. Finally, we present some lessons learned from our 40 years of research in this field about system architecture principles: symmetric multi-modality, no presentation without representation and anticipation feedback loops. We conclude with a discussion of grand challenges such as satisfying conversational maxims and the European Language Equality Act through massive digital multi-linguality—perhaps enabled by interactive machine learning with human trainers. This article is part of a discussion meeting issue 'Cognitive artificial intelligence'.
Article
Full-text available
A hallmark of human intelligence is the ability to understand and influence other minds. Humans engage in inferential social learning (ISL) by using commonsense psychology to learn from others and help others learn. Recent advances in artificial intelligence (AI) are raising new questions about the feasibility of human–machine interactions that support such powerful modes of social learning. Here, we envision what it means to develop socially intelligent machines that can learn, teach, and communicate in ways that are characteristic of ISL. Rather than machines that simply predict human behaviours or recapitulate superficial aspects of human sociality (e.g. smiling, imitating), we should aim to build machines that can learn from human inputs and generate outputs for humans by proactively considering human values, intentions and beliefs. While such machines can inspire next-generation AI systems that learn more effectively from humans (as learners) and even help humans acquire new knowledge (as teachers), achieving these goals will also require scientific studies of its counterpart: how humans reason about machine minds and behaviours. We close by discussing the need for closer collaborations between the AI/ML and cognitive science communities to advance a science of both natural and artificial intelligence. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
Large language models (LLMs) are one of the most impressive achievements of artificial intelligence in recent years. However, their relevance to the study of language more broadly remains unclear. This article considers the potential of LLMs to serve as models of language understanding in humans. While debate on this question typically centres around models’ performance on challenging language understanding tasks, this article argues that the answer depends on models’ underlying competence, and thus that the focus of the debate should be on empirical work which seeks to characterize the representations and processing algorithms that underlie model behaviour. From this perspective, the article offers counterarguments to two commonly cited reasons why LLMs cannot serve as plausible models of language in humans: their lack of symbolic structure and their lack of grounding. For each, a case is made that recent empirical trends undermine the common assumptions about LLMs, and thus that it is premature to draw conclusions about LLMs’ ability (or lack thereof) to offer insights on human language representation and understanding. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
General mathematical reasoning is computationally undecidable, but humans routinely solve new problems. Moreover, discoveries developed over centuries are taught to subsequent generations quickly. What structure enables this, and how might that inform automated mathematical reasoning? We posit that central to both puzzles is the structure of procedural abstractions underlying mathematics. We explore this idea in a case study on five sections of beginning algebra on the Khan Academy platform. To define a computational foundation, we introduce Peano, a theorem-proving environment where the set of valid actions at any point is finite. We use Peano to formalize introductory algebra problems and axioms, obtaining well-defined search problems. We observe existing reinforcement learning methods for symbolic reasoning to be insufficient to solve harder problems. Adding the ability to induce reusable abstractions (‘tactics’) from its own solutions allows an agent to make steady progress, solving all problems. Furthermore, these abstractions induce an order to the problems, seen at random during training. The recovered order has significant agreement with the expert-designed Khan Academy curriculum, and second-generation agents trained on the recovered curriculum learn significantly faster. These results illustrate the synergistic role of abstractions and curricula in the cultural transmission of mathematics. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.
Article
Full-text available
Statistical machine learning usually achieves high-accuracy models by employing tens of thousands of examples. By contrast, both children and adult humans typically learn new concepts from either one or a small number of instances. The high data efficiency of human learning is not easily explained in terms of standard formal frameworks for machine learning, including Gold’s learning-in-the-limit framework and Valiant’s probably approximately correct (PAC) model. This paper explores ways in which this apparent disparity between human and machine learning can be reconciled by considering algorithms involving a preference for specificity combined with program minimality. It is shown how this can be efficiently enacted using hierarchical search based on identification of certificates and push-down automata to support hypothesizing compactly expressed maximal efficiency algorithms. Early results of a new system called DeepLog indicate that such approaches can support efficient top-down construction of relatively complex logic programs from a single example. This article is part of a discussion meeting issue ‘Cognitive artificial intelligence’.