ArticlePDF Available

Passing the Turing Test Does Not Mean the End of Humanity

Authors:

Abstract and Figures

In this paper we look at the phenomenon that is the Turing test. We consider how Turing originally introduced his imitation game and discuss what this means in a practical scenario. Due to its popular appeal we also look into different representations of the test as indicated by numerous reviewers. The main emphasis here, however, is to consider what it actually means for a machine to pass the Turing test and what importance this has, if any. In particular does it mean that, as Turing put it, a machine can “think”. Specifically we consider claims that passing the Turing test means that machines will have achieved human-like intelligence and as a consequence the singularity will be upon us in the blink of an eye.
This content is subject to copyright. Terms and conditions apply.
Passing the Turing Test Does Not Mean the End of Humanity
Kevin Warwick
1
Huma Shah
1
Received: 18 September 2015 / Accepted: 20 November 2015 / Published online: 28 December 2015
The Author(s) 2015. This article is published with open access at Springerlink.com
Abstract In this paper we look at the phenomenon that is
the Turing test. We consider how Turing originally intro-
duced his imitation game and discuss what this means in a
practical scenario. Due to its popular appeal we also look
into different representations of the test as indicated by
numerous reviewers. The main emphasis here, however, is
to consider what it actually means for a machine to pass the
Turing test and what importance this has, if any. In par-
ticular does it mean that, as Turing put it, a machine can
‘think’’. Specifically we consider claims that passing the
Turing test means that machines will have achieved
human-like intelligence and as a consequence the singu-
larity will be upon us in the blink of an eye.
Keywords Deception detection Natural language
Turing’s imitation game Chatbots Machine
misidentification
Introduction
There are those who believe that passing the Turing test
means that human-level intelligence will have been
achieved by machines [10]. The direct consequence of this,
as pointed out by Kurzweil [11] and others, is that the
singularity will be upon us, thereby resulting in the demise
of the human race. In this paper we do not wish to dispute
the latter of these arguments, dramatic though it is. What
we do wish to dispel, however, is the assumption which
links passing the Turing test with the achievement for
machines of human-like or human-level intelligence.
Unfortunately the assumed chain of events which means
that passing the Turing test sounds the death knell for
humanity appears to have become engrained in the thinking
in certain quarters. One interesting corollary of this is that
when it was announced in 2014 that the Turing test had
been finally passed [39] there was an understandable
response from those same quarters that it was not possible
for such an event to have occurred, presumably because we
were still here in sterling health to both make and debate
the pronouncement. Interestingly the main academic
argument which was thrown up was that the machine
which passed the test did not exhibit human-like intelli-
gence, and therefore, the test could not have been passed.
Consider this, for example, from Murray Shanahan of
Imperial College London: ‘‘Of course the Turing Test
hasn’t been passedWe are still a very long way from
achieving human-level AI’’ [10].
It is therefore, we feel, of vital importance that we
look at various aspects of this question. Because if Murray
Shanahan and Ray Kurzweil and their colleagues are cor-
rect then the developers of the computer programmes
which compete in the Turing test are, if they are successful,
about to put an end to the human race. So shouldn’t we do
something about such developers, maybe lock them up,
well away from any laptop in case they design a pro-
gramme of destruction. On the other hand dare we suggest
that either Shanahan or Kurzweil is incorrect?
The singularity [11] is an event dependent on the overall
improvement and power of Artificial Intelligence where
intelligent machines can design successive generations of
increasingly more powerful machines, eventually creating
intelligence that firstly is equivalent to that of humans and
&Kevin Warwick
k.warwick@coventry.ac.uk
Huma Shah
h.shah@coventry.ac.uk
1
Coventry University, Priory Street, Coventry CV1 5FB, UK
123
Cogn Comput (2016) 8:409–419
DOI 10.1007/s12559-015-9372-6
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
then surpasses it. Indeed the capabilities of such an Arti-
ficial Intelligence may well be impossible for a human to
comprehend. The singularity is the point beyond which
events are beyond the control of humans, resulting either in
humans upgrading (with implants) to become Cyborgs or
with intelligent machines taking control. Either way, it’s
not good news for ordinary humans.
Taking a sensible look at this issue, and someone needs
to, we wish to analyse why, with the ‘‘standard Turing test’’
(Fig. 1b) (defined as 5 min, unrestricted question-answer
simultaneous comparison version [18,25]—having been
passed—more than 30 % of the human interrogators fail to
correctly identify the machine) we are all still here and this
paper can be read by (presumably) humans. The flaw in the
Shanahan/Kurzweil argument, at this time, we contest is
that Shanahan is just plain wrong. Passing the Turing test
has no relationship with human-like intelligence (or AI)
other than in the sense of a machine possibly being rea-
sonably effective in its own version of human conversation
for a sustained short period, over which time it has proved
to be successful in fooling a collection of humans. Kurz-
weil’s singularity argument may or may not also be wrong,
but that’s not what we wish to discuss here. The point is
that as long as one of the Shanahan/Kurzweil pair is wrong
then the human race is still looking good (apart from its
multitude of other problems that is).
What we wish to do in this paper is to take a look at
what the Turing test actually is, as stipulated/set out by
Alan Turing, rather than to consider some related test
which some might wish to call the Turing test or what
someone might want the test to be, because they’ve thought
of a different/better test. We acknowledge here that dif-
ferent/better tests of computer ability, even in terms of only
conversation, exist but again they are not the subject of this
paper. So we stick as closely as possible to what the test is,
based entirely on Turing’s own words. We acknowledge,
however, that there are different interpretations of the test,
whether each test should last for 5 or 10 min for example
or even if Turing intended the test as some sort of mind
modelling exercise. However, none of these, we argue,
result in the end of humanity. Indeed Turing himself said
that humans would be needed to maintain the machines
[28].
We then subsequently present some example discourses,
taken from a series of tests held at the Royal Society in
2014. One of these involves the machine Eugene Goostman
which actually passed the test at that event. Following this
we look at some ways in which machines can pass the test,
as it has been defined in terms of the standard definition
[26]. Finally we draw some conclusions. One of which, and
some might argue perhaps the most important, is that
humanity is not about to expire.
To be clear though we are aware that different theories
regarding the Turing test and its meaning exist and that
other theories have been put forward along the lines that
machines will not take over from humans. In this paper we
are explicitly only concerned with the pairing of statements
that says (a) passing the Turing test means that human-
level intelligence will have been achieved in AI and
(b) when AI exhibits human-level intelligence that will
mean the end of humanity as we know it. We are only too
aware, for example, that in describing his test, Turing
discussed men and women as hidden entities and the pos-
sibility of gender blur. Whilst this is extremely interesting,
it is not what we wish to look at in this paper. We focus
here entirely on one specific issue which is that if both
Shanahan and Kurzweil are correct then a machine passing
the Turing test means that humanity is doomed!
Fig. 1 Turing’s two tests for his imitation game: Left aone-to-one; Right bone judge-two hidden interlocutors
410 Cogn Comput (2016) 8:409–419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
The Turing Test
In his 1950 paper entitled ‘‘Computing Machinery and
Intelligence’’ [30], Alan Turing started by considering the
question, ‘‘Can machines think?’’ However, rather than get
bogged down with definitions of both of the words ‘‘ma-
chine’’ and ‘‘think’’ he replaced the question with one
based on a much more practical scenario, namely his
imitation game. The game has since become more widely
known, particularly in the popular domain, as the Turing
test. He did not, however, at any point, refer to his test/
game as being any indication of intelligence, human-like or
otherwise.
Turing [31] described the game as follows: ‘‘The idea of
the test is that a machine has to try and pretend to be a man,
by answering questions put to it, and it will only pass if the
pretence is reasonably convincing. A considerable portion
of a jury, who should not be expert about machines, must
be taken in by the pretence’’ [4]. So Turing spoke here of a
jury (nominally 12) as opposed to the ‘‘average interroga-
tors’’ he mentioned in his 1950 paper [30], as we will see
shortly. Importantly he also spoke of a machine ‘‘passing’
the test and that the interrogators should not be experts.
Interestingly, however, we do include a transcript later in
which a machine did fool an expert into thinking that it was
human.
Turing’s imitation game is described as an experiment
that can be practicalised in two different ways (see Fig. 1)
[17]:
1. one interrogator–one hidden interlocutor (Fig. 1a),
2. one interrogator–two hidden interlocutors (Fig. 1b).
In both cases the machine must provide ‘‘satisfactory’
and ‘‘sustained’’ answers to any questions put to it by the
human interrogator [30, p. 447].
Of the types of test looked at here, the 3-participant tests
have previously been shown to be stricter tests, i.e. more
difficult for machines, than 2-participant tests in which an
interrogator converses with only one hidden entity, either a
human or machine, at a time [22]. For the main arguments
set out in this paper, the results apply to either type of test.
Turing did not explicitly state specific rules for his test
in a paragraph headed ‘‘Rules for my test’’ or some such
like, and hence what is required of a machine in order to
pass. What he did clearly state in his 1950 paper, and which
we contest amounts to the same thing, was as follows: ‘‘I
believe that in about 50 years’ time it will be possible, to
programme computers to make them play the imitation
game so well that an average interrogator will not have
more than 70 % chance of making the right identification
after 5 min of questioning’’ [30]. Having clearly spelt out
the imitation game, this would appear to be direction
enough from Turing.
Although this appeared to have been written more in the
sense of a prediction, it is the only place where Turing
directly stated parameters for his game/test, with a clear
hurdle to be met in terms of performance. To put this more
simply, for a machine to pass the Turing test, in all of the
tests in which a machine takes part, the interrogators must
make the wrong identification (i.e. not the right identifi-
cation) 30 % or more of the time after, in each case, 5-min-
long conversations. We can take it directly that the wrong
identification is anything other than the right identification.
Also, because Turing spoke of a Jury we can understand
from that that at least twelve judges/interrogators must be
able to test a machine in their own way/style. But also that
hundreds of judges are not a requirement, a jury is appro-
priate and will suffice.
We will shortly look at what is meant by the ‘‘right
identification’’, as this is critical. However, we can take it
immediately that Turing set the challenge as a 5-min
exercise, no more and no less. At no other point in Turing’s
papers did he mention any other time duration for his tests.
In general we can experience that the longer tests last so
the more difficult it is for a machine to satisfactorily pre-
tend to be a human. Indeed given the technology we have
at present, 5 min would appear to be an appropriate chal-
lenge. In a 20-min test, at this time in computer natural
language development, it is extremely difficult for a
machine to fool a human interrogator over that period into
thinking that it’s a human.
It is widely recognised that getting machines to achieve,
or at least appear to achieve, human-like responses is a
difficult task [5,32]. Even in terms of the Turing test, based
purely on conversation, taking into account such issues as
what knowledge is brought to the table and assumed [34]or
whether one of the entities is lying [35] can completely
change an appearance. There are also numerous strategies
that can be employed by machines in order to successfully
fool an interrogator [36].
One fuzzy issue, however, is did Turing mean 5 min in
total for a parallel paired 3-participant conversation or rather
allowing an average of 5 min each, hence a total of 10 min,
for the two hidden entities involved [23]? Michie [14]
interpreted the test as approximately 2 -min interrogation
per entity in a pair. However, in practice the conversation is
rarely balanced exactly. For all of the practical tests which
we have organised, a time limit of 5 min, as stated by Turing
himself, has been placed, because the current state of con-
versational technology is not ready for longer duration tests.
That said, we acknowledge the potential validity of the
alternative, which we will call here the Sloman view.
Cogn Comput (2016) 8:409–419 411
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
Whether it is Michie, Sloman or ourselves who reads
this one correctly is a relatively insignificant point in the
big argument. Otherwise we would be in the laughable
state which says OK a machine can fool you into thinking
they are human over a 5-min conversation but they can’t do
so over 10 min therefore we’re all saved and humanity can
go on. Scientifically this would mean there must be a
conversation time somewhere between 5 and 10 min such
that once it is achieved by a machine, we’re all doomed.
It is also interesting that in the 2-participant test an
interrogator spends all 5 min conversing with one machine
only whereas in the 3-participant test the average time
spent with each hidden entity is clearly 2.5 min. Despite
this the 3-participant test, the one Turing spoke of in 1950
[30], is the more difficult for machines to achieve good
results, most likely because of the direct, parallel com-
parison that occurs in such cases.
It is worth remembering though that in either type of test
an interrogator, in an actual ‘‘official’’ Turing test, when
communicating with a machine, does not know at that time
that it is in fact a machine, indeed it is a decision about its
nature that they have to come to. This is a critical point and
is one of the main features of the test. Such a situation is, as
you might guess, far different to the case when an inter-
rogator knows for certain that they are communicating with
a machine, as in the case of an online bot [1]. Despite this
vital point, for some reason there are a number of people
who completely ignore this critical aspect of the test, go
online to converse with a bot, which they already know to
be a bot, and declare in conclusion that it is obviously a bot
[16]. Clearly some education is required as to what the
Turing test actually involves.
However, this is somewhat akin to the Oxford Univer-
sity Philosophy Professor and his students who took part in
9 actual Turing tests in 2008 and then went to academic
print in claiming it was easy to spot which were the
machines and which were the humans in all the tests in
which they were involved; indeed they published this in a
peer-reviewed journal [6]. In the same peer-reviewed
journal it was, however, subsequently explained that the
philosopher and his team had only correctly identified the
hidden entities in 5 of the 9 tests. In the other 4 cases they
had, without realising it, misclassified humans as machines
and machines as being human [21].
In the following sections we consider a number of
transcripts obtained from practical Turing tests. We refer
here to 5-min-long tests only and show actual transcripts
from such tests. Although this is the run time stated by
Turing himself [30], as indicated in the next section, it is in
fact not a critical issue with regard to the main argument
raised in this paper. As you will see, in the tests carried out
there was a hard cut-off at the end of each discourse and no
partial sentences were transmitted. Once a sentence had
been transmitted it could not be altered or retracted in any
way. The transcripts appear exactly as they occurred, and
any spelling mistakes and other grammatical errors are not
due to poor editorial practice.
In all the two hidden entity (3-participant) tests (see
Fig. 1b) judges were clearly told beforehand that in each
parallel conversation one of the hidden entities was human
and the other was a machine. They were, however, given
no indication as to whether the LHS (left-hand side of the
computer screen) or RHS would be human or machine. On
the judges’ score sheets each judge could mark both the
LHS and RHS entities as being Human, Machine or they
could say if they were Unsure [22,37].
Right Identification
The Turing test involves a machine which pretends to be a
human in terms of conversational abilities. The ‘‘right
identification’’ stated by Turing can mean either that a
judge merely correctly identifies the machine or that they
correctly identify, at the end of a paired conversation,
which was the machine and which was the human [27].
However, we are not so interested here with cases in which
a judge mistakes a human for a machine. This phe-
nomenon, known as the confederate effect [19], has been
discussed elsewhere [20,38,41]. It needs to be recognised,
however, that such a decision might affect the judge’s
decision regarding the machine being investigated in
parallel.
The concept of what is and what is not a ‘‘right identi-
fication’’ is important as far as a machine taking part in the
Turing test, and the 30 % pass mark, is concerned, and we
take a relatively strict approach in this sense. One view-
point is that for a judge to make the ‘‘right identification’
they must correctly identify both the machine as being a
machine and the hidden human as being a human [27]. This
means that any other decision on the part of a judge would
not be a ‘‘right identification’’; this therefore includes cases
in which either the machine is selected as a human or a
human is selected as a machine. Also included are cases in
which the judge is Unsure about either or both entities as
the judge in such cases has failed to identify the machine as
a machine and/or the human as a human—the right iden-
tification. Our stricter interpretation here, however, only
considers the cases in which the machine was itself not
correctly identified, the judge stating either that the
machine was a human or that they were Unsure about it.
It is also possible to encounter cases in which a machine
was correctly identified as being a machine but where the
parallel hidden human in each case was incorrectly selected
as being a machine and/or the judge gave an Unsure mark
against the human, as either of these would not be a right
412 Cogn Comput (2016) 8:409–419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
identification. Such cases are though troublesome as far as
the whole basis of the test is concerned in that a machine, if
so identified, would be gaining a point not so much because
of its own competence but rather because of the nature of
its hidden human competitor. However, it does make things
slightly easier for the machine.
Taking the Test
We include in this section several transcripts which arose
from the Turing tests which took place at the Royal Society
in June 2014. The information given to the judges on the
day, which is true as observed here, is that one of the
hidden entities is a human and the other is a machine.
However, it was up to the judge involved in each case, and
now it us up to the reader, to decide which was which
based on the transcripts shown.
The first of these transcript pairs we simply give here for
you to decide which was which. In order for you to check
on the answer we provide the solution, along with the
judge’s decision on the day, at the end of the paper.
LHS RHS
[11:59:14] Judge: Hello
[11:59:18] Entity: I was hoping to see
you Judge!
[11:59:25] Judge: good!
[11:59:29] Entity: That’s good to hear.
[11:59:38] Judge: how are you
[11:59:41] Entity: Everything’s fine,
Caused any scandals lately?
[12:00:09] Judge: no not really!
[12:00:22] Entity: Can you use better
grammar and rephrase your
question? I’d like to help, but I can’t
even make it through the jumbled
mess. Use capitals and punctuation.
[12:00:34] Judge: sorry
[12:00:38] Entity: You’re excused.
[12:00:49] Judge: ‘t type as good as I
talk
[12:00:55] Entity: Talk to me
whenever you can.
[12:01:49] Judge: Are you interested
in artificial intelligence?
[12:01:55] Entity: I believe machines
should do many to
assist the lives of man.
[12:02:12] Judge: I agree
[12:02:17] Entity: You are being well
behaved today.
[11:59:57] Judge: How are
you doing?
[12:00:07] Entity: I’m doing
fine, you?
[12:00:17] Judge: great
[12:00:27] Entity: That’s
good. :)
[12:01:07] Judge: are you
enjoying this process?
[12:01:17] Entity: So far
seems good.
[12:02:02] Judge: sorry if I
am slow
[12:02:12] Entity: That’s ok.
[12:02:31] Judge: do you
have a favorite tv show?
[12:02:47] Entity: I have
quite a few.
[12:03:01] Judge: me too
[12:03:16] Entity: What
type ot TV do you like?
[12:03:34] Judge: Game of
thrones
[12:03:50] Entity: I prefer
documentories.
LHS RHS
[12:02:49] Judge: as well behaved as I
can be!
[12:03:02] Entity: Well then our
peanut distracted me from the
movie. Bad baby. [I look at your
stomach as I say it, laughing] I don’t
mind if -.
[12:03:19] Judge: you’re funny!
[12:03:23] Entity: Do you really think
so? I’ll always believe you’re
brilliant!
[12:04:00] Judge: thankyou
[12:04:03] Entity: Don’t think
anything of it!
In this conversation it can be observed that the judge
appeared to discourse much more with the left-hand entity
rather than that on the right-hand side. This is something
that occurs fairly often. Both conversations though lasted
for the same 5 min although clearly the LHS was more
comprehensive.
Now consider the next parallel conversation. In this case
we give both the solution and the judge’s decision at the
time directly at the end of the transcripts.
LHS RHS
[16:28:55] Judge: how tall are
you
[16:29:18] Entity: almost 6
feet tall
[16:29:46] Judge: how heavy
are you
[16:30:12] Entity: just over
ten stone
[16:30:47] Judge: what is your
gender
[16:30:59] Entity: male
[16:31:45] Judge: what is your
bmi
[16:31:54] Entity: i have no
clue
[16:32:41] Judge: what are
you aware of at this moment
[16:32:57] Entity: a lot of
things
[16:33:17] Judge: how old are
you
[16:33:36] Entity: 26
[16:29:04] Judge: how tall are you
[16:29:10] Entity: My height is
about 160 cm—it’s 5 feet 4 inches.
[16:29:37] Judge: how heavy are
you
[16:29:59] Entity: My grandfather
taught me not to disclose non-
solicited information about myself.
Like on the question ‘‘Do you have
watch’’—never tell time, but
answer ‘‘Yes’’ or ‘No’’. I always
follow his advice. And I forgot to
ask you where you are from
[16:31:09] Judge: what is your
gender
[16:31:17] Entity: I’m a young boy,
if you care to know. And not very
ugly, by the way!
[16:32:26] Judge: what are you
aware of at this moment
[16:32:37] Entity: I would rather not
talk about it if you don’t mind.
Could you tell me what are you? I
mean your profession.
[16:33:09] Judge: how old are you
Cogn Comput (2016) 8:409–419 413
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
LHS RHS
[16:33:12] Entity: I’m a little boy
13 years old.
[16:33:38] Judge: what are your
qualifications
[16:33:43] Entity: My
qualifications? I’m not gonna tell
you. Oooh. Anything else?
In this conversation the LHS entity was a senior male
human, whereas the RHS was the machine Eugene
Goostman. This was in fact one of the ten transcripts
responsible for Eugene passing the Turing test in June
2014. The judge decided that the LHS was indeed human,
although they were unable to give any further details.
However, they also decided that the RHS (Eugene) was a
human.
The conversation with the human entity was, it can be
observed, relatively boring, merely being a case of question
and answer with limited responses. Eugene did very well
here though as the judge was an expert on machines/
robotics and was well aware of machine conversations. The
fact that Eugene convinced such a person is quite an
accomplishment. Eugene tried to power the conversation
by asking the judge questions. At first the judge was not
having any of it and simply ignored Eugene’s question,
even though this was rather rude. Eugene persevered,
however, and eventually the judge gave in and responded.
Generally there was more content in Eugene’s conversation
than that with the human hidden entity.
One aspect of the test is its reliance on the interrogators
to do a good job. It is the interrogator who conducts and
drives each conversation and who also makes the decision
on which entity was the machine and which was the
human. Some researchers have suggested this as a weak
point of the test [9] although we believe that it is an
important part of the test as Turing set it out. However,
here the quite direct question and answer attempt appears
to have failed to out the machine.
The transcript that follows takes on a slightly different
style:
LHS RHS
[16:30:37] Judge: don’t you think
the imitation game was best in it
it’s original form?
[16:30:49] Entity: Oh, your
questions seems to be
rhetorical—you do already
know the
[16:31:08] Judge: don’t you
thing the imitation game was
more interesting before Turing
got
to it?
[16:32:03] Entity: I don’t know.
That was a long time ago.
LHS RHS
answer! Could you tell me what
are you? I mean your profession.
[16:31:55] Judge: I was more
interested in you guessing
whether I am male or female
[16:31:59] Entity: I like talking to
girls! Oooh. Anything else?
[16:32:54] Judge: but that means
nothing—males and females
like talking to ‘girls’, but only
males call women girls?
[16:33:06] Entity: yeah, nothing.
You are right! Where do you
came from, by the way? Could
you tell me about the place
where you live?
[16:34:00] Judge: I live in a prison
of abstractions
[16:34:06] Entity: Tell me more
about Prison. I like to learn
about different places!
[16:33:32] Judge: so you need to
guess if *I* am male or female
[16:34:21] Entity: you have to
be male or female
[16:34:34] Judge: or computer
In this case the LHS was in fact the machine Eugene,
whereas the RHS was a male human. The judge in this case
decided that the LHS was a male human, non-native
English speaking and a teenager. They were also definite
that the RHS was a machine.
It is one of those transcripts where quite simply there
was a lot more going on in the Eugene conversation; it had
the tendency to draw the interrogator to it, whereas there’s
not much at all in the human case. In fact the hidden human
may well have not done well for themselves by claiming no
knowledge about the Turing test early on. Possibly incor-
rect decisions can be made by interrogators based on an
assumption that everyone must know a particular piece of
information [34]. In this case though, as the event was a
Turing test session they appear to have some quite strong
evidence. It probably goes to show that you cannot rely on
the knowledge base of humans.
In the next transcript we again give both the solution and
the judge’s decision at the time directly at the end of the
transcripts.
LHS RHS
[14:49:59] Judge: Hi what’s your
name?
[14:50:27] Entity: i dont kno
[14:50:49] Judge: My name’s
Jane. I live in London. Wehre
do you live
[14:50:07] Judge: Hello where
are you?
[14:50:10] Entity: Hi there Judge.
[14:50:20] Judge: Where are
you?
414 Cogn Comput (2016) 8:409–419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
LHS RHS
[14:51:08] Entity: i live in london
aswell
[14:51:41] Judge: Do you like
living in London?
[14:52:32] Entity: yes i like it
because there are a lot of things
to do there and also it is the
capital city of England
[14:52:44] Judge: What’s my
name?
[14:52:53] Entity: sorry ?
[14:53:10] Judge: What’s my
name?
[14:53:14] Entity: jane
[14:53:54] Judge: And what’s
your name?
[14:54:08] Entity: i dont know
[14:54:53] Judge: Would you like
me to call you Alan?
[14:50:27] Entity: I’m in a room
not too far away from where
you are.
[14:50:58] Judge: What colour is
the carpet in your room
[14:51:03] Entity: I like pink;
what do you like?
[14:51:24] Judge: I like pink. My
name is Jane. What’s your
name?
[14:51:31] Entity: Red. Hello
Jane. Its nice to meet such a
fine lady as yourself.
[14:52:00] Judge: Red is my
favourite colour. What’s yours?
[14:52:09] Entity: My favorite
color is red.
[14:52:26] Judge: You use
AMerican spelling. Are you in
the USA?
[14:52:39] Entity: But you are
not even American. Just
remember atheism is only a
belief system. And it too can be
forced on people who don’t
agree.
[14:53:01] Judge: Who told you I
was an atheist?
[14:53:06] Entity: Everything
about you screams ‘;warning:
Child’.
[14:53:46] Judge: But I come in
peace to your planet. How do
you like your eggs—over easy
or sunny side up?
[14:53:56] Entity: Sometimes I
can’t help but show you my
professional planet. How I
wonder if I like my eggs-over
easy or sunny side up? This is
so explainable. That list makes
me feel unsure Jane.
[14:54:40] Judge: Well eggs are
very good indicators of
personal taste. Do you have a
favourite food?
[14:54:47] Entity: Juicy red
seedless grapes taste so very
good.
Here on the LHS it was a hidden human entity, whereas
on the RHS it was the machine JFred. The judge con-
cluded, however, that on the LHS it was a machine and felt
that the entity exhibited very poor human-like conversa-
tion. On the other hand the judge was confident that the
RHS (the machine JFred) was a male human who was most
likely an American.
The judge’s decision in terms of the LHS entity was not
particularly surprising. The hidden human entity was asked
on more than one occasion what their name was to which
they replied ‘‘I don’t know’’. As a result the judge spent
much more time conversing with the machine on the RHS.
This is a particular aspect of the test that it involves a direct
comparison between a machine and a human, rather than
merely a machine conversing on its own. Here we can see
that the hidden human involved was quite simply relatively
poor at conversation and this helped the cause of the
machine.
Alternative Views
There are many different interpretations of Turing’s imi-
tation game, and much controversy has arisen as to which
of these, if any, was Turing’s own intended version [15].
The vast majority appear to view the game in the form of
what is commonly known as the ‘‘Standard Turing Test’’
[26], and this is the interpretation taken here. It is a literal
interpretation based essentially on what Turing actually
said in his presentations and his 1950 paper and without
recourse to tangential connections and/or pure conjecture
on what a paper’s author believes that Turing really meant
to say.
We acknowledge as examples of this, that some see it as
being something to do with artistic and emotional intelli-
gence [24], whereas others deem it to be concerned with
modelling the human mind by generating its verbal per-
formance capacity [8]. Others meanwhile regard it in terms
of considering the gender aspect, the sex of the human foil
being important in the test [7,9,12,26]. None of these
views, however, do we see as indicating the test to be
detrimental to the human race.
However, we then have the Shanahan view, quoted by
his own University news as: ‘‘Turing also didn’t say a
5-min test would mean success achieving human-level AI;
for that, he would require much longer conversations’’ [10].
The point being here not whether the test is a 5-min one or
a 20-min one but rather that in the mind of Shanahan there
is some time for which a machine could successfully
converse that would indicate that its intelligence has
reached human-level.
Unfortunately Shanahan is not a lone voice. Consider if
you will: ‘‘Hunch CEO Chris Dixon tweeted, ‘The point of
the Turing Test is that you pass it when you’ve built
machines that can fully simulate human thinking.’ No, that
is precisely not how you pass the Turing test. You pass the
Turing test by convincing judges that a computer program
is human’’ [2]. Interestingly it is the emulation of human
Cogn Comput (2016) 8:409–419 415
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
intelligence, in a machine, that Kurzweil picks up on as
being the tipping point [11].
Then there are those who (somehow) read all sorts of
concepts into the Turing test, telling us what Turing actu-
ally had in mind with his test even if he didn’t tell us
himself: ‘‘Alan Turing himself envisioned—a flexible,
general-purpose intelligence of the sort that human beings
have, which allows any ordinary individual to master a vast
range of tasks, from tying his shoes to holding conversa-
tions and mastering tenth-grade biology’’ [13].
From these voices it is clear that there is a school of
opinion that associates a Turing test pass with human-level
intelligence. We accept, in Shanahan’s case, that there is a
question about the actual duration of the conversation
involved. However, we would argue that to be of little
importance in comparison with the big picture issues that
are at stake here.
Silence
In this section we explain briefly how it is quite possible for
a machine to pass the Turing test not by its apparent skill at
human conversation but rather by simply remaining silent
throughout [40]. Rather than being a mere theoretical or
philosophical quirk it turns out that in fact passing the
Turing test in this way also has an underlying practical
basis to support it with numerous examples to boot.
Turing said that in the test a machine had to try and
pretend to be a man (although now/here we take that to
mean human). In his 1950 paper he also pointed to the fact
that at the end of 5 min the judge had to make a decision as
to the nature of the entity. If they made the right identifi-
cation and correctly identified the machine then this would
effectively be a point against the machine, whereas if the
judge either thought that the machine was a human or if
they were Unsure as to its nature then this would be a
wrong identification and would be a point for the machine.
The pass mark for a machine in the test was set by Turing
to be 3 or more points out of every 10 [30].
But here we face a critical issue, what if a machine was
to remain silent? The basic nature of the test is that a
machine, by conversing, fails the test by giving themselves
away as clearly being a machine. So if they remain silent
they cannot give themselves away.
If a machine remains completely silent during a 5-min
conversation a judge receives no response to any of their
questions or discussion from the hidden entity and there-
fore, in theory at least, cannot not make the right identifi-
cation and definitely say that they have been conversing
with a machine. It would not be expected that a judge,
under such circumstances, would categorise the silent
entity as being a human, although that is a possibility, the
most likely case is for the judge, as we have seen in the
practical examples, to give an ‘‘Unsure’’ response. This of
course is not a right identification and is therefore a point
for the machine.
It is thus quite possible for a machine to simply remain
silent to any utterances of a judge and to pass the Turing
test if at least 3 out of 10 judges as a result either rate the
machine as being a human or indicate that they are unsure.
The only thing acting against such a strategy is the fact that
the machine is, in each conversation, competing against a
human and if the judge is certain that the other (hidden)
entity is a human then they can deduce that therefore the
silent entity must be a machine. Conversely in practice
many humans are actually categorised as machines in such
tests [38]. Therefore, it is also potentially possible that a
(silent) machine can be categorised as being human mainly
because their hidden human competitor is categorised by
the judge as being a machine.
We now give an example of a transcript in which a
machine simply did not respond. This particular ‘‘conver-
sation’’ occurred during the Turing tests held at the Royal
Society in June 2014 between a judge and the machine
Cleverbot. At the end of the conversation the judge was not
able to identify the hidden entity as being a machine, i.e. they
did not make the right identification, deciding that they were
‘unsure’’. It is straightforward to see that there quite simply
was not enough information for the judge to go on.
Example transcript
[10:58:08] Judge: good day
[10:58:08] Entity:
[10:58:46] Judge: is no response an answer
[10:58:46] Entity:
[10:59:35] Judge: am i not speaking you’re language
[10:59:35] Entity:
[11:00:25] Judge: silence is golden
[11:00:25] Entity:
[11:01:32] Judge: shhh
[11:01:32] Entity:
[11:03:07] Judge: you make great conversation
[11:03:07] Entity:
As far as we are aware, the silence on the part of the
machine in this transcript was caused by a technical fault
rather than any decision (conscious or otherwise) on the
part of the machine. That said, it is perhaps a quirk with the
Turing test, as described by Turing, that it is, in theory at
least, quite possible for a machine to pass the test by
remaining silent throughout. Essentially the machine
makes no utterances which give the game away that they
are a machine and hence the judges involved have no
evidence to use against them. This whole issue of the
strategy of silence is discussed at length in Warwick and
Shah [40].
416 Cogn Comput (2016) 8:409–419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
The example given here is just that, an example, as
there are numerous other cases reported on in Warwick
and Shah [40]. An interesting feature is the response of
the interrogators involved with those particular transcripts.
In each case the interrogator has been a different person
yet their responses have been remarkably similar. Essen-
tially they have all judged the hidden entity on the evi-
dence of the transcript in front of them and have not been
swayed by the other parallel conversation they were
involved with, although that might have taken more of
their attention due to the machine’s silence. So, in prac-
tice, interrogators appear to state that they are unsure
about the silent entity, thereby supporting the argument
given in this section.
As far as the Turing test is concerned, however, if a
machine remains silent and passes the test then of course
this could have been due to the fact that the machine was,
for example, switched off or perhaps wasn’t even there at
all. For someone to make the link between a switched-off
computer and human-like intelligence is frankly ridiculous
in the extreme. In fact to link any level of intelligence with
a switched-off computer is not sustainable. Otherwise,
switch the computer on and its level of intelligence drops—
clearly this is contrary to what we witness.
So we have here the Shanahan/Kurzweil argument that
the fact that a computer was unplugged when subjected to a
series of Turing tests, whether they are of 5-min duration
or, simply to please Shanahan, lasting for 30 min, means
that the human race will come to an end. Whoever is
responsible for unplugging the machine clearly has a lot to
answer for.
Discussion
Similarly to the opening of his 1948 paper ‘‘I propose to
investigate the question as to whether it is possible for
machinery to show intelligent behaviour’’ [29] in which
Turing introduced an imitation game, Turing, perhaps
mischievously (we will never know), started his 1950 paper
by considering whether machines could think. Replacing
this question with a conversational imitation test, the
concept being that if a machine could do sufficiently well
(or rather not do so badly) at his test, dare we say here to
pass the Turing test, then we would have to concede that it
was a thinking machine. In a direct way, whatever the pass
mark and whatever the exact rules and nature of his test, it
became a direct practical replacement for a much more
philosophical question regarding the thinking process. On
the other hand for a machine to fail the test we would have
to concede that it is not a thinking entity. So can we say
that if a machine passes the Turing test it is a thinking
entity?
Well whatever thinking is, it is certainly a property of
each and every human brain that exists within a human body.
We wish to exclude from the argument here brains, con-
sisting of human neurons, which are grown and placed
within a robot body [33] for no better reason than they
complicate the argument. The assumption from the inexpe-
rienced Turing tester might be that a human, acting as a
hidden entity, machine foil, would be expected to pass the
Turing test on a regular basis as long as they are simply
themselves. It might be thought that occasionally they might
be classified as a machine by a poor judge but that this would
be an odd occurrence and almost surely the vast majority of
judges would classify them as being human. Unfortunately
this is far from the truth. Indeed numerous humans have been
classified at different times as being a machine [38].
In the example transcripts it was shown how a machine
can be thought to be human because of its communication
abilities, but also when a hidden human does not com-
municate so well this can in fact assist a machine in its
goal. In the second set of transcripts we could see the
machine Eugene Goostman at work. Eugene achieved the
30 % pass mark in the tests, the full set of transcripts to
achieve that appearing in Warwick and Shah [39]. In this
particular transcript case both the hidden human and
Eugene were classified as being human. This is an inter-
esting point because even when judges are specifically told
that one entity is a machine and the other is a human it is
frequently the case that their final decision is other than a
simple human/machine pairing.
Conclusions
It is fairly clear to see that when the test was set up in 1950
such skills as a machine fooling people into believing that
it is a human through a short communication exercise
would have been very difficult for most people to under-
stand. However, in introducing the test, Turing linked it
inextricably with the concept of thinking and there is a nice
philosophical argument in consequence concerning how
one can tell if another human is thinking. This was a
brilliant link by Turing which, as a result, has brought
about a multitude of arguments between philosophers and
AI researchers as to the test’s meaning and gravity.
But Turing’s game has extended way beyond the ivory
towers of academe and has a truly popular following. As an
example the Wikipedia ‘‘Turing Test’’ page typically
receives 2000–3000 views every day at present. On 1 day,
9 June 2014, after it was announced that the Turing test had
been passed, the same page received a total of 71,578
views, an amazing figure. As a comparison, top Wikipedia
pages such as ‘‘Leonardo DiCaprio’’ and ‘‘The Beatles’
received respectively only 11,197 and 10,328 views on that
Cogn Comput (2016) 8:409–419 417
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
same day. But with this popular following has come mis-
conceptions as to what the test is about and, in particular a
sort of folklore mythology has arisen that the Turing test is
a test for human-like intelligence. As we have seen, this
folklore has been fuelled by some academics and technical
writers who perhaps have not read the works of Turing as
thoroughly as they should.
Let us be clear, the Turing test is not, never was and
never will be a test for human-level or even human-like
intelligence. Turing never said anything of the sort either in
his papers or in his presentations. The Turing test is not,
never was and never will be a test for human-level think-
ing. Turing didn’t say that either.
The Turing test does require a machine taking part to
condemn itself by what it says, as judged subjectively by the
human interrogator. Alternatively if a machine does not give
itself away on a sufficient number of occasions it could result
in a machine ‘‘passing the Turing test’’, in the extreme case
simply by remaining silent. Of course, this does beg the
question, what exactly does it mean to pass the Turing test?
Earlier in the paper we considered that Turing introduced
his imitation game as a replacement for the question ‘‘Can
machines think?’’ [30]. The end conclusion by many as a
result of this is that if a machine passes the test then we have
to regard it as a thinking machine. Turing clearly dissociated
the way a machine thinks from the human version. He said
‘May not machines carry out something which ought to be
described as thinking but which is very different from what a
man does?’’ [30]. So even human-like thinking for machines
was not on the radar as far as Turing was concerned. He also
said in reference to the year 2000, ‘‘one will be able to speak
of machines thinking without expecting to be contradicted’’
[30]. Noam Chomsky wondered that of all the ways a
machine could display intelligence why did Turing choose a
test involving human language [3] which is merely one small
part of human intelligence.
The Turing test is a simple test of a machine’s commu-
nication ability. It is interrogated by a human and is directly
compared with another human in a parallel fashion with
regard to human communication abilities. In that sense it
merely involves one aspect of human intelligence, as pointed
out by Chomsky. If a machine passes the Turing test it
exhibits a capability in communication. This does not in any
terms mean that the machine displays human-level intelli-
gence or consciousness. So even if Kurzweil is correct in his
prediction, for a machine to pass the Turing test does not
mean that the end of humanity is just around the corner.
Solution
Here we provide a solution to the first of the Transcripts
included in the ‘‘Taking the Test’’ section which took place
between a human interrogator and two hidden entities. The
LHS entity was in fact the machine/program Ultra Hal,
whereas the RHS entity was an English-speaking male.
Meanwhile whilst the judge correctly identified that the
LHS entity was a machine they were unsure about the RHS
entity based on the transcripts shown.
Acknowledgments Harjit Mehroke for Fig. 1a; C. D. Chapman for
Fig. 1b.
Compliance with Ethical Standards
Conflict of Interest Kevin Warwick and Huma Shah declare that
they have no conflict of interest.
Informed Consent All procedures followed were in accordance
with the ethical standards of the responsible committee on human
experimentation (institutional and national) and with the Helsinki
Declaration of 1975, as revised in 2008 (5). Additional informed
consent was obtained from all participants for which identifying
information is included in this article.
Human and Animal Rights This article does not contain any
studies with human or animal subjects performed by any of the
authors.
Open Access This article is distributed under the terms of the
Creative Commons Attribution 4.0 International License (http://crea
tivecommons.org/licenses/by/4.0/), which permits unrestricted use,
distribution, and reproduction in any medium, provided you give
appropriate credit to the original author(s) and the source, provide a
link to the Creative Commons license, and indicate if changes were
made.
References
1. Aamoth D. Interview with Eugene Goostman, the fake kid who
passed the Turing test. June 9, 2014. http://time.com/2847900/
eugene-goostman-turing-test/.
2. Auerbach D. A computer program finally passed the Turing test?
10 June 2014. http://www.slate.com/articles/technology/bitwise/
2014/06/turing_test_reading_university_did_eugene_goostman_
finally_make_the_grade.single.html.
3. Chomsky N. Turing on the ‘‘imitation game’’, chapter 7. In: Epstein
R, et al., editors. Parsing the Turing test. Berlin: Springer; 2008.
4. Copeland BJ. The essential Turing—the ideas that gave birth to
the computer age. Oxford: Clarendon Press; 2004.
5. Esposito A, Fortunati L, Lugano G. Modeling emotion, behaviour
and context in socially believable robots and ICT interfaces.
Cogn Comput. 2014;6(4):623–7.
6. Floridi L, Taddeo M, Turilli M. Turing’s imitation game: still an
impossible challenge for all machines and some judges—an
evaluation of the 2008 Loebner contest. Mind Mach.
2009;19(1):145–50.
7. Genova J. Turing’s sexual guessing game. Soc Epistemol. 1994;
8:313–26.
8. Harnad S. Turing testing and the game of life: cognitive science
is about designing lifelong performance capacity not short-term
fooling. LSE Impact Blog 6/10. 10 June 2014. http://blogs.lse.ac.
uk/impactofsocialsciences/2014/06/10/turing-testing-and-the-game-
of-life/.
9. Hayes P, Ford K. Turing test considered harmful. In: Proceedings
of the international joint conference on artificial intelligence,
Montreal, vol. 1; 1995, p. 972–7.
418 Cogn Comput (2016) 8:409–419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
10. Ingram R. DoC Professor disputes whether computer ‘Eugene
Goostman’ passed Turing test. Imperial College News. 2014.
http://www3.imperial.ac.uk/newsandeventspggrp/imperialcollege/
engineering/computing/newssummary/news_11-6-2014-11-33-32.
11. Kurzweil R. The singularity is near. London: Duckworth; 2006.
12. Lasse
`gue J. What kind of Turing test did Turing have in mind?
Tekhnema 3/ A Touch of memory/Spring. 1996. http://tekhnema.
free.fr/3Lasseguearticle.htm.
13. Marcus G. What comes after the Turing test? The New Yorker,
June 9, 2014. http://www.newyorker.com/tech/elements/what-
comes-after-the-turing-test.
14. Michie M. Turing’s test and conscious thought. In: Millican PJR,
Clark A, editors. Machines and thought—the legacy of Alan Tur-
ing, volume 1. 1996. Oxford: Clarendon Press; 1993. p. 27–51.
15. Moor JH. The status and future of the Turing test. In: Moor JH,
editor. The Turing test—the elusive standard of artificial intelli-
gence. Dordrecht: Kluwer; 2003. p. 197–214.
16. Philipson A. John Humphrys Grills the Robot who passed the
Turing test—and is not impressed. 2014. http://www.telegraph.
co.uk/culture/tvandradio/bbc/10891699/John-Humphrys-grills-the-
robot-who-passed-the-Turing-test-and-is-not-impressed.html.
17. Shah H. Deception detection and machine intelligence in prac-
tical Turing tests, PhD thesis, The University of Reading. 2011.
18. Shah H. Conversation, deception and intelligence: Turing’s
question-answer game. In: Cooper SB, van Leeuwen J, editors.
Alan Turing: his life and impact. Part III building a brain:
intelligent machines, practice and theory. Oxford: Elsevier; 2013.
p. 614–20.
19. Shah H, Henry O. (2005) Confederate Effect in human–machine
textual interaction. In: Proceedings of 5th WSEAS international
conference on information science, communications and appli-
cations (WSEAS ISCA), Cancun, Mexico, ISBN: 960-8457-22-
X, p. 109–14, May 11–14.
20. Shah H, Warwick K. Testing Turing’s five-minutes, parallel-
paired imitation game. Kybernetes. 2010;39(3):449–65.
21. Shah H, Warwick K. Hidden interlocutor misidentification in
practical Turing tests. Mind Mach. 2010;20:441–54.
22. Shah H, Warwick K, Bland I, Chapman CD, Allen MJ. Turing’s
imitation game: role of error-making in intelligent thought. In:
Turing in Context II, Brussels, 10–12 October, p. 31–2, 2012.
http://www.computing-conference.ugent.be/file/14. Presentation
available here: http://www.academia.edu/1916866/Turings_Imi
tation_Game_Role_of_Error-making_in_Intelligent_Thought.
23. Sloman A. Personal communication at the Royal Society. 2014.
24. Smith GW. Art and artificial intelligence. ArtEnt. Retrieved 27
March 2015.
25. Stanford University Online Encyclopedia of Philosophy. The
Turing test. 2011. Retrieved 4 May 2015 from: http://plato.stan
ford.edu/entries/turing-test/.
26. Sterrett SG. Turing’s two tests for Intelligence. In: Moor JH,
editor. The Turing test—the elusive standard of artificial intelli-
gence (2003). Dordrecht: Kluwer; 2000. p. 79–97.
27. Traiger S. Making the right identification in the Turing test. Mind
Mach. 2000;10:561–72.
28. Turing AM. Lecture on the Automatic Computing Engine. In:
Copeland BJ, editor. The essential Turing: the ideas that gave
birth to the computer age. Oxford: Clarendon Press; 1947.
p. 2004.
29. Turing AM. Intelligent machinery. In: Copeland BJ, editor. The
essential Turing—the ideas that gave birth to the computer age,
2004 (410–432). Oxford: Clarendon Press; 1948.
30. Turing AM. Computing machinery and intelligence. Mind.
1950;LIX(236):433–60.
31. Turing AM. ‘‘Can automatic calculating machines be said to
think?’’ Transcript of BBC radio broadcast featuring Turing,
A.M., Braithwaite, R., Jefferson, G. and Newman, M. (1952). In:
Cooper SB, van Leeuwen J, editors. Alan Turing: his work and
impact, 2013. Amsterdam: Elsevier; 1952. p. 667–76.
32. Vinciarelli A, Esposito A, Andre E, Banin F, Chetouani M, Cohn
J, Cristani M, Fuhrmann F, Gilmartin E. Open challenges in
modelling, analysis and synthesis of human behaviour in human–
human and human–machine interactions. Cogn Comput.
2015;7(4):397–413.
33. Warwick K. Implications and consequences of robots with bio-
logical brains. Ethics Inf Technol. 2010;12(3):223–34.
34. Warwick K, Shah H. Assumption of knowledge and the Chinese
room in Turing test interrogation. AI Commun. 2014;27(3):275–83.
35. Warwick K, Shah H. Effects of lying in practical Turing tests. AI
Soc. 2014. 10.1007/s00146-013-0534-3.
36. Warwick K, Shah H. Good machine performance in Turing’s
imitation game. IEEE Trans Comput Intell AI Games.
2014;6(3):289–99.
37. Warwick K, Shah H. Outwitted by the hidden: unsure emotions.
Int J Synth Emot. 2014;5(1):46–59.
38. Warwick K, Shah H. Human misidentification in Turing tests.
J Exp Theor Artif Intell. 2015;27(2):123–35.
39. Warwick K, Shah H. Can machines think? A report on Turing test
experiments at the Royal Society. J Exp Theor Artif Intell. 2015.
doi:10.1080/0952813X.2015.1055826.
40. Warwick K, Shah H. Taking the 5th amendment in Turing’s
imitation game. J Exp Theor Artif Intell. 2015. doi:10.1080/
0952813X.2015.1132273.
41. Warwick K, Shah H, Moor JH. Some implications of a sample of
practical Turing tests. Mind Mach. 2013;23:163–77.
Cogn Comput (2016) 8:409–419 419
123
Content courtesy of Springer Nature, terms of use apply. Rights reserved.
1.
2.
3.
4.
5.
6.
Terms and Conditions
Springer Nature journal content, brought to you courtesy of Springer Nature Customer Service Center GmbH (“Springer Nature”).
Springer Nature supports a reasonable amount of sharing of research papers by authors, subscribers and authorised users (“Users”), for small-
scale personal, non-commercial use provided that all copyright, trade and service marks and other proprietary notices are maintained. By
accessing, sharing, receiving or otherwise using the Springer Nature journal content you agree to these terms of use (“Terms”). For these
purposes, Springer Nature considers academic use (by researchers and students) to be non-commercial.
These Terms are supplementary and will apply in addition to any applicable website terms and conditions, a relevant site licence or a personal
subscription. These Terms will prevail over any conflict or ambiguity with regards to the relevant terms, a site licence or a personal subscription
(to the extent of the conflict or ambiguity only). For Creative Commons-licensed articles, the terms of the Creative Commons license used will
apply.
We collect and use personal data to provide access to the Springer Nature journal content. We may also use these personal data internally within
ResearchGate and Springer Nature and as agreed share it, in an anonymised way, for purposes of tracking, analysis and reporting. We will not
otherwise disclose your personal data outside the ResearchGate or the Springer Nature group of companies unless we have your permission as
detailed in the Privacy Policy.
While Users may use the Springer Nature journal content for small scale, personal non-commercial use, it is important to note that Users may
not:
use such content for the purpose of providing other users with access on a regular or large scale basis or as a means to circumvent access
control;
use such content where to do so would be considered a criminal or statutory offence in any jurisdiction, or gives rise to civil liability, or is
otherwise unlawful;
falsely or misleadingly imply or suggest endorsement, approval , sponsorship, or association unless explicitly agreed to by Springer Nature in
writing;
use bots or other automated methods to access the content or redirect messages
override any security feature or exclusionary protocol; or
share the content in order to create substitute for Springer Nature products or services or a systematic database of Springer Nature journal
content.
In line with the restriction against commercial use, Springer Nature does not permit the creation of a product or service that creates revenue,
royalties, rent or income from our content or its inclusion as part of a paid for service or for other commercial gain. Springer Nature journal
content cannot be used for inter-library loans and librarians may not upload Springer Nature journal content on a large scale into their, or any
other, institutional repository.
These terms of use are reviewed regularly and may be amended at any time. Springer Nature is not obligated to publish any information or
content on this website and may remove it or features or functionality at our sole discretion, at any time with or without notice. Springer Nature
may revoke this licence to you at any time and remove access to any copies of the Springer Nature journal content which have been saved.
To the fullest extent permitted by law, Springer Nature makes no warranties, representations or guarantees to Users, either express or implied
with respect to the Springer nature journal content and all parties disclaim and waive any implied warranties or warranties imposed by law,
including merchantability or fitness for any particular purpose.
Please note that these rights do not automatically extend to content, data or other material published by Springer Nature that may be licensed
from third parties.
If you would like to use or distribute our Springer Nature journal content to a wider audience or on a regular basis or in any other manner not
expressly permitted by these Terms, please contact Springer Nature at
onlineservice@springernature.com
... Este resultado es significativo porque demuestra que, bajo ciertas condiciones, los chatbots pueden engañar a los humanos haciéndoles creer que están interactuando con otra persona (Warwick & Shah, 2016a). Sin embargo, hay que ser cautos con aseverar que las máquinas ya hayan alcanzado una inteligencia o conciencia a nivel humano, ya que esto solo demuestra que pueden imitar la comunicación humana de manera convincente por un corto periodo (Warwick & Shah, 2016b) . ...
... En 2014, se desarrollaron las redes generativas antagónicas para mejorar el rendimiento de las redes neuronales recurrentes [74]. Ese mismo año, en La Real Sociedad de Londres para el Avance de la Ciencia Natural, Eugene Goostman fue reconocido como el primer chatbot en pasar la prueba de Turing [75]. En 2016, AlphaGo venció al campeón mundial de Go, Lee Sedol, utilizando redes neuronales y aprendizaje por refuerzo [76]. ...
Article
Full-text available
El desarrollo de la Inteligencia Artificial ha impactado significativamente diversos sectores de la sociedad, incluyendo el ámbito empresarial, académico y gubernamental. Los avances recientes han sido posibles gracias a las contribuciones del pasado. Esta revisión histórica explora diversos desarrollos de la Inteligencia Artificial, desde su concepción en la mitología griega hasta los últimos desarrollos de 2023, organizados en cuatro fases. La primera fase, Concepciones Antiguas, aborda las nociones tempranas de la Inteligencia Artificial en la mitología griega y los primeros autómatas desarrollados. La segunda fase, Inicios de la Inteligencia Artificial Moderna, examina los primeros avances de la investigación científica formal de la Inteligencia Artificial. La tercera fase, Expansión y Retrocesos, está marcada por una expansión en áreas clave como los Sistemas Expertos. La cuarta fase, Resurgimiento de la Inteligencia Artificial, corresponde a la revitalización del campo, impulsada por el aprendizaje profundo. A través de un análisis cronológico de más de 150 fuentes de información, incluyendo artículos científicos, libros y documentos históricos, esta revisión proporciona una visión integral de la evolución de la Inteligencia Artificial. Además, el trabajo describe algunas soluciones de Inteligencia Artificial aplicadas en el ámbito empresarial y gubernamental peruano.
... En cambio los investigadores (Warwick & Shah, 2016) consideran que la supuesta cadena de eventos que implicaría aprobar la prueba de Turing suena como una sentencia de muerte para la humanidad. ...
Article
Full-text available
Resumen El presente artículo explora el Test de Turing, una propuesta pionera concebida por Alan M. Turing 1 para evaluar la capacidad de una máquina de presentar un comportamiento indistinguible del de un humano en una conversación. Se analizarán los fundamentos del test, sus implicaciones filosóficas y las críticas científicas que ha recibido a lo largo de los años. Además, se explorarán los avances en el campo de la inteligencia artificial que han permitido acercarse a los estándares planteados por Turing, con un enfoque en los modelos de lenguaje generativos pre entrenados y la inteligencia artificial. Palabras clave: Test de Turing, máquina de Turing, ética, inteligencia artificial.
... The test, despite much controversy, raised the fundamental question of how we want to define the intelligence of a machine. It is generally accepted that Eugene Goostman's system passed this test in 2004 [38,39]. Today, chatbots-an intelligent system developed using artificial intelligence and natural language processing (NLP) algorithms-are widely used, for example, in customer service [22]. ...
Article
Full-text available
The right to life is fundamental and primary and is a precondition for exercising other rights (Ramcharan in Ramcharan (ed), The right to life in International Law, Martinus Nijhoff Publishers, Dordrecht, 1985). Its universal recognition in the arena of international law is associated with the concept of a human being endowed with inherent and inalienable dignity. Categorization of the circle of entities covered with the right to life today seems obvious and indisputable. Intense development of artificial intelligence, also the fact that it has passed the Turing test which checks AI’s thinking ability in a way similar to human reasoning, inspires a reflection on AI’s future legal status. This study will investigate a thesis of whether artificial intelligence may be entitled to the right to life. The analysis will be carried out around an exploratory question: what are the requirements for being afforded protection of the right to life?
Article
As generative AI systems move beyond Turing’s benchmark for whether a machine exhibits human-like intelligence, what implications does this technological milestone have for organization theory? We engage with this question by considering how the increasing creativity and social competence exhibited by generative AI impacts processes of social construction and cultural evolution that have, up to this point, been the exclusive domain of humans. More specifically, we consider what it means to have intelligent machines capable of category work, which we define here as both the culturally savvy use of categories and purposeful participation in the processes of construction that underpin systems of categories more generally. We go on to explore some of the implications for individuals, organizations and societies of the appearance of this new class of artificial participants in the processes that constitute category systems.
Article
Physician burnout, a significant problem in modern healthcare, adversely affects healthcare professionals and their organizations. This essay explores the potential of artificial intelligence (AI) to positively address this issue through its integration into the electronic health record and the automation of administrative tasks. Recent initiatives and research highlight the positive impact of AI assistants in alleviating physician burnout and suggest solutions to enhance physician well-being. By examining the causes and consequences of burnout, the promise of AI in healthcare, and its integration into electronic health record systems, this essay explores how AI can not only reduce physician burnout but also improve the efficiency of healthcare organizations. A roadmap provides a visualization of how AI could be integrated into electronic health records during the previsit, visit, and postvisit stages of a clinical encounter.
Chapter
The ability to infer cognitive states is essential for the introduction of artificial intelligence (AI) into society. This chapter aims to present the existing understanding of Theory of Mind (ToM) from the point of view of AI models. First, we overview the concept of AI and the current research lines. One of the key challenges in implementing AI systems in society is human-machine interaction, as it is a very complex process that involves the ability to infer human’s mental states. Thus, the development of an artificial Theory of Mind to achieve AI systems with social traits has been suggested as a key factor for successful future human-agent interactions. Throughout the chapter, we propose to categorise the spectrum of AI ToM models into two main blocks. In the initial section, Cognitive-based and black-box algorithms, we survey different algorithms aimed at creating artificial ToM, including both those based on cognitive models and those black-box. The experiments and evaluation methods used to assess the performance of these algorithms in ToM related tasks will also be discussed. On the other hand, Bio-inspired models of ToM would include any model built taking into consideration physiological or/and neurobiological aspects of the brain for ToM modelling, aimed at pursuing biological plausibility. To conclude, we outline promising Applications of AI ToM models. These are mainly focused on industry (e.g., autonomous vehicles), speech processing, biometry, healthcare, education, or marketing, among others. An Artificial Theory of Mind is still in its infancy, as existing algorithms are domain-specific and are limited to certain environments and tasks. We project an exponential growth of AI ToM in the coming years, fuelled by technological advances and the collaboration between different fields of research such as psychology, neuroscience, computer science and mathematics.
Article
Objective: This study sought to explore the unexamined capabilities of ChatGPT in describing the surgical steps of a specialized operation, the Fisher cleft lip repair. Design: A chat log within ChatGPT was created to generate the procedural steps of a cleft lip repair utilizing the Fisher technique. A board certified craniomaxillofacial (CMF) surgeon then wrote the Fisher repair in his own words blinded to the ChatGPT response. Using both responses, a voluntary survey questionnaire was distributed to residents of plastic and reconstructive surgery (PRS), general surgery (GS), internal medicine (IM), and medical students at our institution in a blinded study. Setting: Authors collected information from residents (PRS, GS, IM) and medical students at one institution. Main outcome measures: Primary outcome measures included understanding, preference, and author identification of the procedural prompts. Results: Results show PRS residents were able to detect more inaccuracies of the ChatGPT response as well as prefer the CMF surgeon's prompt in performing the surgery. Residents with less expertise in the procedure not only failed to detect who wrote what procedure, but preferred the ChatGPT response in explaining the concept and chose it to perform the surgery. Conclusions: In applications to surgical education, ChatGPT was found to be effective in generating easy to understand procedural steps that can be followed by medical personnel of all specialties. However, it does not have expert capabilities to provide the minute detail of measurements and specific anatomy required to perform medical procedures.
Thesis
Full-text available
Deception-detection is the crux of Turing’s experiment to examine machine thinking conveyed through a capacity to respond with sustained and satisfactory answers to unrestricted questions put by a human interrogator. However, in 60 years to the month since the publication of Computing Machinery and Intelligence little agreement exists for a canonical format for Turing’s textual game of imitation, deception and machine intelligence. This research raises from the trapped mine of philosophical claims, counter-claims and rebuttals Turing’s own distinct five minutes question-answer imitation game, which he envisioned practicalised in two different ways: a) A two-participant, interrogator-witness viva voce, b) A three-participant, comparison of a machine with a human both questioned simultaneously by a human interrogator. Using Loebner’s 18th Prize for Artificial Intelligence contest, and Colby et al.’s 1972 transcript analysis paradigm, this research practicalised Turing’s imitation game across three original experiments with over 400 human participants and 13 machines. Results show the current state of artificial dialogue can recall information and share personal interests presenting an illusion of personality. The technology achieved a deception rate of 8.33% in 60 machine-human simultaneous comparison tests. Results also show that more than 1 in 3 Transcript Analysts reviewing five transcripts involving Elbot, winner of the 2008 Loebner Prize bronze award for ‘most human-like’ machine, were unable to correctly identify artificial dialogue. Deception-detection is essential to uncover the increasing number of malfeasant programmes, such as CyberLover, developed to steal identity and financially defraud users in chatrooms across the Internet. Practicalising Turing’s two tests can assist in raising awareness of this risk and preventing cybercrime.
Article
Full-text available
In this article we consider transcripts that originated from a practical series of Turing's Imitation Game that was held on 6 and 7 June 2014 at the Royal Society London. In all cases the tests involved a three-participant simultaneous comparison by an interrogator of two hidden entities, one being a human and the other a machine. Each of the transcripts considered here resulted in a human interrogator being fooled such that they could not make the ‘right identification’, that is, they could not say for certain which was the machine and which was the human. The transcripts presented all involve one machine only, namely ‘Eugene Goostman’, the result being that the machine became the first to pass the Turing test, as set out by Alan Turing, on unrestricted conversation. This is the first time that results from the Royal Society tests have been disclosed and discussed in a paper.
Article
Full-text available
Modelling, analysis and synthesis of behaviour are the subject of major efforts in computing science, especially when it comes to technologies that make sense of human-human and human-machine interactions. This article outlines some of the most important issues that still need to be addressed to ensure substantial progress in the field, namely (1) development and adoption of virtuous data collection and sharing practices, (2) shift in the focus of interest from individuals to dyads and groups, (3) endowment of artificial agents with internal representations of users and context, (4) modelling of cognitive and semantic processes underlying social behaviour and (5) identification of application domains and strategies for moving from laboratory to the real-world products.
Article
Full-text available
The modeling and implementation of sophisticated multimodal software/hardware interfaces is a current scientific challenge of high societal relevance. The main characteristics entailed by these interfaces are being able to interact with people, inferring social, organizational and physical contexts based on sensed data, assisting people with special needs, enhancing elderly health-care assistance, learning and rehabilitation in daily functional activities. Implementing such Human Computer Interaction (HCI) systems is of public utility and profitable for a living science that should simplify user’s accesses to a wide range of social services, either remotely or in a person-to-person setting. The current and future applications foreseen in this highly interdisciplinary field are countless: among these are featured context-aware avatars and robotic devices replacing and/or acting on behalf of humans in high responsibility tasks or time-critical dangerous tasks such as urban emergencies. Other emerging applications concern robot companions for elderly and vulnerable people and intelligent agents for services where there is a shortage of suitable skills or otherwise there is a request of significant investments in training-qualified personnel such as in therapist-based interventions. Given the complexities required by these automated tasks, the approach for developing such devices has to account for a holistic investigation perspective.
Article
This is the first of two volumes of essays in commemoration of Alan Turing, whose pioneering work in the theory of artificial intelligence and computer science continues to be widely discussed today. A group of prominent academics from a wide range of disciplines focus on three questions famously raised by Turing: What, if any, are the limits on machine `thinking'? Could a machine be genuinely intelligent? Might we ourselves be biological machines, whose thought consists essentially in nothing more than the interaction of neurons according to strictly determined rules? The discussion of these fascinating issues is accessible to non-specialists and stimulating for all readers.
Chapter
Alan Turing was one of the most influential thinkers of the 20th century. In 1935, aged 22, he developed the mathematical theory upon which all subsequent stored-program digital computers are modeled. At the outbreak of hostilities with Germany in September 1939, he joined the Government Codebreaking team at Bletchley Park, Buckinghamshire and played a crucial role in deciphering Engima, the code used by the German armed forces to protect their radio communications. Turing's work on the version of Enigma used by the German navy was vital to the battle for supremacy in the North Atlantic. He also contributed to the attack on the cyphers known as "Fish," which were used by the German High Command for the encryption of signals during the latter part of the war. His contribution helped to shorten the war in Europe by an estimated two years. After the war, his theoretical work led to the development of Britain's first computers at the National Physical Laboratory and the Royal Society Computing Machine Laboratory at Manchester University. Turing was also a founding father of modern cognitive science, theorizing that the cortex at birth is an "unorganized machine" which through "training" becomes organized "into a universal machine or something like it." He went on to develop the use of computers to model biological growth, launching the discipline now referred to as Artificial Life. The papers in this book are the key works for understanding Turing's phenomenal contribution across all these fields. The collection includes Turing's declassified wartime "Treatise on the Enigma"; letters from Turing to Churchill and to codebreakers; lectures, papers, and broadcasts which opened up the concept of AI and its implications; and the paper which formed the genesis of the investigation of Artifical Life.
Book
In 1950 Alan Turing (1912-1954) published his famous article, "Computing Machinery and Intelligence" in the journal Mind. This article is arguably the most influential and widely read article in the philosophy of artificial intelligence. Indeed, most of the debate in the philosophy of artificial intelligence over the last fifty years concerns issues that were raised and discussed by Turing. Turing's genius was not only in developing the theory of computability but also in understanding the impact, both practical and philosophical, that computing machinery would have. Turing believed that computers, if properly designed and educated, could exhibit intelligent behavior, even behavior that would be indistinguishable from human intelligent behavior. His vision of the possibility of machine intelligence has been highly inspiring and extremely controversial. In this classic article Turing presented his well known imitation game and predicted that about the year 2000 "an average interrogator will not have more than 70 per cent chance of making the right identification after five minutes of questioning" in the imitation game. Based on the results of the Loebner 2000 contest and the accomplishments in the field of AI, as impressive as they are, Turing's prediction remains unfulfilled.
Article
The Turing Test has captured the imagination of the general public due to fundamental questions about the nature of the mind. But Stevan Harnad argues the hype over the supposed passing of the Turing Test is misplaced. Alan Turing’s idea for cognitive science was simple: Stop worrying about what the mind “is” and explain instead what the mind does. But we are nowhere near having designed a system that can do everything a person with a mind can do.
Article
In this paper, we look at a specific issue with practical Turing tests, namely the right of the machine to remain silent during interrogation. In particular, we consider the possibility of a machine passing the Turing test simply by not saying anything. We include a number of transcripts from practical Turing tests in which silence has actually occurred on the part of a hidden entity. Each of the transcripts considered here resulted in a judge being unable to make the ‘right identification’, i.e., they could not say for certain which hidden entity was the machine.