Conference PaperPDF Available

The senior companion: a semantic web dialogue system.

Authors:
  • Recordsure

Abstract

The Senior Companion (SC) is a fully implemented Windows application intended for intermittent use by one user only (a senior citizen) over potentially many years. The thinking behind the SC is to make a device that will give its owner comfort, company, entertainment, and some practical functions. The SC will typically be installed at home, either as an application on a personal computer, or on a dedicated device (like a Chumby) or an intelligent coffee table (like Microsoft's Surface). By means of multimodal input and output, and a graphical interface, the SC provides its 'owner' with different functionalities, which currently include: • conversing with the user about his personal photos • learning about the user, user's family, and life history • telling the user jokes • reading the news (via RSS feed from the internet)
13831383
The Senior Companion: a Semantic Web Dialogue System
Debora Field
d.¿eld@shef.ac.uk Roberta Catizone
roberta@dcs.shef.ac.uk WeiWei Cheng
w.cheng@dcs.shef.ac.uk
Alexiei Dingli
alexiei@dingli.org
Simon Worgan
s.worgan@dcs.shef.ac.uk
Lei Ye
l.ye@dcs.shef.ac.uk
Yorick Wilks
yorick@dcs.shef.ac.uk
1. APPLICATION DOMAIN
The Senior Companion (SC) is a fully implemented Win-
dows application intended for intermittent use by one user
only (a senior citizen) over potentially many years. The
thinking behind the SC is to make a device that will give its
owner comfort, company, entertainment, and some practical
functions. The SC will typically be installed at home, either
as an application on a personal computer, or on a dedicated
device (like a Chumby1) or an intelligent coffee table (like
Microsoft’s Surface). By means of multimodal input and
output, and a graphical interface, the SC provides its ‘owner’
with different functionalities, which currently include:
conversing with the user about his personal photos
learning about the user, user’s family, and life history
telling the user jokes
reading the news (via RSS feed from the internet)
Chatting about photos is currently the SC’s main activity.
This initial direction was chosen on the assumption that
senior citizens enjoy browsing photos and being reminded of
events and people from their lives. The SC will currently
also tell its owner jokes and read the news, if requested.
2. AGENT TECHNIQUES
Goals
The goals of the SC are more vague than those of most
dialogue systems, which are often focused on task fulfilment.
The overarching goal of the SC is to be a friendly, entertain-
ing, and useful companion for its owner. One subgoal of this
overarching goal is to encourage the user to talk about his
life, while using personal photos as a prompt. Another sub-
goal is for the SC to learn the details of significant events in
the user’s life, so as to be able to construct a story or time-
line of the user’s life, and to be a knowledge source for other
family members (grandchildren, for example). Another sub-
goal is for the SC to learn about the user’s personal photos
Affiliation of all authors is Department of Computer Sci-
ence, University of Sheffield, S1 4DP, UK.
1http://www.chumby.com
Cite as: The Senior Companion: a Semantic Web Dialogue System,
D G Field, R Catizone, W Cheng, A Dingli, S Worgan, L Ye, Y Wilks,
Proc.of8thInt.Conf.onAutonomousAgentsandMultia-
gent Systems (AAMAS 2009), Decker, Sichman, Sierra and Castel-
franchi (eds.), May, 10–15, 2009, Budapest, Hungary, pp. XXX-XXX.
Copyright c
2008, International Foundation for Autonomous Agents and
Multiagent Systems (www.ifaamas.org). All rights reserved.
and what they depict, so that the SC can retrieve photos on
the basis of what they depict, to enhance conversation.
Dialogue Manager
The agency of the SC is embodied in its dialogue man-
ager (DM). The DM uses the stack architecture (after [2]
and COMIC2) to manage dialogue, employing a set of hand-
crafted augmented transition networks we call ‘Dialogue Ac-
tion Forms’ (DAFs). DAFs are individuated by conversa-
tional topic—when the conversation moves to a new topic,
a new DAF is pushed onto the stack. Although the DAFs
are hand-crafted, their design is informed by data from a
spoken dialogue corpus,3and we are making preparations
to add decision theory to the DAFs to enable the DM to
make probabilistic decisions based on current context.
At login the SC speaks to the user, welcoming him, and
begins a conversation. The SC asks the user questions, re-
members some content from the user’s answers, and makes
statements. Some statements are motivated by inferences
the SC has made, and some come from a chatbot. User ut-
terances are processed first as input to an Automatic Speech
Recogniser (ASR). The DM sends the output of the ASR
to the Natural Language Understander (NLU). We are us-
ing an Information Extraction approach to NLU exploiting
GATE[1] plug-ins. Named Entity Recognition is a main fea-
ture of the NLU. In many cases, the DM sends the NLU
a specification of the type of information it is expecting to
be contained in the user response. For example, the DM
might tell the NLU that it is looking for a family relation-
ship in the user’s utterance content. The information types
the NLU can recognise are as follows:
person names
person relationships (family and other)
location names
prepositional phrases that describe locations
dates
time phrases not containing an explicit date
occasions (e.g., weddings, funerals, birthdays)
If the DM does not receive from the NLU the information
it is expecting at that point, it will typically apologise and
repeat the question to the user up to three times, or it will
invoke the chatbot.
2http://www.hcrc.ed.ac.uk/comic
3Collected under Companions [3].
Cite as: The Senior Companion: a Semantic Web Dialogue System,
Debora Field, Roberta Catizone, WeiWei Cheng, Alexiei Dingli, Simon
Worgan, Lei Ye, Yorick Wilks, Proc. of 8th Int. Conf. on Autonomous
Agents and Multiagent Systems (AAMAS 2009), Decker, Sichman, Sierra
and Castelfranchi (eds.), May, 10–15, 2009, Budapest, Hungary, pp. 1383 – 1384
Copyright © 2009, International Foundation for Autonomous Agents and
Multiagent Systems (www.ifaamas.org), All rights reserved.
AAMAS 20098th International Conference on Autonomous Agents and Multiagent Systems 10–15 May, 2009 Budapest, Hungary
1384
Th
e sys
t
em can recogn
i
se w
h
en
th
euser
su
tt
erance s
h
ows
he has changed his mind about something he has just said
(typically by using phrases like “Oh no” and “that’s not
right”). The system will then use a clarification routine to
ensure that it has understood the correct information.
Knowledge and reasoning
Information that the DM learns from a user is stored
as triples (binary predicates) in an RDF triplestore. The
knowledge base is essentially monotonic, with the exception
of the clarification routine just mentioned above, in which
the system replaces a fact it discovers is false with one it
now thinks is true.
The knowledge base also contains a small set of inference
rules describing family relationships. At the end of each user
utterance the DM calls a reasoner to infer new information.
The reasoner is called forwards to infer everything it can
about everything it knows. The newly inferred information
is then exploited by the DM in its subsequent utterances at
appropriate points.
With regard to ascertaining the date when a particular
photo was taken, if the user uses a time phrase that is not an
explicit date (for example, “It was taken six years ago”), the
system invokes procedures for working out the approximate
date when the photo was taken.
Object recognition
As the conversation progresses, the system not only learns
about events and people in the user’s life, it also makes an
inventory of what is depicted in each photo, and it links the
photos and the things depicted in them to the facts being
stored in its knowledge base. In order to recognise that dis-
tinct objects are depicted in a photo, the system has to be
able to see the photo, at least in some limited sense. Cur-
rently the system is able to detect front-facing human faces
(using OpenCV4), but it cannot distinguish between faces,
and so cannot recognise people. We are, however, about
to replace this with a face recognition system developed by
Polar Rose.5We also intend to add the ability to recognise
other object types, such as monuments and landmarks.
3. INNOVATIONS
The key innovation of the Senior Companion is the use
of Semantic Web technologies to build a dialogue system.
This approach was chosen in order to provide a seamless
join between a dialogue system and the internet, and there-
fore to maximise the potential for exploiting open access
knowledge when planning the content of a system’s conver-
sational utterances. Currently, the ways in which we exploit
the internet to enhance conversational abilities are:
to enable easy access to one’s own or other people’s
personal digital photographs via Facebook.
to look for tourist attractions near a place mentioned
by the user, so as to chat to the user about them.
to invoke an online chatbot at appropriate points.
to provide live, up-to-date news.
to supply particular kinds of jokes.
The above are just the first steps towards an internet-
driven dialogue system.
Other innovations under development include:
4http://sourceforge.net/projects/opencvlibrary/
5http://www.polarrose.com
us
i
ng mac
hi
ne
l
earn
i
ng
(ML) t
o
d
eve
l
op a
th
eory o
f
how to monitor and respond to user emotions during
conversation, and building that into the DM.
using ML to derive dialogue structure from a corpus.
using reasoning to guide the conversation towards the
system’s topics of greatest ignorance about the user
(which we call ‘grounding in the user’).
4. LIVE AND INTERACTIVE ASPECTS
Hardware
The SC is a Windows application that runs on a personal
computer. To interact with the SC, a microphone is needed
for user input, and speakers for system output.
Personal photos
Before a user interacts with the SC some of his personal
photos must have been uploaded. They can either be stored
statically on the hard disk, or they can be uploaded from a
Facebook album via the internet.
Input modalities
The principal input modality is speech. The first time
only that the user interacts with the SC, the SC leads him
through a ten-minute voice training session with the ASR.6
The user may also type his utterances into a text box,
which is provided mainly for error correction. For example,
if the ASR repeatedly misinterprets part of a user utterance,
and this leads to the DM not receiving anything useful from
the NLU, the user can type his utterance. (The user can see
from the interface when the ASR makes mistakes.)
A third input modality not fully exploited yet is touch.
The user can point on the touch-sensitive screen with an
electronic pen. He might typically do this when saying, for
example, “This is my sister”, while touching the image of
his sister on the screen. The DM knows the co-ordinates of
where the screen has been touched, and knows which areas
on a particular image depict faces. By aligning these, the
system will be able to exploit the user’s touch input in its
conversational utterances and inferences.
5. ACKNOWLEDGMENTS
This work was funded by Companions [3], European Com-
mission Sixth Framework Programme Information Society
Technologies Integrated Project IST-34434.7
6. REFERENCES
[1] H. Cunningham, D. Maynard, K. Bontcheva, and
V. Tablan. GATE: A framework and graphical
development environment for robust NLP tools and
applications, 2002. In Proc. 40th Anniversary Meeting
of the Association for Computational Linguistics
(ACL).
[2] O. Lemon, A. Bracey, A. Gruenstein, and S. Peters.
The WITAS multimodal dialogue system, 2003. In
Proc. Eu rospeech 20 03.
[3] Y. Wilks. Artificial companions, 2005. Interdisciplinary
Science Reviews , June, Vol. 30, pp. 145–152.
6The current ASR is Dragon Naturally Speaking.
7http://www.companions-project.org/
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly (Field et al., 2009), and a health and fitness companion (Stahl et al., 2009). In our work, together with industry partners , we are developing a speech-enabled companion toy for children. ...
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly ( Field et al., 2009), and a health and fitness companion ( Stahl et al., 2009). In our work, together with industry partners, we are developing a speech-enabled companion toy for children. ...
Article
Full-text available
In line with the growing interest in conversational agents as companions, we are developing a toy companion for children that is capable of engaging interactions and of developing a long-term relationship with them, and is extensible so as to evolve with them. In this paper, we investigate the importance of personalising interaction both for engagement and for long-term relationship development. In particular, we propose a framework for representing, gathering and using personal knowledge about the child during dialogue interaction.
... The first and foremost challenge [7] of the chatbot is to handle NLP issue by mastering their syntax. If we ask them that " what's the weather?". ...
Conference Paper
Full-text available
In the modern era of technology, Chatbots are the next big thing in the era of conversational services. Chatbots is a virtual person who can effectively talk to any human being using interactive textual skills. Currently, there are many cloud base Chatbots services which are available for the development and improvement of the chatbot sector such as IBM Watson, Microsoft bot, AWS Lambda, Heroku and many others. A virtual person is based on machine learning and Artificial Intelligence (AI) concepts and due to dynamic nature, there is a drawback in the design and development of these chatbots as they have built-in AI, NLP, programming and conversion services. This paper gives an overview of cloud-based chatbots technologies along with programming of chatbots and challenges of programming in current and future Era of chatbot.
... As dialog systems support a broad range of applications in business enterprises, education healthcare and entertainment [2]- [4], the study incorporated a chatbot to a dialog system to automate resume based employment interviews. The researchers chose this type of employment interviews because the user answers can be verified via their resume. ...
Article
Full-text available
Traditional resume based recruitment interviews and in-person interviews only allow companies to handle only a limited number of job applicants at a time. As a result, a substantial amount of time and money is misdirected on interviewing unqualified job applicants. The proponents developed a resume based employment interview chatbot, using an enhanced example based dialog model, to evaluate job applicants' consistency in their resume details and interview answers. The chatbot will replace the HR interviewer while maintaining the fundamental quality and naturalness of a resume based interview. The study aimed to improve the current hiring process, specifically the initial resume based interview conducted during job applicants' screening and also utilized the potential of chatbots to further improve their simulation of intelligent conversations by having an in-depth analysis on the content and meaning of the user's input. Index Terms—Chatbots, dialog system, expert system, natural language processing.
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly (Field et al., 2009), and a health and fitness companion (Stahl et al., 2009). In our work, together with industry partners, we are developing a speech-enabled companion toy for children. ...
Article
Full-text available
I start from a perspective close to that of the EC COMPANIONS project, and set out its aim to model a new kind of human-computer relationship based on long-term interaction, with some tasks involved although a Companion should not be inherently task-based, since there need be no stopping point to its conversation. Some demonstration of its functionality will be given but the main purpose here is an analysis of what it is people might want from such a relationship and what evidence we have for whatever we conclude. Is politeness important? Is an attempt at emotional sympathy important or achievable? Does a user want a consistent personality in a Companion or a variety of personalities? Should we be talking more in terms of a "cognitive prosthesis (or orthosis)?"---something to extract, organize, and locate the user's knowledge or personal information---rather than attitudes?
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly ( Field et al., 2009), and a health and fitness companion ( Stahl et al., 2009). In our work, together with industry partners, we are developing a speech-enabled companion toy for children. ...
Article
A technology demonstrator is one thing but having people use a technology is another, and the result reported here is that people often ignore our lovingly crafted handiwork. The SERA project - Social Engagement with Robots and Agents - was set up to look explicitly at what happens when a robot companion is put in someone's home. Even if things worked perfectly, there are times when a companion's human is simply not engaged. As a result we have separated our "dialog manager" into two parts: the dialog manager itself that determines what to say next, and an "interaction manager" that determines when to say it. This paper details the design of this SALT-E architecture.
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly (Field et al., 2009), and a health and fitness companion (Stahl et al., 2009 ). In our work, together with industry partners , we are developing a speech-enabled companion toy for children. ...
Article
Full-text available
We describe a 'How was your day?' (HWYD) Companion whose purpose is to establish a comforting and supportive relationship with a user via a conversation on a variety of work-related topics. The system has several fairly novel features aimed at increasing the naturalness of the interaction: a rapid 'short loop' response primed by the results of acoustic emotion analysis, and an 'interruption manager', enabling the user to interrupt lengthy or apparently inappropriate system responses, prompting a replanning of behaviour on the part of the system. The 'long loop' also takes into account the emotional state of the user, but using more conventional dialogue management and planning techniques. We describe the architecture and components of the implemented prototype HWYD system.
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly (Field et al., 2009), and a health and fitness companion (Stahl et al., 2009). In our work, together with industry partners, we are developing a speech-enabled companion toy for children. ...
Article
Full-text available
The purpose of this research was to advance the understanding of the behavior of small groups in online chat rooms. The research was conducted using Internet chat data collected through planned exercises with recruited participants. Analysis of the collected data led to construction of preliminary models of social behavior in online discourse. Some of these models, e.g., how to effectively change the topic of conversation, were subsequently implemented into an automated Virtual Chat Agent (VCA) prototype. VCA has been demonstrated to perform effectively and convincingly in Internet conversation in multiparty chat environments.
... In recent years there has been an increasing interest in so-called Companion agents: agents that are intelligent, and built to interact naturally (via speech and other modalities) with their user over a prolonged period of time, personalising the interaction to them and developing a relationship with them. The EU Companions project 2 is the most well known such project, with applications such as a companion for the elderly (Field et al., 2009), and a health and fitness companion (Stahl et al., 2009 ). In our work, together with industry partners , we are developing a speech-enabled companion toy for children. ...
Conference Paper
Full-text available
We present a family of Embodied Conversa- tional Agents (ECAs) using Talking Head technology, along with a program of associ- ated research and user trials. Whilst antece- dents of our current ECAs include "chatbots" desgined to pass the Turing Test (TT) or win a Loebner Prize (LP), our current agents are task-oriented Teaching Agents and Social Companions. The current focus for our re- search includes the role of emotion, expres- sion and gesture in our agents/companions, the explicit teaching of such social skills as recognizing and displaying appropriate ex- pressions/gestures, and the integration of template/database-based dialogue managers with more conversational TT/LP systems as well as with audio-visual speech/gesture rec- ognition/synthesis technologies.
Article
In this review of Stephen Cowley’s special issue of Pragmatics & Cognition [Cowley, S.J., 2009. Distributed language and dynamics. Pragmatics & Cognition 17 (3), 495–508] we present an overview of the extensive overlap between the distributed language approach and the development of artificial dialogue systems. After presenting a brief summary of each of the articles in the Distributed Language special issue of Pragmatics & Cognition we will discuss how this new approach could contribute to existing artificial models of language and the production of effective human–computer conversational interactions. In doing so we will briefly highlight the mutual advantages of an interdisciplinary program of research, one that could unify the varied strands of distributed language and contribute to current Artificial Intelligence research programs.
Conference Paper
This paper presents a real-time implementation of an automatic dialogue system called ‘Senior Companion’, which is not strictly task-oriented, but instead it is designed to ‘chat’ with elderly users about their family photographs. To a large extent, this task has lost the usual restriction of dialogue systems to a particular (narrow) domain, and thus the speech and natural language processing components had to be designed to cover a broad range of possible user and system utterances. Keywordsautomatic dialog systems-human-computer interaction-speech technologies-natural language processing
Conference Paper
Full-text available
What will an artificial Companion be like? Who will need them and how much good or harm will they do? Will they change our lives and social habits in the radical way technologies have in the past: just think of trains, phones and television? Will they force changes in the law so that things that are not people will be liable for damages; up till now, it is the case that if a machine goes wrong, it is always the maker or the programmer, or their company, which is at fault. Above all, how many people, with no knowledge of technology at all, such as the old and very young, will want to go about, or sit at home, with a companion that may look like a furry handbag on the sofa, or a rucksack on the back, but which will keep track of their lives by conversation, and be their interface to the rather elusive mysteries we now think of as the Internet or Web.
Conference Paper
Full-text available
We present the first demonstration version of the WITAS dia- logue system for multi-modal conversations with autonomous mobile robots, and motivate several innovations currently in de- velopment for version II. The human-robot interaction setting is argued to present new challenges for dialogue system en- gineers, in comparison to previous work in dialogue systems under the travel-planning paradigm, in that dialogues must be asynchronous, mixed-initiative, open-ended, and involve a dy- namic environment. We approached these general problems in a dialogue interface to the WITAS robot helicopter, or UAV ('Un- manned Aerial Vehicle').1 We present this system and the mod- elling ideas behind it, and then motivate changes being made for version II of the system, involving more richly structured dia- logue states and the use of automated reasoning systems over task, ability, and world-state models. We argue that these sorts of enhancement are vital to the future development of conversa- tional systems.
Conference Paper
Full-text available
The paper concerns the closely related topics of the possibility of machines having identifiable personalities, and the possible future legal responsibilities of machines. As win become clear, I wish to explore these topics in terms which are basically those of computational linguistics; and by 'machines' here, I intend software entities, rather than robots, and in particular the sort of software agents now being encountered on the web, ranging at present from technical advisers to mere chatbots. I call them Companions. The body of the paper explores the following related aspects of a Companion in more detail: what kind and level of personality should be in a machine agent so as to be acceptable to a human user, more particularly to one who may fear technology and have no experience of it; and what levels of responsibility and legal attribution for responsibility can we expect or desire from entities like web agents in the near future?
The WITAS multimodal dialogue system
  • O Lemon
  • A Bracey
  • A Gruenstein
  • S Peters
O. Lemon, A. Bracey, A. Gruenstein, and S. Peters. The WITAS multimodal dialogue system, 2003. In Proc. Eurospeech 2003.