Conference PaperPDF Available

What’s cooking and Why? Behaviour Recognition during Unscripted Cooking Tasks for Health Monitoring

Authors:

Abstract and Figures

Nutrition related health conditions can seriously decrease quality of life; a system able to monitor the kitchen activities and eating behaviour of patients could provide clinicians with important indicators for improving a patient’s condition. To achieve this, the system has to reason about the person’s actions and goals. To address this challenge, we present a behaviour recognition approach that relies on symbolic behaviour repre- sentation and probabilistic reasoning to recognise the person’s actions, the type of meal being prepared and its potential impact on a patient’s health. We test our approach on a cooking dataset containing unscripted kitchen activities recorded with various sensors in a real kitchen. The results show that the approach is able to recognise the sequence of executed actions and the prepared meal, to determine whether it is healthy, and to reason about the possibility of depression based on the type of meal.
Content may be subject to copyright.
What’s cooking and Why? Behaviour
Recognition during Unscripted Cooking Tasks for
Health Monitoring
Kristina Yordanova1,2, Samuel Whitehouse2, Adeline Paiement2,
Majid Mirmehdi2, Thomas Kirste1, Ian Craddock2
1University of Rostock, 18051 Rostock, Germany
2University of Bristol, Bristol BS8 1UB, UK
April 3, 2017
Abstract
Nutrition related health conditions can seriously decrease quality of life; a sys-
tem able to monitor the kitchen activities and eating behaviour of patients could
provide clinicians with important indicators for improving a patient’s condition.
To achieve this, the system has to reason about the person’s actions and goals.
To address this challenge, we present a behaviour recognition approach that relies
on symbolic behaviour representation and probabilistic reasoning to recognise the
person’s actions, the type of meal being prepared and its potential impact on a
patient’s health. We test our approach on a cooking dataset containing unscripted
kitchen activities recorded with various sensors in a real kitchen. The results show
that the approach is able to recognise the sequence of executed actions and the pre-
pared meal, to determine whether it is healthy, and to reason about the possibility
of depression based on the type of meal.
1 Introduction
Nutrition affects our health and is an important factor for having a healthy lifespan
[7]. Nutrition related diseases thus can impact our well-being and reduce the quality of
life. This is particularly true for long term physical conditions, such as diabetes, eating
disorders, or mental conditions such as depression, that affect the patient’s willingness
to prepare and consume healthy food, or people suffering from dementia disorders
whose ability to prepare food is hampered by the disease’s progression [10]. To reduce
the costs associated with hospitalisation and treatment of these conditions, different
works have attempted to provide automated home monitoring of the patient that besides
reducing the hospitalisation costs, potentially improves the wellbeing of the patient as
they can be monitored and treated in home settings [2].
To build such systems, one needs to recognise the user actions and goals as well as
the causes behind the observed behaviour [5]. To achieve that, different works make use
1
of knowledge-based models [9]. In contrast to data-driven approaches that rely on large
amounts of sensor data and which can learn only situations that are present in the data,
knowledge-based approaches have the advantage that they are able to reason beyond
the sensor data. Thanks to the rules that define the possible behaviour, they can provide
information about the person’s situation, e.g. caused by the progression of the disease
[10]. The main challenge for rule-based approaches is that they are usually unable to
cope with the problems associated with real world scenarios: (a) the variability of the
user behaviour results in complex models that are often computationally infeasible, and
(b) the presence of imperfect sensors makes the purely symbolic models unable to cope
with the ambiguity.
To address these problems, some works propose approaches that combine rules and
probabilistic inference, such as [3, 8]. This type of models is also known as computa-
tional state space models (CSSMs) [4]. CSSMs combine symbolic representation with
probabilistic reasoning to cope with the combination of behaviour variability and sen-
sor noise [3, 4]. One challenge with CSSMs is that so far they have only been applied to
scripted scenarios which implies simplified settings that do not address the challenges
of complexity and behaviour variability present in real settings. Another challenge is
the reconstruction of the user activities and goals from low level sensor data. As [11]
point out, “bridging the gap between noisy, low-level data and high-level activity mod-
els is a core challenge”. In this work, we address the above challenges by presenting
first empirical results showing that CSSMs are able to reason about the user’s actions
and goals in unscripted kitchen scenarios based on low level sensor data.
2 Related Work
There are different rule-based approaches that can reason about the person’s actions
and goals based on context information. One such approach is where ontology-based
behaviour libraries are explicitly provided by human experts [9]. One problem with
this type of approaches is that “library-based models are inherently unable to solve
the problem of library completeness caused by the inability of a designer to model all
possible execution sequences leading to the goal” [13].
A second option for arriving at a suitable model is to mine action sequences from
observations of human behaviour. Such approaches manually define an initial library
of behaviours. Later, behaviour variations are added or removed based on observations
of the user activities [1]. Although this approach provides a solution to the problem
of keeping behaviour libraries up-to-date, it still relies on initial manual definitions of
behaviour variations.
To address the problem of designing models that represent the behaviour variability
without relying on large amounts of sensor data, some works propose computational
state space models [3, 8, 4, 13]. CSSMs describe the actions in terms of preconditions
and effects, and some of them allow probabilistic reasoning about the user state, goals
and context. The manually defined model is very compact as it requires the definition
of several action templates that are automatically expanded into different execution
sequences based on the causal relations between them. This provides an alternative
solution to the problem of manually defining all execution sequences, or relying on
2
large amounts of annotated sensor data to learn them.
So far, CSSMs have been used only in controlled experiments with predefined ex-
ecution sequences. Such experiments limit the behaviour variability typical for real-
world problems, which in turn simplifies the CSSM model needed to recognise this
behaviour. To our knowledge, so far there is no empirical evidence that CSSMs are
able to cope with the behaviour complexity typical for a real world daily activities.
What is more, so far CSSMs have been used for goal recognition only based on sim-
ulated data [3, 8]. In a previous work we proposed a CSSM model that is able to
recognise the protagonist’s activities during unscripted kitchen tasks [12]. The model
was tested on simulated sensor data. In this work, we extend this model for activity
recognition based on real sensor data by building the appropriate observation model
and actions’ durations. Furthermore, we extend the model for goal recognition and
we show first empirical evidence that it is possible to perform both activity and goal
recognition based on noisy sensor data in real-world everyday scenarios.
3 Computational Causal Behaviour Models
The CSSM approach we chose for our problem is called Computational Causal Be-
haviour Models (CCBM) which has been shown to perform adequate activity recogni-
tion in problems with large state spaces and noisy sensor observations [4, 13]. CCBM
relies on the idea of Bayesian filtering to recognise the person’s actions and goals based
on observations. Figure 1 shows the dynamic Bayesian network (DBN) structure of a
CCBM model. Informally, CCBM can be divided into two parts: observation model,
which provides the probability of having an observation given a certain system state
and system model, which provides the probability of the current system model state
given the previous state. The observation model is defined through Yt= (Wt, Zt),
the observation data for time step t, e.g sensor data collected from low level sensors or
simulated data. So far CSSMs had either performed only activity recognition using real
sensor data [4, 13] or they have used simulated data to perform goal recognition [3, 8].
In this work we use low level sensor data to perform both activity and goal recognition.
The system model consists of causal model (expressed through Gt,At,St), duration
model (expressed through Dtand Ut), and action selection heuristics (which are not
represented in Figure 1). Gtis the current goal the person is pursuing. In difference to
[4, 13] who assume that the current goal is a constant, in this work we follow different
goals and the goal can change dynamically over time. Stis the high-level model state
(the state describing the person’s behaviour). It is either the result of applying a new
action Ator carrying over the old state St1. Actions can last longer than a single
time step, i.e. they have durations. Utdenotes the starting time of an action, while the
boolean random variable Dtindicated the termination status of the previous action. Vt
is the time stamp associated to the DBN slice.
In CCBM the causal model is presented in terms of rules that describe the possible
initial and goal states, the conditions that have to hold in order for an action to be
executable and the changes to the world after the action is executed. To select a new
action CCBM uses action selection heuristics such as goal distance, cognitive heuristic
etc. The duration model is expressed through probability distribution indicating the
3
t
t1
Dt1
Vt1
St1
At1
Ut1
Wt1
Zt1
Gt1
Dt
Vt
St
At
Ut
Wt
Zt
Gt
Xt
Xt1
Yt
Yt1
Figure 1: DBN structure of a CCBM model. Adapted from [4].
probability of terminating the action given the starting time of the action. For more
details on CCBM see [4, 13].
4 Experimental Setup
Data Collection A sensor dataset consisting of 15 runs of kitchen activities was
recorded in the SPHERE House which is part of the SPHERE project (a Sensor Plat-
form for HEalthcare in a Residential Environment) [14]. The SPHERE House in Bris-
tol (UK) is a 2-bedroom house equipped with a variety of environmental sensors. The
sensor network in the kitchen of the house collects data on temperature,humidity,light
levels,noise levels,dust levels,motion within the room,cupboard and room door state,
and water and electricity usage. A head-mounted camera was used to record the actions
of the participants to allow for annotation of the observations. The resulting dataset can
be downloaded from [6].
Data Processing The original sensor data is in JSON format. The data was converted
so that for each type of sensor there is a separate column. This conversion produced
4
multiple rows with the same timestamp (in milliseconds). Rows with the same times-
tamp were then combined as long as per sensor type there was only one unique value.
As this new format produces NAs for some sensors at a given time (due to the way
in which the data is initially collected from the sensors), the NAs between two sensor
readings were replaced with the first value. As the state of the most sensors is being
read at a certain sampling rate but additionally there is a reading when a change in the
state is detected, we believe that this simple replacement of NAs is sufficient. The re-
sulting data contained identical observations for different action labels. To reduce the
impact of this artefact on the model performance, a sliding window of 5 time steps with
overlapping of 50% was used and the observations in this window were represented by
the maximum value for each sensor in the window.
CCBM Models In this work we use extended version of the model proposed in [12]
where they use it for activity recognition on simulated data. Here, we extend the model
by adding probabilistic action durations and goal recognition and use it with real sensor
data for following different goals.
Causal model: 15 specialised models, specifically fitted for the corresponding exe-
cution sequence, were developed. Furthermore, a general model was developed which
can handle all sequences in the dataset. Each of the models can recognise the fol-
lowing action classes: clean,drink,eat,get,move,prepare,put,unknown. The model
dimensions for the two model implementations can be seen in Table 1. Some additional
Table 1: Parameters for the different models.
Parameters General model Specialised model
Action classes 8 8
Ground actions 92 10 – 28
States 450 144 40 – 1288
Valid plans 21 889 393 162 – 15 689
discussion on the models can be found in [12].
Goals in the model: The model has three types of goals: 1) the type of meal the
person is preparing (13 goals); 2) whether the meal / drink is healthy or not (4 goals);
3) whether the person is depressed or not (2 goals). For 3) we rely on the assumption
that the person is depressed when they are preparing ready meals instead of cooking.
Duration model: The durations of the model were calculated based on the annota-
tion. Empirical probability was assigned to each action class, indicating how long the
model can stay in the same state before transitioning to another state.
Observation model: Two types of observation models were trained with a decision
tree: 1) OMo: All data was used both for training the OM and for testing the CCBM
model. 2) OMp: The first run was used for training the OM and the remaining runs
were used for testing the CCBM. The first run was chosen because it is the only run
where all action classes appear.
The decision tree for OMowithout any additional underlying model achieved mean
accuracy of .52. The decision tree for the OMpachieved mean accuracy of .39. This is
5
to be expected as the tree was trained only on the first run and in each of the remaining
experiments, a different meal was prepared usually by a different person.
Experiments For each of the observation models, the following experiments were
conducted: 1) activity recognition of the action classes based on: the specific CCBM
model (we call this model CC BMs); the general CCBM model with one goal1(we
call this model CC BMg); the general CCBM model with multiple goals2(we call
this model CC BMg1). 2) goal recognition on: the different meals and drinks that
can be prepared (the goal recognition is done with the CCBMg1model); whether
the prepared meal is healthy (we call this model CC BMg2); whether the person is
depressed (we call this model CC BMg3).
5 Results
Figure 2 shows the results from the activity recognition with the different observation
and system models. It can be seen that for both OM, the CCBM models performed bet-
ter than the classification with decision tree. For most models, Shapiro-Wilk normality
test did not reject the null hypothesis that the samples come from normal distribution
with the exception of CC BMsand CC BMg1with OMp(p value 0.03). For that
reason, to test whether the results significantly differ from each other we performed
both signed t test and Wilcoxon test. Both tests showed that all CCBM models with
OModo not significantly differ from each other (p 0.84 for t test and p 0.74 for
Wilcoxon test). This indicates that the general models do not significantly reduce the
recognition rate in comparison to the specialised model. This however showed that all
three CCBM models with OMosignificantly differ from the decision tree tested on the
training data (p 0.006 for t test and p 0.003 for Wilcoxon test). On the other hand,
OM−o
OM−p
0.00
0.25
0.50
0.75
1.00
accuracy
models
CCBM−g
CCBM−g1
CCBM−s
DT
Figure 2: Accuracy for the activity recognition.
both tests illustrated that the CCBM models with OMpand the decision tree where
a test and train dataset was used do not significantly differ (p 0.73 for t test and p
0.86 for Wilcoxon test). This stands to show that the very ambiguous observation
1This means that all possible meals are described with “OR” statement.
2This means that all possible meals are described as separate goals.
6
model does not allow for the system model to improve the activity recognition perfor-
mance. However, it also shows that despite the inaccurate OM, the system models do
not reduce the recognition performance.
Figure 3 shows the results from the goal recognition for the different observation
models. Here we measure the F score, as for some of the experiments more goals
were recognised than were followed during the experiment. Surprisingly, OMpdid not
reduce recognition of the goal for healthy meal and the type of meal. On the contrary,
despite the ambiguous OM and the low activity recognition results, the models were
able to perform better than when using OMo. On the other hand, the recognition of
depressed
healthy
meal
0.00
0.25
0.50
0.75
1.00
f.measure
models
OM−o
OM−p
Figure 3: Accuracy for the goal recognition.
whether the person is depressed or not was lower with OMpthan with OMo. To test
whether the results for the two OM significantly differ, we performed a Wilcoxon test
(the Shapiro-Wilk normality test rejected the null hypothesis that the samples come
from a normal distribution with a p 0.001). The Wilcoxon test showed that results
for the two OMs for meal recognition and healthy meal do not significantly differ (p
=0.86 for CC BMg1, and p =0.40 for CCBMg2). This means that the models
perform comparably and that the inaccurate observation model OMpdoes not reduce
the accuracy of recognising the person’s goal. The results however showed that the
OM has influence on the recognition of whether the person is depressed or not (p =
0.004 for CC BMg3). In other words the more accurate OMosignificantly improved
the recognition of the cause for the prepared meal. Figure 4 shows an example of the
probability of preparing a healthy meal / drink for one of the experiments. The true
goal is “healthy meal” and it can be seen that after time step 125 the model converges
to the real goal. Although in this example, the protagonist pursues only one goal, in
4 of the experiments the goal changes throughout the experiment as the protagonist
prepares more than one meal or drink.
6 Conclusion and Future Work
This work investigates the applicability of CSSMs to real world everyday activities. We
applied the approach to a sensor dataset containing 15 unscripted meal preparations.
The results showed that the approach is able to perform activity recognition without
reducing the recognition quality compared to the performance of decision tree. They
also showed the approach is able to perform goal recognition and to accurately reason
7
Figure 4: Evolving of the goals’ probability with the accumulation of new observations
during the experiment execution.
about the type of meal, whether it is healthy and whether the person preparing the meal
is depressed even in the case of poor activity recognition results. These first results
show that the approach has the potential to reason about the person’s behaviour and the
causes behind it that could hint at (the progression of) medical conditions.
In the future, we intend to compare the results from the goal recognition to state of
the art approaches, such as HMMs. Furthermore, we intend to add the data from depth
cameras to the observations and to investigate the influence of the type of sensor on the
model performance.
7 Acknowledgments
This work is partially funded by the UK Engineering and Physical Sciences Research
Council (EPSRC) within the context of the SPHERE IRC, grant number EP/K031910/1.
References
[1] L. Chen, C. Nugent, and G. Okeyo. An ontology-based hybrid approach to activ-
ity modeling for smart homes. IEEE Transactions on Human-Machine Systems,
44(1):92–105, Feb 2014.
[2] A. Helal, D. J. Cook, and M. Schmalz. Smart home-based health platform for
behavioral monitoring and alteration of diabetes patients. Journal of Diabetes
Science and Technology, 3(1):141–148, January 2009.
[3] L. M. Hiatt, A. M. Harrison, and J. G. Trafton. Accommodating human variability
in human-robot teams through theory of mind. In Proceedings of IJCAI, pages
2066–2071. AAAI Press, 2011.
8
[4] F. Kr¨
uger, M. Nyolt, K. Yordanova, A. Hein, and T. Kirste. Computational state
space models for activity and intention recognition. a feasibility study. PLoS
ONE, 9(11):e109381, 11 2014.
[5] F. Kr¨
uger, K. Yordanova, C. Burghardt, and T. Kirste. Towards creating assistive
software by employing human behavior models. Journal of Ambient Intelligence
and Smart Environments, 4(3):209–226, May 2012.
[6] M. Mirmehdi, T. Kirste, S. Whitehouse, A. Paiement, and K. Yor-
danova. Sphere unscripted kitchen activities. University of Bristol, 2016.
https://data.bris.ac.uk/data/dataset/raqa2qzai45z15b4n0za94toi.
[7] S. D. Ohlhorst, R. Russell, D. Bier, D. M. Klurfeld, Z. Li, J. R. Mein, J. Milner,
A. C. Ross, P. Stover, and E. Konopka. Nutrition research to affect food and a
healthy lifespan. Advances in Nutrition: An International Review Journal, 4:579–
584, September 2013.
[8] M. Ramirez and H. Geffner. Goal recognition over pomdps: Inferring the inten-
tion of a pomdp agent. In Proceedings of IJCAI, pages 2009–2014, 2011.
[9] P. C. Roy, S. Giroux, B. Bouchard, A. Bouzouane, C. Phua, A. Tolstikov, and
J. Biswas. A Possibilistic Approach for Activity Recognition in Smart Homes for
Cognitive Assistance to Alzheimer’s Patients, pages 33–58. Atlantis Press, 2011.
[10] A. Serna, H. Pigot, and V. Rialle. Modeling the progression of alzheimer’s dis-
ease for cognitive assistance in smart homes. User Modeling and User-Adapted
Interaction, 17(4):415–438, September 2007.
[11] Gita Sukthankar, Christopher Geib, Hung Hai Bui, David Pynadath, and Robert P.
Goldman. Plan, Activity, and Intent Recognition: Theory and Practice, chapter
Introduction, pages xix–xxxv. Morgan Kaufmann Publishers Inc., San Francisco,
CA, USA, 1st edition, 2014.
[12] S. Whitehouse, K. Yordanova, A. Paiement, and M. Mirmehdi. Recognition of
unscripted kitchen activities and eating behaviour for health monitoring. In Pro-
ceedings of the 2nd IET International Conference on Technologies for Active and
Assisted Living (TechAAL 2016), London, UK, October 2016. INSPEC.
[13] K. Yordanova and T. Kirste. A process for systematic development of symbolic
models for activity recognition. ACM Transactions on Interactive Intelligent Sys-
tems, 5(4):20:1–20:35, December 2015.
[14] N. Zhu, T. Diethe, M. Camplani, L. Tao, A. Burrows, N. Twomey, D. Kaleshi,
M. Mirmehdi, P. Flach, and I. Craddock. Bridging e-health and the internet of
things: The sphere project. IEEE Intelligent Systems, 30(4):39–46, July 2015.
9
... Abbreviations as folow: CS: contact sensor, THS: thermal sensor, ACC: accelerometer, LRS: laser range sensors, IRC: infrared cam, CCDC: CCD camera, IRSMD: infrared sensors for movement detection, IMU: Inertial measurement unit, TEMP: temperature, CO2: CO2 presence, MAG -magnetic switch, PIRMS: passive infrared movement sensor, LUM: luminiscence sensor, HUM: humidity, USD: ultra sound sensor to measure distance, ELCONS: electrical consumption sensor, WCONS water consumption sensor, WUS: water usage sensor, MS: movement sensor, RS: reed sensor, WPS: water pressure sensor, PS: pressure sensor, LLS: light level sensor, NLS: noise levels sensor, DLS: dust levels sensor, LIM: limit switch,RFID Radio-Frequency Identification, RFIDT: RFID-tags, DC: RGB camera with depth information, PWD: power load sensor, VS: vision sensor, C: RGB camera, SS: smoke sensor, SWACC: smart-watch accelerometer, MACCG: mobile device accelerometer and gyroscope, MACCGMAG: mobile accelerometer, gyroscope, magnetometer, PT: paper-tags, SC: stereoscopic camera, FS: float sensor. and SPHERE [51], among others. Surprisingly, the majority of researchers relied on their own data and observations, without making them accessible to others. ...
... This reflects the recognition that activities are often deducible by identifying sequences of atomic actions and the objects involved in their execution. For example, in a study by Yordanova et al. [51], the coarse-grained action of ''preparing healthy food'' was discerned through the observation of fine-grained actions like ''drink'', ''clean'', and ''put''. Another instance is the approach by Chatterjee et al. [31], where coarse-grained actions (referred to as macro-activities) were disaggregated into finer-grained actions (micro-activities). ...
Article
Full-text available
With the global population aging, the demand for technologies facilitating independent living, especially for those with cognitive impairments, is increasing. This paper addresses this need by conducting a comprehensive evaluation of the Rostock Kitchen Task Assessment dataset, a pivotal resource in kitchen task activity recognition. Our study begins with an in-depth introduction, emphasizing the increasing prevalence of neurodegenerative disorders and the crucial role of assistive technologies. Our contributions encompass a systematic literature review, design and implementation of a working prototype of our envisioned system, refinement of the Rostock Kitchen Task Assessment dataset, creation of a semantically annotated dataset, extraction of statistical features, comparative analysis, and rigorous model performance assessment. The core of our work is the thorough evaluation and benchmarking of different activity recognition approaches using the aforementioned Rostock Kitchen Task Assessment dataset. Our experimental results demonstrate that despite encountering an imbalance problem in the dataset, the fusion of the Hidden Markov Model and Random Forest leads to superior results, achieving a weighted-averaged F 1 -score of 74.10% for all available activities and 81.40% for the most common actions in the Rostock Kitchen Task Assessment dataset. Moreover, through systematic analysis, we identify strengths and suggest potential refinements, thereby advancing the field of kitchen activity recognition. This offers valuable insights for researchers and practitioners in assistive and remote care technologies.
... Identifying users' goals and intentions based on their current and past activities is an important task in this context. While there is some work on goal recognition in the context of smart assistance systems (Yordanova, Whitehouse, Paiement, Mirmehdi, Kirste, & Craddock, 2017;Yordanova, Lüdtke, Whitehouse, Krüger, Paiement, Mirmehdi, Craddock, & Kirste, 2019;Krüger, Nyolt, Yordanova, Hein, & Kirste, 2014), so far research has mostly focused on recognizing users' current activities (Helaoui, Riboni, & Stuckenschmidt, 2013;Rashidi, Cook, Holder, & Schmitter-Edgecombe, 2010;Hoque & Stankovic, 2012;Yao, Nie, Sheng, Gu, Li, & Wang, 2016;Sztyler & Stuckenschmidt, 2017). In this paper, we address the problem of identifying user goals as a basis for automatic support. ...
... We have shown that these PRAP approaches indeed show good performance in a real-world setting, but have problems in capturing relations between observations and user goals that cannot be properly modeled manually. Some other recent approaches to goal recognition in smart environments also belong to this class of approaches (Yordanova et al., 2017(Yordanova et al., , 2019. Consequently, they have the same problems as the approaches considered in this work. ...
Preprint
Full-text available
An important feature of pervasive, intelligent assistance systems is the ability to dynamically adapt to the current needs of their users. Hence, it is critical for such systems to be able to recognize those goals and needs based on observations of the user's actions and state of the environment. In this work, we investigate the application of two state-of-the-art, planning-based plan recognition approaches in a real-world setting. So far, these approaches were only evaluated in artificial settings in combination with agents that act perfectly rational. We show that such approaches have difficulties when used to recognize the goals of human subjects, because human behaviour is typically not perfectly rational. To overcome this issue, we propose an extension to the existing approaches through a classification-based method trained on observed behaviour data. We empirically show that the proposed extension not only outperforms the purely planning-based- and purely data-driven goal recognition methods but is also able to recognize the correct goal more reliably, especially when only a small number of observations were seen. This substantially improves the usefulness of hybrid goal recognition approaches for intelligent assistance systems, as recognizing a goal early opens much more possibilities for supportive reactions of the system.
... In a previous work, we shortly presented our concept and we showed first preliminary empirical results from CSSMs applied to an unscripted scenario [14]. In this work, we extend our previous work by providing 1. detailed formal description of the proposed methodology; 2. detailed analysis on the influence of different sensors on the CSSM's performance; 3. extension of the model to allow the recognition of single pooled and multiple goals; 4. detailed comparison between state-of-the-art methods and our proposed approach. ...
... In a previous work, we showed preliminary results of a CSSM approach called Computational Causal Behaviour Models (CCBM), which reasons about one's activities in real cooking scenarios [14,38]. In this work, we extend our previous work by providing detailed information on the approach and the developed model. ...
Article
Full-text available
Wellbeing is often affected by health-related conditions. Among them are nutrition-related health conditions, which can significantly decrease the quality of life. We envision a system that monitors the kitchen activities of patients and that based on the detected eating behaviour could provide clinicians with indicators for improving a patient’s health. To be successful, such system has to reason about the person’s actions and goals. To address this problem, we introduce a symbolic behaviour recognition approach, called Computational Causal Behaviour Models (CCBM). CCBM combines symbolic representation of person’s behaviour with probabilistic inference to reason about one’s actions, the type of meal being prepared, and its potential health impact. To evaluate the approach, we use a cooking dataset of unscripted kitchen activities, which contains data from various sensors in a real kitchen. The results show that the approach is able to reason about the person’s cooking actions. It is also able to recognise the goal in terms of type of prepared meal and whether it is healthy. Furthermore, we compare CCBM to state-of-the-art approaches such as Hidden Markov Models (HMM) and decision trees (DT). The results show that our approach performs comparable to the HMM and DT when used for activity recognition. It outperformed the HMM for goal recognition of the type of meal with median accuracy of 1 compared to median accuracy of 0.12 when applying the HMM. Our approach also outperformed the HMM for recognising whether a meal is healthy with a median accuracy of 1 compared to median accuracy of 0.5 with the HMM.
... Several researchers have addressed the specific problem of recognizing cooking activities [4,13,16,23,18,19,21,22]. Pham et al. [4] propose a real-time approach to classify fine-grained cooking activities such as, e.g., peeling, slicing and dicing, using accelerometer data. ...
Chapter
In this paper, we present an automatic approach to recognize cooking activities from acceleration and motion data. We rely on a dataset that contains three-axis acceleration and motion data collected with multiple devices, including two wristbands, two smartphones and a motion capture system. The data is collected from three participants while preparing sandwich, fruit salad and cereal recipes. The participants performed several fine-grained activities while preparing each recipe such as cut and peel. We propose to use the multi-class classification approach to distinguish between cooking recipes and a multi-label classification approach to identify the fine-grained activities. Our approach achieves 81% accuracy to recognize fine-grained activities and 66% accuracy to distinguish between different recipes using leave-one-subject-out cross-validation. The multi-class and multi-label classification results are 27 and 50% points higher than the baseline. We further investigate the effect on classification performance of different strategies to cope with missing data and show that imputing missing data with an iterative approach provides 3% point increment to identify fine-grained activities. We confirm findings from the literature that extracting features from multi-sensors achieves higher performance in comparison to using single-sensor features.
... Hence, other techniques rely on body-worn or environmental sensors for activity recognition. In the context of the SPHERE project [39], Yordanova et al. propose the use of computational state space models (i.e., probabilistic models combining symbolic representation with probabilistic reasoning) for recognizing cooking activities based on data acquired from different sensors in a smart kitchen [36,37]. In particular, the authors rely on temperature, humidity, light/noise/dust levels, individual's motion, and usage of certain objects, water, and electricity. ...
Article
Full-text available
Abstract Unhealthy behaviors regarding nutrition are a global risk for health. Therefore, the healthiness of an individual’s nutrition should be monitored in the medium and long term. A powerful tool for monitoring nutrition is a food diary; i.e., a daily list of food taken by the individual, together with portion information. Unfortunately, frail people such as the elderly have a hard time filling food diaries on a continuous basis due to forgetfulness or physical issues. Existing solutions based on mobile apps also require user’s effort and are rarely used in the long term, especially by elderly people. For these reasons, in this paper we propose a novel architecture to automatically recognize the preparation of food at home in a privacy-preserving and unobtrusive way, by means of air quality data acquired from a commercial sensor. In particular, we devised statistical features to represent the trend of several air parameters, and a deep neural network for recognizing cooking activities based on those data. We collected a large corpus of annotated sensor data gathered over a period of 8 months from different individuals in different homes, and performed extensive experiments. Moreover, we developed an initial prototype of an interactive system for acquiring food information from the user when a cooking activity is detected by the neural network. To the best of our knowledge, this is the first work that adopts air quality sensor data for cooking activity recognition.
Article
With the rapid development of artificial intelligence, the Internet of Things and big data, etc., the smart home enters the Era 3.0, which profoundly relies on user-based big data mining and deep learning. At present, smart interaction, intelligent perception, and complex decision-making in the smart home urgently need new supporting technologies and methods. This paper integrated a large amount of frontier research worldwide and summarized and refined the three elements required for the future smart home: "perception" "thinking" and "execution". This paper began with an overview of the frontier platforms for smart homes in China and abroad, followed by an elaboration and review of the research on data collection, feature extraction, perception in smart home scenarios, and multimodal human-computer interaction and innovation scenarios for smart homes, respectively. The authors presented a series of related research work of Future Lab in the smart home field for researchers in related fields.
Chapter
Improvements in global wealth and well-being result in an increase in the life expectancy of the population and, consequently, determine an increase in the number of people with physical or mental impairments. This slice of the population needs daily assistance and monitoring to live a safe and productive life. For this reason, several researchers are focusing their work on health care technologies. A broad collection of sensors has been developed to monitor our health status, our daily behavior, and the environment where we live. In this chapter we present two types of non-invasive sensors used in the health care domain: air quality sensors and cameras. For each type of sensor, we describe the health care application areas, the technology used to implement them, the existing datasets, and the data collection process and issues. Moreover, we analyze a real-life health care application for both air quality sensors and cameras.KeywordsHealth careAir quality sensorsCamerasActive agingDeep learningRobotics
Article
A common way of learning to perform a task is to observe how it is carried out by experts. However, it is well known that for most tasks there is no unique way to perform them. This is especially noticeable the more complex the task is because factors such as the skill or the know-how of the expert may well affect the way she solves the task. In addition, learning from experts also suffers of having a small set of training examples generally coming from several experts (since experts are usually a limited and expensive resource), being all of them positive examples (i.e. examples that represent successful executions of the task). Traditional machine learning techniques are not useful in such scenarios, as they require extensive training data. Starting from very few executions of the task presented as activity sequences, we introduce a novel inductive approach for learning multiple models, with each one representing an alternative strategy of performing a task. By an iterative process based on generalisation and specialisation, we learn the underlying patterns that capture the different styles of performing a task exhibited by the examples. We illustrate our approach on two common activity recognition tasks: a surgical skills training task and a cooking domain. We evaluate the inferred models with respect to two metrics that measure how well the models represent the examples and capture the different forms of executing a task showed by the examples. We compare our results with the traditional process mining approach and show that a small set of meaningful examples is enough to obtain patterns that capture the different strategies that are followed to solve the tasks.
Article
Full-text available
There's a widely known need to revise current forms of healthcare provision. Of particular interest are sensing systems in the home, which have been central to several studies. This article presents an overview of this rapidly growing body of work, as well as the implications for machine learning, with an aim of uncovering the gap between the state of the art and the broad needs of healthcare services in ambient assisted living. Most approaches address specific healthcare concerns, which typically result in solutions that aren't able to support full-scale sensing and data analysis for a more generic healthcare service, but the approach in this article differs from seamlessly linking multimodel data-collecting infrastructure and data analytics together in an AAL platform. This article also outlines a multimodality sensor platform with heterogeneous network connectivity, which is under development in the sensor platform for healthcare in a residential environment (SPHERE) Interdisciplinary Research Collaboration (IRC).
Article
Full-text available
Background: Computational state space models (CSSMs) enable the knowledge-based construction of Bayesian filters for recognizing intentions and reconstructing activities of human protagonists in application domains such as smart environments, assisted living, or security. Computational, i. e., algorithmic, representations allow the construction of increasingly complex human behaviour models. However, the symbolic models used in CSSMs potentially suffer from combinatorial explosion, rendering inference intractable outside of the limited experimental settings investigated in present research. The objective of this study was to obtain data on the feasibility of CSSM-based inference in domains of realistic complexity. Methods: A typical instrumental activity of daily living was used as a trial scenario. As primary sensor modality, wearable inertial measurement units were employed. The results achievable by CSSM methods were evaluated by comparison with those obtained from established training-based methods (hidden Markov models, HMMs) using Wilcoxon signed rank tests. The influence of modeling factors on CSSM performance was analyzed via repeated measures analysis of variance. Results: The symbolic domain model was found to have more than 10(8) states, exceeding the complexity of models considered in previous research by at least three orders of magnitude. Nevertheless, if factors and procedures governing the inference process were suitably chosen, CSSMs outperformed HMMs. Specifically, inference methods used in previous studies (particle filters) were found to perform substantially inferior in comparison to a marginal filtering procedure. Conclusions: Our results suggest that the combinatorial explosion caused by rich CSSM models does not inevitably lead to intractable inference or inferior performance. This means that the potential benefits of CSSM models (knowledge-based model construction, model reusability, reduced need for training data) are available without performance penalty. However, our results also show that research on CSSMs needs to consider sufficiently complex domains in order to understand the effects of design decisions such as choice of heuristics or inference procedure on performance.
Article
Full-text available
Activity models play a critical role for activity recognition and assistance in ambient assisted living. Existing approaches to activity modeling suffer from a number of problems, e.g., cold-start, model reusability, and incompleteness. In an effort to address these problems, we introduce an ontology-based hybrid approach to activity modeling that combines domain knowledge based model specification and data-driven model learning. Central to the approach is an iterative process that begins with “seed” activity models created by ontological engineering. The “seed” models are deployed, and subsequently evolved through incremental activity discovery and model update. While our previous work has detailed ontological activity modeling and activity recognition, this paper focuses on the systematic hybrid approach and associated methods and inference rules for learning new activities and user activity profiles. The approach has been implemented in a feature-rich assistive living system. Analysis of the experiments conducted has been undertaken in an effort to test and evaluate the activity learning algorithms and associated mechanisms.
Article
Full-text available
Proper nutrition offers one of the most effective and least costly ways to decrease the burden of many diseases and their associated risk factors, including obesity. Nutrition research holds the key to increasing our understanding of the causes of obesity and its related comorbidities and thus holds promise to markedly influence global health and economies. After outreach to 75 thought leaders, the American Society for Nutrition (ASN) convened a Working Group to identify the nutrition research needs whose advancement will have the greatest projected impact on the future health and well-being of global populations. ASN's Nutrition Research Needs focus on the following high priority areas: 1) variability in individual responses to diet and foods; 2) healthy growth, development, and reproduction; 3) health maintenance; 4) medical management; 5) nutrition-related behaviors; and 6) food supply/environment. ASN hopes the Nutrition Research Needs will prompt collaboration among scientists across all disciplines to advance this challenging research agenda given the high potential for translation and impact on public health. Furthermore, ASN hopes the findings from the Nutrition Research Needs will stimulate the development and adoption of new and innovative strategies that can be applied toward the prevention and treatment of nutrition-related diseases. The multidisciplinary nature of nutrition research requires stakeholders with differing areas of expertise to collaborate on multifaceted approaches to establish the evidence-based nutrition guidance and policies that will lead to better health for the global population. In addition to the identified research needs, ASN also identified 5 tools that are critical to the advancement of the Nutrition Research Needs: 1) omics, 2) bioinformatics, 3) databases, 4) biomarkers, and 5) cost-effectiveness analysis.
Article
Full-text available
Assistive software becomes more and more important part of our everyday life. As it is not straightforward to create such a system, the engineering of assistive systems is a topic of current research with different applications in healthcare, education and industry. In this paper we introduce three contributions to this field of research. Whereas most assistive systems use approaches for intention recognition based on training data applicable to specific environments and applications, we introduce a training-free approach. We do that by showing that it is possible to generate probabilistic inference systems from causal models for human behavior. Additionally, we collect a list of requirements for context aware assistive software and human behavior modeling for intention recognition and showed that our system satisfies them. We then introduce a software architecture for assistive systems that provides support for this kind of modeling. In addition to introducing the modeling approach and the architecture we show in an experimental way that our approach is suited for smart environments. The collected list of requirements could help a software engineer create a robust and easily adaptable to changes in the environment assistive software.
Chapter
Full-text available
Providing cognitive assistance to Alzheimer’s patients in smart homes is a field of research that receives a lot of attention lately. The recognition of the patient’s behavior when he carries out some activities in a smart home is primordial in order to give adequate assistance at the opportune moment. To address this challenging issue, we present a formal activity recognition framework based on possibility theory and description logics. We present initial results from an implementation of this recognition approach in a smart home laboratory.
Conference Paper
Nutrition related health conditions can seriously decrease the quality of life, and a system able to monitor the kitchen activities and eating behaviour of patients could provide clinicians important information towards the improvement of the patient’s condition. We propose a symbolic model able to describe unscripted kitchen activities and eating behaviour of people in home settings. The model consists of an ontology that describes the problem domain and of a computational state space model that is able to reason in a probabilistic manner about the person’s actions, goals, and causes of problems during action execution. To validate our model, we recorded 15 unscripted kitchen tasks involving 9 subjects and manually annotated the video data according to the proposed ontology schema. We then compared the model’s ability to recognise people’s activities and their goals by generating simulated noisy observations from the annotation of the experiments. The results showed that our model is able to recognise kitchen activities with an average accuracy of 0.8, when using specialised models, and 0.4 when using the general model.
Technical Report
Dataset containing environmental sensor data of several people performing unscripted cooking activities in the Sphere kitchen at the University of Bristol.
Book
PLEASE DO NOT SEND REQUESTS FOR FULL TEXT: THE EDITORS DO NOT OWN THE COPYRIGHT AND CANNOT SHARE THE TEXT. Plan recognition, activity recognition, and intent recognition together combine and unify techniques from user modeling, machine vision, intelligent user interfaces, human/computer interaction, autonomous and multi-agent systems, natural language understanding, and machine learning. Plan, Activity, and Intent Recognition explains the crucial role of these techniques in a wide variety of applications including: personal agent assistants. Computer and network security. Opponent modeling in games and simulation systems. Coordination in robots and software agents. Web e-commerce and collaborative filtering. Dialog modeling. Video surveillance. Smart homes. In this book, follow the history of this research area and witness exciting new developments in the field made possible by improved sensors, increased computational power, and new application areas.
Article
The article can be downloaded from https://mmis.informatik.uni-rostock.de/index.php?title=A_Process_for_Systematic_Development_of_Symbolic_Models_for_Activity_Recognition Several emerging approaches to activity recognition (AR) combine symbolic representation of user actions with probabilistic elements for reasoning under uncertainty. These approaches provide promising results in terms of recognition performance, coping with the uncertainty of observations, and model size explosion when complex problems are modelled. But experience has shown that it is not always intuitive to model even seemingly simple problems. To date, there are no guidelines for developing such models. To address this problem, in this work we present a development process for building symbolic models that is based on experience acquired so far as well as on existing engineering and data analysis workflows. The proposed process is a first attempt at providing structured guidelines and practices for designing, modelling, and evaluating human behaviour in the form of symbolic models for AR. As an illustration of the process, a simple example from the office domain was developed. The process was evaluated in a comparative study of an intuitive process and the proposed process. The results showed a significant improvement over the intuitive process. Furthermore, the study participants reported greater ease of use and perceived effectiveness when following the proposed process. To evaluate the applicability of the process to more complex AR problems, it was applied to a problem from the kitchen domain. The results showed that following the proposed process yielded an average accuracy of 78%. The developed model outperformed state-of-the-art methods applied to the same dataset in previous work, and it performed comparably to a symbolic model developed by a model expert without following the proposed development process.