Conference PaperPDF Available

A Reconfigurable HMM for Activity Recognition

Authors:

Abstract

This paper presents a reconfigurable HMM for activity recognition in a home setting where sensors are deployed in the environment. Sensors activated in a sequence are captured and modeled using the HMM like construct. The model traces out the most likely sensor activation sequence responsible for the activity being performed. We recognize several Activities of Daily Living (ADLs) for Ubiquitous Healthcare.
k-Means Discriminant Maps for Data Visualization and
Classification
Vo Dinh Minh Nhat
Ubiquitous Computing Lab
Kyung Hee University
Suwon, Korea
vdmnhat@uclab.khu.ac.kr
SungYoung Lee
Ubiquitous Computing Lab
Kyung Hee University
Suwon, Korea
sylee@uclab.khu.ac.kr
ABSTRACT
Over the years, many dimensionality reduction algorithms
have been proposed for learning the structure of high dimen-
sional data by linearly or non-linearly transforming it into a
low-dimensional space. Some techniques can keep the local
structure of data, while the others try to preserve the global
structure. In this paper, we propose a linear dimensionality
reduction technique that characterizes the local and global
properties of data by firstly applying k-means algorithm on
original data, and then finding the projection by simultane-
ously globally maximizing the between-cluster scatter ma-
trix and locally minimizing the within-cluster scatter ma-
trix, which actually keeps both local and global structure of
data. Low complexity and structure preserving are two main
advantages of the proposed technique. The experiments on
both artificial and real data sets show the effectiveness and
novelty of proposed algorithm in visualization and classifi-
cation tasks.
Categories and Subject Descriptors
I.5.2 [Computing Methodologies]: Pattern Recognition—
Design Methodology.
General Terms
Algorithms, Design, Experimentation, Performance, The-
ory.
Keywords
Dimensionality Reduction, k-Means, Manifold Learning, Lin-
ear Discriminant Analysis.
1. INTRODUCTION
The purpose of dimensionality reduction is to transform
high dimensional data into a low-dimensional space, while
retaining most of the underlying structure in the data. The
reason for using dimensionality reduction is based on the fact
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, to
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee.
SAC’08 March 16-20, 2008, Fortaleza, Cear´
a, Brazil
Copyright 2008 ACM 978-1-59593-753-7/08/0003 ...$5.00.
that some features may be irrelevant and ”intrinsic” dimen-
sionality of data may be smaller than the number of features.
Dimensionality reduction can also be used to visualize high
dimensional data by transforming the data into two or three
dimensions, thereby giving additional insight into the prob-
lem at hand. With the rapidly increasing demand on dimen-
sionality reduction techniques, it is not surprising to see an
overwhelming amount of research publications on this topic
in recent years. In general, there are linear and nonlinear
dimensionality reduction techniques. Linear dimensionality
reduction methods include Principal Component Analysis
(PCA) developed by Pearson (1901) and Hotelling (1933)
[6][3], and Multi-Dimensional Scaling (MDS) by Torgerson
(1952) and Shepard (1962) [10][8]. While PCA finds a low-
dimensional embedding of the data points that best pre-
serves their variance as measured in the high-dimensional
input space, MDS finds an embedding that preserves the
inter-point distances, which is equivalent to PCA when the
distances are Euclidean. Besides linear methods, there are
also non-linear dimensionality reduction techniques which
have been developed up-to-date. Kernel PCA (KPCA) [4]
map inputs nonlinearly to a new space, then perform PCA.
Laplacian Eigenmaps (LE) [1] preserve nearness relations
as encoded by graph Laplacian. ISOMAP [9] assumes that
the data lie on a (Riemannian) manifold and maps data to
its low-dimensional representation in such a way that the
geodesic distance between two date points is as close to the
Euclidean distance between two respectively points in low-
dimensional space as possible. Diffusion Maps (DM) [5] is
based on defining a Markov random walk on the graph of
the data. In the low-dimensional representation of the data,
the pairwise diffusion distances are retained as well as pos-
sible. Locally Linear Embedding (LLE) [7] maps its inputs
into a single global coordinate system of lower dimensional-
ity by computing low-dimensional, neighborhood preserving
embedding of high-dimensional inputs, and its optimization
does not involve local minima. It actually recovers global
nonlinear structure from locally linear fits. Due to the length
of paper some other techniques most of which are variants of
above reviewed techniques are not covered here. In this pa-
per, we propose a linear dimensionality reduction technique
called k-Means Discriminant Maps (kDM). The algorithm
firstly applies k-Means to cluster the original data, then for
the purpose of keeping both local and global structure of
data, it try to find a desirable projection that simultane-
ously minimizes the within-cluster scatter and maximizes
the between-cluster scatter matrices. Some main contribu-
tions of proposed algorithm can be described as: low com-
plexity due to its linear property, and keeping both local
and global structure of data. The outline of this paper is as
follows. The proposed method is described in Section 2. In
Section 3, experimental results are performed on both arti-
ficial and real data sets to demonstrate the effectiveness of
our method. Finally, conclusions are presented in Section 4.
2. K-MEANS DISCRIMINANT MAPS
The dimension reduction problem is, given a data set
{x1, x2, ..., xN}where xi∈ <n, to find a set of points
{y1, y2, ..., yN}where yi∈ <mand (m¿n), such that
each yi“represents” its counterpart xi. For the convenience
of presentation, we denote the matrix X= [x1, x2, ..., xN]
and correspondingly the matrix Y= [y1, y2, ..., yN]. In this
section, our emphasis is on the description of our proposed
algorithms. Due to the paper length, those previous dimen-
sionality reduction techniques can be referenced in respec-
tive literatures.
2.1 k-Means
The objective it tries to achieve is to minimize total intra-
cluster variance, or, the squared error function
f=
k
X
i=1
X
xjΠi
kxjµik2(1)
where there are kclusters µi,i= 1,2, ..., k and Πiis the
centroid or mean point of all the points xjΠi.
2.2 k-Means Discriminant Maps - kDM
Assume that after clustering data by k-means algorithm
each data sample belongs to one of Ccluster {Π1,Π2, ..., ΠC}.
Let Nibe the number of the samples in cluster Πi(i=
1,2, ..., C), µi=1
NiP
xΠi
xbe the mean of the samples or the
centroid in cluster Πi. Then we define the between-cluster
scatter matrix Sband the within-cluster scatter matrix Sw
as follow
Sb=1
N
C
X
i=1
Ni(µiµ)(µiµ)T(2)
Sw=1
N
C
X
i=1
X
xkΠi
(xkµi)(xkµi)T(3)
For the purpose of keeping both local and global structure
of data, we try is to find a projection which will draw the
close samples (ones in the same cluster) closer together while
simultaneously making the distant samples (ones from dif-
ferent clusters) even more distant from each other. From
this point of view, a desirable projection should be the one
that, at the same time, minimizes the within-cluster scatter
and maximizes the between-cluster scatter matrices. So in
kDM, the projection Wopt is chosen to maximize the ratio of
the determinant of the between-cluster scatter matrix of the
projected samples to the determinant of the within-cluster
scatter matrix of the projected samples, i.e.,
J(w) = arg max
w
wTSbw
wTSww(4)
From the criterion in ( 4), we can find the projection by si-
multaneously globally maximizing the between-cluster scat-
ter and locally minimizing the within-cluster scatter, which
Technique Parameter Settings
PCA None
Kernel PCA κ= (XX T+ 1)3
Diffusion Maps σ= 1
LLE k= 12
LE k= 12, σ = 1
ISOMAP k= 12
kDM k= 3
Table 1: Parameter settings for the experiments
actually keep both local and global structure of data. It
is also easy to realize that the criterion ( 4) is formally
similar to the Fisher criterion since they are both Rayleigh
quotients. However in kDM, we form the between-cluster
scatter matrix Sband the within-cluster scatter matrix Sw
without knowing the class labels of samples. This means
Fisher discriminant projection is supervised, while the pro-
jection determined by kDM can be obtained in an unsu-
pervised manner. The optimal projection for kDM is W=
[w1w2...wm], where {wi|i= 1,2, ..., m}is the set of gener-
alized eigenvectors of Sband Swcorresponding to the m
largest generalized eigenvalues {λi|i= 1,2, ..., m}, i.e.,
Sbwi=λiSwwi
S1
wSbwi=λiwii= 1,2, ..., m (5)
However in some cases, the dimension of the sample space
is typically larger than the number of samples. As a conse-
quence, Swis singular. This problem is known as the “small
sample size (3S) problem”[2]. To solve this problem, we use
the strategy of Direct-LDA[11] to implement our kDM al-
gorithm in the case of 3S problem. The key idea of DLDA
is to discard the null space of Sb, which contains no useful
information, rather than discarding the null space of Sw,
which contains the most discriminative information. Sbis
firstly diagonalized as Sb=UΛUT, where U∈ <n×(C1)
is a matrix whose columns are eigenvectors of Sb,Cis the
number of classes (equivalent to the number of clusters kin
k-Means of kDM, we use kand Cinterchangeably) and Λ
is a diagonal matrix with eigenvalues. The new projected
within scatter matrix is formed as
˜
Sw= Λ1/2UTSwUΛ1/2(6)
Let ˜
Ww= [w1w2...wC1], where {wi|i= 1,2, ..., C 1}is
the set of eigenvectors of ˜
Sw. Then, the optimal projection
for DLDA is Wopt =UΛ1/2˜
Ww.
3. EXPERIMENTS
In this section, a systematic empirical experiments of the
performance of previous techniques and our proposed tech-
nique kDM are performed. We perform the evaluation on
two types of datasets: (1) artificial datasets and (2) real
datasets (ORL face database and PolyU Palmprint database).
3.1 Data Visualization on Artificial Datasets
The artificial datasets on which we performed experiments
are: (1) the Swiss roll dataset and (2) the intersecting dataset.
Some parameters used in this part of experiments can be
seen in Table 1. We perform PCA, Kernel PCA, Diffusion
Maps, LLE, LE, ISOMAP and kDM on 1000 data points of
−15
−10
−505
10
15 −15
−10
−5 0510 15
0
10
20
30
Swiss Roll
0 10 20
−15
−10
−5
0
5
10
15 PCA
t = 0.23438s −1 −0.5 0 0.5
−0.6
−0.4
−0.2
0
0.2
0.4
0.6
Kernel PCA
t = 39.4844s −4 −2 0 2 4
x 10−7
−3
−2
−1
0
1
2
3
4x 10−7
Diffusion Maps
t = 1.2188s
−2 0 2
−2
−1.5
−1
−0.5
0
0.5
1
1.5 LLE
t = 2.375s −0.05 0 0.05
−0.04
−0.02
0
0.02
0.04
LE
t = 1.1719s −60 −40 −20 0 20 40
−15
−10
−5
0
5
10
15 ISOMAP
t = 104.3906s −10 0 10
−15
−10
−5
0
5
10
15 kDM
t = 0.15625s
Figure 1: Two-dimensionality Visualization of the Swiss roll dataset based on variety of techniques
−15
−10
−5
0
5
10
15 −15 −10 −5 0510 15
0
30
60
Swiss Roll
−20 −10 0 10 20
−15
−10
−5
0
5
10
k = 5 −20 −10 0 10 20
−15
−10
−5
0
5
10
15
20
k = 10
−20 −10 0 10 20
−15
−10
−5
0
5
10
15
k = 15 −20 −10 0 10 20
−15
−10
−5
0
5
10
15
k = 20 −20 −10 0 10 20
−10
−5
0
5
10
15
20
k = 25
Figure 2: Performance of kDM versus k= 5,10,15,20
on Swiss roll dataset.
−1.5−1
−0.5 0
0.5 1
1.5 −0.8
−0.6
−0.4
−0.2
00.20.40.6
−2
0
2
4
6
Intersect
−1.5 −1 −0.5 0 0.5 1
−0.8
−0.6
−0.4
−0.2
0
0.2
0.4
k = 5 −1 −0.5 0 0.5 1
−0.8
−0.6
−0.4
−0.2
0
0.2
0.4
k = 10
−1.5 −1 −0.5 0 0.5 1
−0.8
−0.6
−0.4
−0.2
0
0.2
0.4
k = 15 −1 0 1
−0.5
0
0.5
k = 20 −1.5 −1 −0.5 0 0.5 1
−0.5
0
0.5
k = 25
Figure 3: Performance of kDM versus k= 5,10,15,20
on Intersection dataset.
−5000 −4000 −3000
−1000
−500
0
500
1000 PCA (88%)
t = 0.6875s −0.5 0 0.5 1
−0.4
−0.2
0
0.2
0.4
0.6
0.8
Kernel PCA (15%)
t = 0.09375s −4 −2 0 2 4 6
x 10−4
−3
−2
−1
0
1
2
3
x 10−4
Diffusion Maps (91%)
t = 0.25s
−2 0 2
−3
−2
−1
0
1
2LLE (91%)
t = 0.28125s −0.2 0 0.2
−0.2
−0.1
0
0.1
0.2
LE (90%)
t = 0.10938s −4000−2000 0 2000 4000
−3000
−2000
−1000
0
1000
2000
3000 ISOMAP (86%)
t = 0.35938s −3000 −2000 −1000
−1000
−500
0
500
kDM (93%)
t = 0.0825s
Sample Image
Figure 4: Two-dimensionality Visualization of the ORL face database based on variety of techniques
5000 6000 7000
−500
0
500
PCA (87%)
t = 0.40625s −0.5 0 0.5 1
−0.2
0
0.2
Kernel PCA (13%)
t = 0.10938s
−2 0 2 4
x 10−3
−2
0
2
4
x 10−3
Diffusion Maps (70%)
t = 0.29688s
−2 0 2 4
−2
−1
0
1
LLE (95%)
t = 0.29688s −0.2 0 0.2
−0.2
−0.1
0
0.1
0.2 LE (79%)
t = 0.078125s −2000 0 2000 4000
−1000
0
1000
2000 ISOMAP (96%)
t = 0.39063s −1000 −500 0 500
−400
−200
0
200
400 kDM (97%)
t = 0.1125s
Sample Palm
Figure 5: Two-dimensionality Visualization of the PolyU Palmprint database based on variety of techniques
Swiss roll dataset to show the two-dimensional representa-
tions of the Swiss roll dataset which can be seen in Fig. 1.
From the depicted representations, we can see that PCA,
Kernel PCA and Diffusion Maps techniques are not capa-
ble of successfully learning the 2-dimensional structure of
the Swiss roll manifold. While LLE and Laplacian Eigen-
maps are capable of learning the local structure of the man-
ifold, ISOMAP can learn the global structure of data. Also,
from the graph we can see advantages of new dimensionality
reduction technique kDM that it can learn both local and
global structure of Swiss roll dataset, i.e. “close” data points
will be retained “close” and ”far” data points will lie “far” in
the embedding coordinates. Since kDM is a linear method,
the running time is quite very low compared to other non-
linear techniques. We next vary the value of cluster number
k= 5,10,15,25 in kDM algorithm to see how kDM works
(see Fig. 2, 3). It seems to us that , to some extent, kDM
does not systematically depends on the number of clusters
k, which actually is an issue under our investigation.
3.2 Experiment on Biometrics Databases
In this section, we do some experiment on real biometrics
databases which are ORL face database and PolyU Palm-
print database. We choose the value k= 5 in kDM algo-
rithm to perform on both databases, while the parameters
for the other methods are still same as in Table 1. Due
to high dimensionality of biometrics data (3S problem), in
this section the kDM algorithm is implemented based on
the strategy of DLDA as discussed in previous section. In
ORL face database, we randomly select 10 subjects, each of
which contains 10 sample images. All the images were taken
against a dark homogeneous background with the subjects
in an upright, frontal position (with tolerance for some side
movement) and are manually cropped and resized to 50x50
pixel images. Two-dimensionality visualization of the ORL
face database based on variety of techniques are presented in
Fig. 4. It should be noted that the classification error rates
are calculated and put on the title of each subplot in Fig.
4. We can see that kDM give the best accuracy rate (93%),
LLE and Diffusion Maps are second best (91%), while Kernel
PCA gives a very bad performance. The PolyU Palmprint
Database[12] contains 7752 grayscale images corresponding
to 386 different palms. We also select randomly 10 sub jects
and 10 palms for each subject to do the experiment. We
use inscribed circle-based segmentation approach in [13] to
extract palms and resize each palm to radius of 25 pixels.
In the case of palmprint database, from Fig. 5, kDM still
give good performance in terms of both data visualization
and classification with 97% accuracy.
3.3 Discussion
The experiments on both artificial and real biometrics
datasets have been systematically performed. These exper-
iments reveal a number of interesting points as follow:
kDM can be a good candidate for data visualization
because it can learn the whole structure (both local
and global structure) of data.
Though kDM is a unsupervised technique, it still have
the ability of finding discriminative features which is
very helpful in classification taks.
It is quite easy to implement and run fast compared
to the other non-linear techniques.
4. CONCLUSIONS
In this paper, we propose a linear dimensionality reduc-
tion technique that can keep both local and global struc-
ture of data. The experiments on both artificial and real
datasets show its potential in data visualization and clas-
sification tasks. The corner-stone of the idea is the usage
of nice properties from k-means and Fisher criteria. In the
first step of applying k-means, those “close” data samples
will be tendentiously kept in the same cluster, those ”dis-
tant” data samples will be grouped into different clusters.
And this topology of data will be preserved by using Fisher
criteria to embed data into low-dimensional representation.
A future work is obviously the effect of k-means on the kDM
algorithm.
5. ACKNOWLEDGMENTS
This research was supported by the MIC(Ministry of In-
formation and Communication), Korea, Under the ITFSIP
(IT Foreign Specialist Inviting Program) Supervised by the
IITA (Institute of Information Technology Advancement).
6. REFERENCES
[1] M. Belkin and P. Niyogi. Laplacian eigenmaps and
spectral techniques for embedding and clustering.
Advances in Neural Information Processing Systems
14.
[2] K. Fukunaga. Introduction to statistical pattern
recognition. Academic Press Professional, Inc., San
Diego, CA, USA, 1990.
[3] H. Hotelling. Analysis of a complex of statistical
variables into principal components. J. Educational
Psychology, 27:417–441, 1933.
[4] S. Mika, B. Sch¨olkopf, A. J. Smola, K.-R. M¨uller,
M. Scholz, and G. R¨atsch. Kernel pca and de-noising
in feature spaces. pages 536–542, 1998.
[5] B. Nadler, S. Lafon, R. R. Coifman, and I. G.
Kevrekidis. Diffusion maps, spectral clustering and
reaction coordinates of dynamical systems. Applied
and Computational Harmonic Analysis, 21.
[6] K. Pearson. On lines and planes of closest fit to
systems of points in space. Philoshophical Magazine,
2:559–572, 1901.
[7] S. T. Roweis and L. K. Saul. Nonlinear dimensionality
reduction by locally linear embedding. Science,
290(5500):2323–2326, 2000.
[8] R. N. Shepard. The analysis of proximities:
Multidimensional scaling with an unknown distance
function. Psychometrika, 27:125–140, 1962.
[9] J. Tenenbaum, V. de Silva, and J. Langford. A global
geometric framework for nonlinear dimensionality
reduction. 290(5500):2319–2323, December 2000.
[10] W. S. Torgerson. Multidimensional scaling.
Psychometrika, 17:401–419, 1952.
[11] H. Yu and J. Yang. A direct lda algorithm for
high-dimensional data - with application to face
recognition. Pattern Recognition, 34(10):2067–2070,
2001.
[12] D. Zhang. Polyu palmprint palmprint database -
http://www.comp.polyu.edu.hk/ biometrics/.
[13] D. Zhang. Palmprint Authentication. Kluwer
Academic, 2004.
... Wang at al. [40] designed a trac padding algorithm that uses matched package schedules to prevent adversaries from paring incoming and outgoing trac ows. Signicant work [14,[29][30][31] proposed to model user in-home activities using Markov Chain-based approaches. However, due to the empirical modeling of IoT device events and the nature of the random injecting of IoT trac signatures, these approaches may still allow sophisticated attackers to identify the randomly injected "fake" signatures, and thus infer the genuine user private information. ...
... The Autominder System developed by Pollack et al. [25] uses dynamic Bayesian networks as an underlying domain model to coordinate prompts delivering through activities. Capturing ADL through sensors is not new [1], [2], [14], [12], [7], [8], [13], [3], and when combined with AI, Barucha showed that sensing capabilities improve the reability and viability of assistive systems [20]. Nevertheless, there has been less attention devoted to the definition and recognition of subtasks in these activities. ...
Article
Stroke survivors often have difficulties performing Activities of Daily Living (ADL). When trying to complete a task, they tend to rely on caregivers who give them cues when necessary. However, this reliance on caregivers' support may affect their ability to live independently. In our study, we have developed CogWatch - an assistive system designed to provide guidance to stroke survivors during tea-making. An evaluation of the system was carried out on twelve patients. The latter were asked to complete different ADLs with and without CogWatch's assistance. Results showed that patients succeed at the tasks more often when assisted by the system than without guidance. It is anticipated that this system could be installed in the home environment and provide early stage rehabilitation.
... A large amount of the research has been conducted in the area of activity recognition in smart environments. A series of Bayesian derivative models are popularly applied, including naï ve Bayes [6] and Bayesian networks [7] with the strength in encoding causal (dependence) relationships, and Dynamic Bayesian Networks [8], Hidden Markov Models [9] and Conditional Random Fields [10] with the strength in encoding temporal relationships. ...
Conference Paper
Full-text available
Opportunistic Situation Identification (OSI) is new paradigms for situation-aware systems, in which contexts for situation identification are sensed through sensors that happen to be available rather than pre-deployed and application-specific ones. OSI extends the application usage scale and reduces system costs. However, designing and implementing OSI module of situation-aware systems encounters several challenges, including the uncertainty of context availability, vulnerable network connectivity and privacy threat. This paper proposes a novel middleware framework to tackle such challenges, and its intuition is that it facilitates performing the situation reasoning locally on a smartphone without needing to rely on the cloud, thus reducing the dependency on the network and being more privacy-preserving. To realize such intuitions, we propose a hybrid learning approach to maximize the reasoning accuracy using limited phone's storage space, with the combination of two the-state-the-art techniques. Specifically, this paper provides a genetic algorithm based optimization approach to determine which pre-computed models will be selected for storage under the storage constraints. Validation of the approach based on an open dataset indicates that the proposed approach achieves higher accuracy with comparatively small storage cost. Further, the proposed utility function for model selection performs better than three baseline utility functions.
... ADL can be captured through sensors on the patients or their environment, or the objects that they interact with [7][8][9][10][11][12][13][14], but decomposition of an ADL into sub-goals and recognition of these sub-goals has received less attention. The use of sensorised objects promotes an " object-centric " view of action recognition, in which a sub-goal is characterized in terms of how it is " experienced " by the objects involved. ...
Conference Paper
Assisting patients to perform activity of daily living (ADLs) is a challenging task for both human and machine. Hence, developing a computer-based rehabilitation system to re-train patients to carry out daily activities is an essential step towards facilitating rehabilitation of stroke patients with apraxia and action disorganization syndrome (AADS). This paper presents a real-time hidden Markov model (HMM) based human activity recognizer, and proposes a technique to reduce the time-delay occurred during the decoding stage. Results are reported for complete tea-making trials. In this study, the input features are recorded using sensors attached to the objects involved in the tea-making task, plus hand coordinate data captured using KinectTM sensor. A coaster of sensors, comprising an accelerometer and three force-sensitive resistors, are packaged in a unit which can be easily attached to the base of an object. A parallel asynchronous set of detectors, each responsible for the detection of one sub-goal in the tea-making task, are used to address challenges arising from overlaps between human actions. The proposed activity recognition system with the modified HMM topology provides a practical solution to the action recognition problem and reduces the time-delay by 64% with no loss in accuracy.
... Capturing an ADL through sensors [1], [10], [12], [13], [16], [18], [23], [8], or using HMMs to recognize human activity [17] are certainly not new, but decomposition of an ADL into sub-goals and recognition of these sub-goals has received less attention. ...
... Researchers have also used HMM variations such as Coupled HMM [7], Fuzzy HMM [8], Hierarchical HMM [9] and Reconfigurable HMM [10]. Somboon Hongeng, et al. [11] used several layers of naive Bayesian classifiers to gather properties of mobile objects from a sequence of image frames. ...
Article
Full-text available
Pervasive computing envisions a scenario where users are surrounded by hundreds of computers in the form of day-today devices and gadgets and using them with almost no effort on their part. To achieve this vision, a pervasive computing system has to be proactive and self-tuning. To be proactive and self-tuning, the system must be aware of the user's context. The most important aspect of a user's context is the activities being pursued by the user. So activity recognition – the ability to recognize user's activities – is essential for pervasive computing systems. To be able to recognize user activities, the system needs to be aware of how an activity is done by the user. For this, developing methods of representing activities in computer usable format is utmost important. Many structural and probabilistic methods have been developed by researchers. This paper presents a simple method using which the structure and other important details of user activities can be captured from sensor outputs in a smart environment. The captured details can then be used to recognize user activities. The proposed method was tested using a publicly available data set and is found to produce promising results.
Chapter
This chapter describes some essential elements of context-aware applications including context acquisition, context modeling, context reasoning, context distribution, and context adaptation. Context acquisition involves gathering context from the users and the environment. It has to engage with various kinds of sensors. Context modeling which is also called context representation is needed because of the necessary to have the common understanding between the system and its components. It also has to deal with the relationships and dependencies of different types of contexts. Context reasoning involves creating new knowledge and deducing better understanding based on the available context. Context distribution is to deliver the appropriate information or the services to the users. It plays the leading roles, especially for the large-scale network system. Finally, context adaptation involves the adjustment of application’s behavior so that the appropriate response can be obtained.
Conference Paper
This paper describes an approach to real-time human activity recognition using hidden Markov models (HMMs) and sensorised objects, and its application to rehabilitation of stroke patients with apraxia or action disorganisation syndrome (AADS). Results are presented for the task of making a cup of tea. Unlike speech or other sequential decoding problems where HMMs have previously been successfully applied, human actions can occur simultaneously or at least in overlapping time. The solution proposed in this paper is based on a parallel, asynchronous set of detectors, each responsible for the detection of one of the component sub-goals of the tea-making task. The inputs to these detectors are formed from the outputs of sensors attached to the objects involved in that sub-goal, plus hand coordinate data. The sensors, comprising an accelerometer and three force-sensitive resistors, are packaged in a coaster which can be easily attached to the base of a mug or jug. In tests on complete tea-making trials, error rates range from less than 5% for sub-goals where all of the objects involved are sensorised, to up to 30% for detectors that rely on hand-coordinate data alone. The complete set of detectors runs in real-time. It is concluded that a set of parallel HMM-based sub-goal detectors combined with fully sensorised objects, is a viable, accurate and easily deployable approach to real-time object-centred human activity recognition.
Conference Paper
An increasingly popular technique of monitoring activities within a smart environment involves the use of sensor technologies. With such an approach complex constructs of data are generated which subsequently require the use of activity recognition techniques to infer the underlying activity. The assignment of sensor data to one from a possible set of predefined activities can essentially be considered as a classification task. In this study, we propose the application of a cluster-based classifier ensemble method to the activity recognition problem, as an alternative to single classification models. Experimental evaluation has been conducted on publicly available sensor data collected over a period of 26 days from a single person apartment. Two types of sensor data representation have been considered, namely numeric and binary. The results show that the ensemble method performs with accuracies of 94.2% and 97.5% for numeric and binary data, respectively. These results outperformed a range of single classifiers.
Article
Sensor networks have quickly risen in importance over the last several years to become an active field of research, full of difficult problems and applications. At the same time, graphical models have shown themselves to be an extremely useful formalism for describing the underlying statistical structure of problems for sensor networks. In part, this is due to a number of efficient methods for solving inference problems defined on graphical models, but even more important is the fact that many of these methods (such as belief propagation) can be interpreted as a set of message passing operations, for which it is not difficult to describe a simple, distributed architecture in which each sensor performs local processing and fusion of information, and passes messages locally among neighboring sensors. At the same time, many of the tasks which are most important in sensor networks are characterized by such features as complex uncertainty and nonlinear observation processes. Particle filtering is one common technique for dealing with inference under these conditions in certain types of sequential problems, such as tracking of mobile objects.
Conference Paper
In this work, a system for recognizing activities in the home setting using a set of small and simple state-change sensors is introduced. The sensors are designed to be “tape on and forget” devices that can be quickly and ubiquitously installed in home environments. The proposed sensing system presents an alternative to sensors that are sometimes perceived as invasive, such as cameras and microphones. Unlike prior work, the system has been deployed in multiple residential environments with non-researcher occupants. Preliminary results on a small dataset show that it is possible to recognize activities of interest to medical professionals such as toileting, bathing, and grooming with detection accuracies ranging from 25% to 89% depending on the evaluation criteria used.
Article
A central problem in data analysis is the low dimensional representation of high dimensional data and the concise description of its underlying geometry and density. In the analysis of large scale simulations of complex dynamical systems, where the notion of time evolution comes into play, important problems are the identification of slow variables and dynamically meaningful reaction coordinates that capture the long time evolution of the system. In this paper we provide a unifying view of these apparently different tasks, by considering a family of diffusion maps, defined as the embedding of complex (high dimensional) data onto a low dimensional Euclidean space, via the eigenvectors of suitably defined random walks defined on the given datasets. Assuming that the data is randomly sampled from an underlying general probability distribution p(x)=e−U(x), we show that as the number of samples goes to infinity, the eigenvectors of each diffusion map converge to the eigenfunctions of a corresponding differential operator defined on the support of the probability distribution. Different normalizations of the Markov chain on the graph lead to different limiting differential operators. Specifically, the normalized graph Laplacian leads to a backward Fokker–Planck operator with an underlying potential of 2U(x), best suited for spectral clustering. A different anisotropic normalization of the random walk leads to the backward Fokker–Planck operator with the potential U(x), best suited for the analysis of the long time asymptotics of high dimensional stochastic systems governed by a stochastic differential equation with the same potential U(x). Finally, yet another normalization leads to the eigenfunctions of the Laplace–Beltrami (heat) operator on the manifold in which the data resides, best suited for the analysis of the geometry of the dataset regardless of its possibly non-uniform density.
Conference Paper
Ubiquitous computing researchers are increasingly turning to sensor- enabled "living laboratories" for the study of people and technologies in settings more natural than a typical laboratory. We describe the design and operation of the PlaceLab, a new live-in laboratory for the study of ubiquitous technologies in home settings. Volunteer research participants individually live in the PlaceLab for days or weeks at a time, treating it as a temporary home. Meanwhile, sensing devices integrated into the fabric of the architecture record a detailed description of their activities. The facility generates sensor and observational datasets that can be used for research in ubiquitous computing and other fields where domestic con- texts impact behavior. We describe some of our experiences constructing and op- erating the living laboratory, and we detail a recently generated sample dataset, available online to researchers.
Conference Paper
Sensor networks hold the promise of truly intelligent buildings: buildings that adapt to the behavior of their occupants to improve productivity, efficiency, safety, and security. To be practical, such a network must be economical to manufacture, install and maintain. Similarly, the methodology must be efficient and must scale well to very large spaces. Finally, be be widely acceptable, it must be inherently privacy-sensitive. We propose to address these requirements by employing networks of passive infrared (PIR) motion detectors. PIR sensors are inexpensive, reliable, and require very little bandwidth. They also protect privacy since they are neither capable of directly identifying individuals nor of capturing identifiable imagery or audio. However, with an appropriate analysis methodology, we show that they are capable of providing useful contextual information. The methodology we propose supports scalability by adopting a hierarchical framework that splits computation into localized, distributed tasks. To support our methodology we provide theoretical justification for the method that grounds it in the action recognition literature. We also present quantitative results on a dataset that we have recorded from a 400 square meter wing of our laboratory. Specifically, we report quantitative results that show better than 90% recognition performance for low-level activities such as walking, loitering, and turning. We also present experimental results for mid-level activities such as visiting and meeting.