A preview of this full-text is provided by Springer Nature.
Content available from Multimedia Tools and Applications
This content is subject to copyright. Terms and conditions apply.
Vol.:(0123456789)
Multimedia Tools and Applications (2025) 84:8869–8892
https://doi.org/10.1007/s11042-024-18933-2
1 3
Deep similarity segmentation model forsensor-based
activity recognition
AbdulRahmanBaraka1· Mohd HalimMohdNoor1
Received: 24 January 2023 / Revised: 24 January 2024 / Accepted: 13 March 2024 /
Published online: 4 May 2024
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024
Abstract
Signal segmentation is a critical stage in activity recognition. Most existing studies adopted
the fixed-size sliding window method for this stage. However, the fixed-size sliding win-
dow may not produce the most effective segmentation method since human activities have
variable length durations, particularly transitional activities. In this paper, we propose a
novel deep similarity segmentation model that overcomes not only the limitations of the
fixed sliding window method but also the weaknesses of threshold-based segmentation
methods. Specifically, a novel deep learning model is designed to distinguish between tran-
sitional and basic activity by treating the segmentation task as a binary classification task.
The proposed model accepts multiple sequence windows and extracts the local features
automatically for each window using convolutional neural networks. The temporal fea-
tures of windows are extracted by measuring the similarity and differentiation between the
local features of adjacent windows. The local features are combined with the temporal fea-
tures and passed to deep fully connected layers to distinguish the transitional activity from
the basic activity windows. The evaluation relies on two public datasets, SBHARPT and
FORTH-TRACE. According to the experimental findings, the proposed approach can dis-
tinguish between basic and transitional activities with an accuracy of 98.51% and 98.41%,
respectively. Additionally, our method outperformed the fixed sliding window for activity
recognition by 2.93% and 2.24% for both datasets, respectively, achieving an accuracy of
93.35% and 84.96%. These results are significant and outperform the precision of cutting-
edge models.
Keywords Signal segmentation· Deep learning· Transitional activity
* Mohd Halim Mohd Noor
halimnoor@usm.my
AbdulRahman Baraka
abarakeh@qou.edu
1 School ofComputer Sciences, Universiti Sains Malaysia, Gelugor, PulauPinang, Malaysia
Content courtesy of Springer Nature, terms of use apply. Rights reserved.