Science topics: Data MiningActivity Recognition
Science topic
Activity Recognition - Science topic
Activity recognition aims to recognize the actions and goals of one or more agents from a series of observations on the agents' actions and the environmental conditions.
Questions related to Activity Recognition
The project I'm currently working on aims to create a deep learning model for Human Activity Recognition. I'm focusing on system design and implementation. Could someone please help me by sharing some papers or document links to better understand system design and implementation?
Thank you in advance for your assistance.
PKU MMD and NTU RGBD are large skeleton datasets widely used in human action recognition. There are plenty of codes available to process the NTU RGBD dataset. But i cannot find any code available to process PKU MMD data set and convert it to the same format as in NTU RGBD. if anybody knows the preprocessing steps and code to preprocess PKU MMD, please share it.
Thank you.
hi community;
I created a model for HAR (human activity recognition ) , and I executed it on this dataset
and now I want to try it on another dataset; with a similar format
any helps?
- HAR dataset - human activity recognition I have used.
- What is the impact of activity recognition if the dataset is having small number of features?
- I have worked on HAR and HAPT UCI datasets for human activity recognition which has 561 features.
- But with a huge dataset of low dimensional space (2 or 3 dimension), can we do activity recognition and what will be its impact?
This was first published almost 10 years ago.
Conference Paper Should We Treat Workplace Inactivity like Occupational Hazar...
Now almost a decade later, has there been any progress? Chronic workplace inactivity has been a pandemic in developed societies for much longer than a decade. The healthcare and productivity costs of workplace inactivity are all increasingly well documented. Unfortunately, this sentence from 2012 probably still applies: "Employers often provide break time and specific areas for smoking, yet to do this for exercise may be considered distracting, counterproductive, and/or too expensive." .
Thank you for considering this discussion.
I'm working with an accelerometer-based dataset capturing accelerations from the human thigh. Most HAR processes will have window segmentation to break down the signal into samples, and pre-processing of the signals (i.e filtering) prior to feature extraction. By doing both processes on the same signal from my dataset and subtracting the difference, I've found that the resulting output signal is different depending on the order these two operations are carried out.
Which of these processes should be carried out first? In the literature, I have seen several instances of either approach being taken.
Are there Simulation tools for User Activity recognition in IoT Environment?
I am interested in simulation tools based on In IoT and Non-IoT environment.
I am looking for Dataset regarding PARM.. someone can share links or Email me on
People with Dementia (PwD) have difficulty living their daily lives. And to help PwD, the caregiver is one of several solutions. However, caregivers also have many challenges in helping PwD. Because the memory and thinking decreased dramatically, PwD usually has many symptoms such as Agitation and Anxiety, repetitive questions, depression, hallucinations, sleep disturbances, etc. which make PwD refuse to be helped by caregivers. Therefore, approaches or methods that can help caregivers are needed so that their efforts to support PwD are successful. I have read "humanitude" which is one of the most successful methods. But are there other methods you might know about? Please share. Thank you.
I am working on activity recognition using wearable sensor data. Actually, I am confused to correctly specify the window size for my activities. Here, I am considering a sliding window technique for my work.
The accurate window size plays a vital role in the detection of activity; it affects the features, and whenever any features get affected, it directly hinders the performance of a classifier. I am working on four activities (Ac1, Ac2, Ac3, and Ac4), which are totally different in nature. In the AC1, the average person’s time is at least12 s, the maximum being 20 s, to complete one cycle of AC1. On the other hand, AC2 and AC3 activities are not regular activities compared to AC1. User lasts for 4 to 6 seconds to complete one circle of these two activities. In the Ac4, the average person time is 10 second to complete the activity.
So, my question is what should be my window size for this kind of activities to correctly process?
A reply would be greatly appreciated.
I am currently investigating means to assess human interaction interest on a mobile robotic platform before approaching said humans.
I already found several sources which mainly focus on movement trajectories and body poses.
Do you know of any other observable features which could be used for this?
Could you point me to relevant literature in this field?
Thanks in advance,
Martin
I have collected the data from different exercise activities by using accelerometer sensor through MATLAB mobile sensor support application. I faced a problem when I process the data, which as follows:
- Sensor recording are not synchronized and the time range is not accurate either. Sample rate of sensor is not exactly 50Hz and not consistent.
To make all the sensor data synchronized at the exact same sample rate, we interpolated and re-sampled the data.
My Question is:
Is there any side-effect of this interpolation and re-sampling on data?
I would like to know the standard technique to give a performance value for a person affected with ASD as to compare to the normal.
Hello,
I will need to implement Transport Mode Detection (TMD) on a smartphone accelerometers, gps, etc. in order to detect if the user is traveling on foot, on bicycle, by car, by bus, etc. Here are my two questions:
1) Is there some public data available to train and benchmark algorithms for this task?
2) Is there some commercially usable libraries/services implementing TMD? Something under MIT licence would be great but commercial solutions could also work.
Thanks for your time,
Bruno
NB: Something like
Hi, I am doing human activity recognition. In my task, feature scaling gives low accuracy, rather than keeping original feature values. But my feature values are not in either [-1 1] or [0 1]. So, why do I get low accuracy after feature scaling?
I am choosing between RNN and CNN to train an AI model, for a Video Images Human Activities Recognition System. Which of those (RNN, CNN) should be used?
I am a research student working in the field of human action recognition. I need help or guidance in extracting skeleton joints and plotting them on images of following dataset
- I need your help in extracting and plotting joints point (i.e. skeleton_pos.txt) in depth or RGB images of SBU Kinect Interaction Dataset.
- Plotting skeleton joints using text files (a01_s01_e01_realworld & a01_s01_e01_screen) Mendeley Data - KARD - Kinect Activity Recognition Dataset.
- I have used Skeleton visualization codes for Cornell Activity Datasets: CAD-60 & CAD-120
Hi, I am trying to do Activity Recognition with a labelled dataset containing data coming from an accelerometer, 30 binary sensors and proximity beacons.
A row example from the dataset would be:
x, y, z, s1, s2, s3, s4, s5, b1, b2
where x,y,x are continuous values coming from the accelerometer, s1,..,s5 are sensor with values 1 or 0 and b1, b2 are proximity beacons represented by their rssi value.
My biggest question is: how to use all this data together?
I tried:
- a cnn using only x, y, z
- a cnn using the sensors data
But I was wondering if it was possible to do something more complex considering the different sources of data.
Can anyone point me to an algorithm or a model that can detect body movement from the accelerometer data on a wristband.
I use a dataset about activities that an old person was doing during a year. it has features of start time, end time and activity name as below:
08:52:12 - 08:55:38 - Washing hand/face
08:57:36 - 09:05:53 - Make coffee
09:07:38 - 09:12:52 - Washing hand/face
09:13:57 - 09:21:10 - Make sandwich
09:23:08 - 09:43:11 - Eating
..
I want to insert abnormal situations in which an activity lasts longer than usual or increase frequencies of doing an activity during a day.
i'm programming in python. what should i do?
if i want to insert an abnormal record, should i change the time of all of records that came after that record?
The frame activity is used for many applications to separate signals, detection and so on.
UCI Human Activity Recognition (HAR) Data set is easily available on internet as well as on kaggle if someone had worked on it then do let know.
I am trying to use the HAR dataset (see attached link) to test my activity recognition algorithm. However, before using it to test my own method, I am trying to make sense of the data, and understand the features that are extracted.
As a simple test I tried to reproduce them, starting with the mean. I calculated the arithmetic mean of the first row of values in the data/train/Inertial Signals/body_acc_x_train.txt file. If I understand the explanation correctly, this should be the first value of the first line of data/train/X_train.txt However when computing the mean, I obtained 0.00226869, whereas the value in the X_train.txt file is 0.28858
This same discrepancy occurs for the y and z values. If I omit the division by 128 (number of samples in a window) then the value is nearer (at least of the same order of magnitude), but still further off than floating point errors should account for (just to be sure, I used the bigfloat package in my Python code with a precision of 15 to ensure the rounding errors were not the problem on my side.
I understand this is a rather niche question and sent it to the admin of the data set, unfortunately, he's out of office until the end of August, so thought I'd ask here in case someone has experience with using this dataset.
i want to recognize human activities in multiple camera environment. I am taking 2 camera views for experiment. I want to fuse information extracted both views to get a precise feature vector. But i am facing following confusions:
As i am using supervised learning, I have to label activities of each person in each frame. In first camera view at some time t, two persons are observed as very close to each other so i label it as interacting. But in second camera view at same time t,it seems that those persons are not close to each other i.e distance is high.. so i labeled it as non- interacting.
How can i fuse two feature vectors (from 2 views) having 2 different labels at same time?
I'm doing some research using smartphones to help control something. I cannot explain the difference between detect a gesture (moving left and then move right) with user's activity (for simple, call moving hand left and right).
Hello everyone,
Does anyone know if I can use video sequences from movies in my research? More specifically, is it legal to trim video segments from movies and run computer vision algorithms on them or would I have copyright issues when I am going to present my experimental results in a international conference? Is there a company that could grants permits for digital media? Does the same law applies for EU projects?
I'm attempting an analysis on sampling rates and window sizes of accelerometer for human activity recognition (HAR). I'm looking for a good test for statistical significance. My data can't fulfill the sphericity assumption for the repeated measures two way ANOVA. As for the Friedman test, I have replicated 10-fold data which means it's not "unreplicated block data". Are there any alternatives I should look at? Or perhaps some way to adjust the dataset to fulfill the requirements of either test?
Thanks.
I'm doing recognition of activities from RF-signals which is device-free and I'm using USRP device.So what kind of activities do you think is suitable to measure with,because during the experiments they are many interference and noises.
I'm working on the topic of analyzing important information for motion representation. For example you have a dataset of human typical activities as: walking, running, climbing, etc and you are asked to represent each class (type) with as sufficient information as possible.
in other words, we want to find prototypes for each activity (ex: like in clustering). Knowing that each motion is consist of body features like movements of legs, hands, head, etc , Then do you think it'd be a practical idea if:
1- we consider all the features together and combine them like what we do in PCA or considering the total distance of an activity to others (like in clustering)?
or
2- considering features separately and find the features which describe each activity more efficiently. For example: considering leg movements for walking or running while hand movements in punching or handshake?
From my point of view, the first one is computationally efficient, while the latter might give more precise/semantic representation for each prototype of activity.
However i'm so interested to hear from experienced people in motion analysis about which way the prefer to handle the problem and why?
We are working on the problem of outdoor activity recognition, for this purpose, we need to test our approaches using a dataset that contains users' mobility traces, we need both the continuous GPS recordings and the visited places.
I have a tracker that outputs the trajectory (x,y,z) of an object (e.g., a can).
I want to use these trajectories to train a classifier (i.e., SVM) in order to infer the activity that the person manipulating the object is performing (e.g., drinking from a can or pouring from a can )
Which kind of features should I use to quantize these trajectories?
Assuming:
- the number of users is unknown at run time.
- in all group activities users perform similar actions.
I'm using Quality Function Deployment (QFD) to perform network (Access Point) prioritisation and selection, semantically (context aware). Any ideas for advancement, extension or comparison?
Conference Paper Seamless semantic service provisioning mechanism for Ambient...
Currently, we have acquired video data of human actions performing martial arts movements. We want to segment the video frames into different actions (sequentially). Can anyone suggest what the best method so far is for this problem? Some good links are also welcomed. Thank you.
Is there a comprehensive taxonomy that can explain the state of the art of current abnormal events detection techniques from video?
Are there articles on how religious norms can be subject to mutual recognition (as in goods and services) in religious marriage contracts?
I have tried to biotinylate some Fab fragments but I am afraid the Biotinylation is happening also at the recognition site, thus impeding the binding of my target peptides.
I would like to have some advices, suggestions, how to deal with this problem.
I am looking for literature on how to carry out research on human activity recognition.
Real time Activity Recognition using a tri-axial accelerometer
Using only an accelerometer and limited data storage capacity I would like to be able to determine what activity is being performed in real time. The focus for this activity recognition is animal activity.
I want to know the last techniques of classification in images and videos to enhance the precision of classification.
I am working on Physical Activity Recognition using data acquired from smartphone sensors (gyroscope, magnetometer and accelerometer) and would like to compare different classifier performance on the dataset, but wondering which evaluation matrix would be best to use: True Positive Rate (TPR), False Positive Rate (FPR), Precision, Recall, F-score or overall classification accuracy? This is a six class problem (6 different activities).
I would like to reduce sequences of about 50 RGB frames of 640x480 pixels to get a representation of the data I could feed into an deep neural network. The goal is to recognize a performed activity/gesture.
I have seen many examples for individual images with static gestures but I struggle to find practical examples using whole sequences with a dynamic gesture.
I have worked through the tutorial here* and they use the MNIST dataset to train the network. So their input are images of the size 28x28 pixel. I would like to use my data as input but I don't really know how to reduce and how much reduction is enough/necessary.
What I did until now is remove the background and then perform edge detection using the openCV Canny edges algorithm** which works fine but still leaves me with a lot of data.
I tried using image flow to generate something like a heatmap but I am not very happy with the results. I read about DTW or Space Time Shapes, but have not yet found a way to apply the theory.
So, do you have any hints, tips or links to papers, tutorials, presentations or whatever to help me reducing the video sequences without loosing to much data? I would prefer practical examples.
Thank you!
Im trying to work for gait analysis of humans.
Does anyone have experience in using Microsoft's face tracking SDK which utilizes AAM's? Any comments on robustness?
I am interested in modelling human activities using sensor data with HMMs and would like to incorporate prior knowledge during inference. The normal procedure is to model K different activities with K separate HMMs. To test an unknown sequence, compute its likelihood from each of the HMMs and the HMM with maximum value is assigned as the class label. This is all done under the assumption that priors over HMM are uniform.
A problem can arise when one of the class is rare or unusual and its prior probability may be very low in comparison to other classes and therefore the uniform priors may not be a good assumption. Therefore, I am interested in posterior probability and not just the likelihood to capture the combined effect. My observations are continuous (features extracted from sensor) and not discrete values. My questions are:
1. Can inference be done using a bayesian network type approach that include multiplication of prior with likelihood?
2. In my case the prior will be the count of activities available per HMM. Can that be estimated using a dirichlet prior to avoid zero-count problem for rare class (assuming I approximate an HMM for a rare class). Does that make sense?
3. The multivariate observation data is approximated using single gaussian (and not mixtures), in that case likelihood will be gaussian?, can it be mixed with dirichlet prior to compute posterior probability? or the likelihood is still multinomial as it represents K different outcomes from K different HMMs?
Sorry if I have mixed with some of the basic concepts, I am new and I seek guidance to move further.
Considering that video recordings took place in a home based environment. Do you believe that skin detection could actually segment human persons from background objects with similar colour (e.g. a skin colored closet, or table)? Could anyone recommend a relevant publication?
I have noticed that someone can find a lot of work on Human activity recognition, but just a few ones focus on human activity detection problem (also referred in literature as activity localization or action spotting). This renders human activity recognition useless for real-life applications, as most videos are unsegmented and cannot be annotated as global entities that contain just one action. Do you have any suggestions - ideas concerning how this problem might be solved?
I am looking for activity recognition data sets that are captures through sensors (i.e. accelerometer, gyroscope etc) or using a smartphone. Most of the publicly available data sets contain data from normal activities of daily living (e.g. walking, running, cycling etc), however I am interested in data sets that shall also contain data from unusual/abnormal activities such as fall or stroke apart from normal activities. Currently I am using DLR Human Activity Recognition data set, and looking for some other similar data sets. I would appreciate if you would please direct me to any such data you are aware of.