| Literature DB >> 35017796 |
Eman Shalaby1, Nada ElShennawy1, Amany Sarhan1.
Abstract
In recent years, channel state information (CSI) in WiFi 802.11n has been increasingly used to collect data pertaining to human activity. Such raw data are then used to enhance human activity recognition. Activities such as lying down, falling, walking, running, sitting down, and standing up can now be detected with the use of information collected through CSI. Human activity recognition has a multitude of applications, such as home monitoring of patients. Four deep learning models are presented in this paper, namely: a convolution neural network (CNN) with a Gated Recurrent Unit (GRU); a CNN with a GRU and attention; a CNN with a GRU and a second CNN, and a CNN with Long Short-Term Memory (LSTM) and a second CNN. Those models were trained to perform Human Activity Recognition (HAR) using CSI amplitude data collected by a CSI tool. Experiments conducted to test the efficacy of these models showed superior results compared with other recent approaches. This enhanced performance of our models may be attributable the ability of our models to make full use of available data and to extract all data features, including high dimensionality and time sequence. The highest average recognition accuracy reached by the proposed models was achieved by the CNN-GRU, and the CNN-GRU with attention models, standing at 99.31% and 99.16%, respectively. In addition, the performance of the models was evaluated for unseen CSI data by training our models using a random split-of-dataset method (70% training and 30% testing). Our models achieved impressive results with accuracies reaching 100% for nearly all activities. For the lying down activity, accuracy obtained from the CNN-GRU model stood at 99.46%; slightly higher than the 99.05% achieved by the CNN-GRU with attention model. This confirmed the robustness of our models against environmental changes.Entities:
Keywords: Channel state information; Convolution neural network; Gated recurrent unit; Human activity recognition
Year: 2022 PMID: 35017796 PMCID: PMC8739002 DOI: 10.1007/s00521-021-06787-w
Source DB: PubMed Journal: Neural Comput Appl ISSN: 0941-0643 Impact factor: 5.606
Fig. 1Examples of CSI Amplitude Variations due to Human Activities: a Fall, b Run, and c Stand up
Summary of related work
| Reference no | Activities | Classification method | Accuracy |
|---|---|---|---|
| [ | Lie down, Fall, Walk, Run, Sit down and Stand up | LSTM | 75% |
| [ | Run, Walk, Sit down, Open refrigerator, Fall, Box, Push one hand, and Brush teeth | CARM | 96% |
| [ | Lie down, Fall, Walk, Run, Sit down and Stand up | ABLSTM | 95% |
| [ | Lie down, Fall, Walk, Run, Si down and Stand up | CNN + LSTM | 95% |
| [ | Push, Dodge, Strike, Pull, Drag, Kick, Circle, Punch (twice), and Bowl | RF | 89.147% for NLOS |
| [ | Sit down and Stand up | SVM | 98.4% |
| [ | Sign language gestures and FallDeFi Data: Fall, Walk, Jump, pick up, Sit down, and Stand up | CSIGAN | 84.17% for sign language 86.27% for FallDeFi |
| [ | Bend, Halve squat, Step, Stretch leg, and Jump | SVM with DTW | 96.6% for LOS 92% for NLOS |
| [ | Push, Wave, Kick, Run, Fall, Box, Sit, Pick, Walk, and Empty | ELM | 94.2% |
| [ | Walk, Sit, Stand, Run and Fall | SVM, LSTM | 95% and more for the rest of activities, 100% for fall |
| [ | Bend, Box, Clap and Wave, Fall, pick up, Run, Sit down, Stand up and Walk | CNN + BLSTM | 96.96% for four activities for three regions 90% for six activities except stand up is 86% |
| [ | Lay down, Fall, Walk, Run, Sit down and Stand up | GAN + LSTM | 87.2% With 50% accuracy and 50% Synthetic 92.8 for all data accurate |
Fig. 2LSTM architecture
Fig. 3BLSTM Architecture
Fig. 4GRU circuit
Fig. 5Architecture CSI-based deep learning activity recognition system
Fig. 6The CNN-GRU human activity recognition model
Fig. 7The CNN-GRU-attention human activity recognition model
Fig. 8The CNN-GRU-CNN human activity recognition model
Fig. 9The CNN-LSTM-CNN human activity recognition model
Number of samples per activity
| Activity | No. of Samples |
|---|---|
| Lie Down | 1318 |
| Fall | 889 |
| Walk | 2931 |
| Run | 2408 |
| Sit down | 812 |
| Stand up | 601 |
| Total | 8959 |
Confusion matrix of the (LSTM) model [11]
| LSTM [ | |||||||
|---|---|---|---|---|---|---|---|
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
| Actual | Lie down | 0.01 | 0.01 | 0.01 | 0.00 | 0.02 | |
| Fall | 0.01 | 0.05 | 0.00 | 0.00 | 0.00 | ||
| Walk | 0.00 | 0.01 | 0.04 | 0.01 | 0.01 | ||
| Run | 0.00 | 0.00 | 0.02 | 0.01 | 0.00 | ||
| Sit down | 0.03 | 0.01 | 0.05 | 0.02 | 0.07 | ||
| Stand up | 0.01 | 0.00 | 0.03 | 0.05 | 0.07 | ||
Confusion matrix of the (ABLSTM) model [12]
| ABLSTM [ | |||||||
|---|---|---|---|---|---|---|---|
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
| Actual | Lie down | 0.0 | 0.01 | 0.0 | 0.02 | 0.01 | |
| Fall | 0.0 | 0.0 | 0.01 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0.02 | 0.0 | 0.0 | ||
| Run | 0.0 | 0.0 | 0.02 | 0.0 | 0.0 | ||
| Sit down | 0.01 | 0.01 | 0.01 | 0.0 | 0.02 | ||
| Stand up | 0.01 | 0.0 | 0.0 | 0.0 | 0.01 | ||
Confusion matrix of the CNN – GRU learning model:( a) with k-fold, and (b) with randomly splitting of the dataset
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
|---|---|---|---|---|---|---|---|
| Actual | Lie down | 0.0 | 0.0015 | 0.0 | 0.0008 | 0.0008 | |
| Fall | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0 | 0.0 | 0.0 | ||
| Run | 0.0 | 0.0 | 0.0 | 0.0 | 0.0004 | ||
| Sit down | 0.0037 | 0.0024 | 0.0 | 0.0 | 0.0024 | ||
| Stand up | 0.0017 | 0.0 | 0.0 | 0.0017 | 0.0033 | ||
| Actual | Lie down | 0.0 | 0.0025 | 0.0 | 0.0 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0 | 0.0 | 0.0 | ||
| Run | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Sit down | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Stand up | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
Fig. 10CNN-GRU model performance results
Confusion matrix of the CNN-GRU-Attention deep learning model:( a) with k-fold, and (b) with randomly splitting of the dataset
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
|---|---|---|---|---|---|---|---|
| Actual | Lie down | 0.0 | 0.0008 | 0.0 | 0.0015 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0003 | 0.0003 | 0.0003 | 0.0 | ||
| Run | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Sit down | 0.0012 | 0.0 | 0.0 | 0.0 | 0.0037 | ||
| Stand up | 0.0 | 0.0017 | 0.0017 | 0.0 | 0.0033 | ||
| Actual | Lie down | 0.0 | 0.0051 | 0.0 | 0.0 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0023 | 0.0 | 0.0 | 0.0 | 0.0023 | ||
| Run | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Sit down | 0.0 | 0.0 | 0.0041 | 0.0 | 0.0 | ||
| Stand up | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
Fig. 11CNN-GRU-Attention model performance results
Confusion matrix of the CNN-GRU-CNN deep learning model:( a) with k-fold and (b) with randomly splitting of the dataset
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
|---|---|---|---|---|---|---|---|
| (a) CNN-GRU-CNN with k-fold Predicted | |||||||
| Actual | Lie down | 0.0 | 0.0 | 0.0008 | 0.0053 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0011 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0.0010 | 0.0 | 0.0 | ||
| Run | 0.0004 | 0.0 | 0.0008 | 0.0 | 0.0 | ||
| Sit down | 0.0025 | 0.0 | 0.0012 | 0.0 | 0.0073 | ||
| Stand up | 0.0 | 0.0017 | 0.0017 | 0.0 | 0.01033 | ||
| (b) CNN-GRU-CNN with randomly splitting of the dataset Predicted | |||||||
| Actual | Lie down | 0.0 | 0.0 | 0.0 | 0.0127 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0037 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0.0079 | 0.0 | 0.0046 | ||
| Run | 0.0 | 0.0 | 0.0028 | 0.0 | 0.0 | ||
| Sit down | 0.0 | 0.0 | 0.0041 | 0.0 | 0.0082 | ||
| Stand up | 0.0 | 0.0 | 0.0 | 0.0 | 0.0111 | ||
Fig. 12CNN-GRU-CNN model performance results
Confusion matrix of the CNN-LSTM-CNN deep learning model:( a) with k-fold, and (b) with randomly splitting of the dataset
| Lie down | Fall | Walk | Run | Sit down | Stand up | ||
|---|---|---|---|---|---|---|---|
| Actual | Lie down | 0.0008 | 0.0061 | 0.0 | .0061 | 0.0099 | |
| Fall | 0.0 | 0.0011 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0003 | 0.0 | 0.0017 | 0.0 | 0.0003 | ||
| Run | 0.0004 | 0.0012 | 0.0120 | 0.0 | 0.0008 | ||
| Sit down | 0.0111 | 0.0012 | 0.0062 | 0.0 | 0.0172 | ||
| Stand up | 0.0 | 0.0 | 0.0017 | 0.0 | 0.0116 | ||
| Actual | Lie down | 0.0 | 0.0 | 0.0 | .0101 | 0.0 | |
| Fall | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Walk | 0.0 | 0.0 | 0.0034 | 0.0 | 0.0 | ||
| Run | 0.0 | 0.0 | 0.0 | 0.0 | 0.0 | ||
| Sit down | 0.0041 | 0.0 | 0.0082 | 0.0 | 0.0 | ||
| Stand up | 0.0 | 0.0 | 0.0 | 0.0 | 0.0056 | ||
Fig. 13CNN-LSTM-CNN model performance results
Average Accuracy for our four proposed models
| Models | Accuracy with k-fold method (%) | Accuracy with split method (%) |
|---|---|---|
| CNN-GRU | 99.31 | 99.46 |
| CNN-GRU-Attention | 99.16 | 99.05 |
| CNN-GRU-CNN | 98.88 | 99.05 |
| CNN-LSTM-CNN | 98.71 | 98.99 |
Performance metrics results for the four proposed deep learning models
| Proposed Model | Loss | Accuracy | Precision | Recall | AUC |
|---|---|---|---|---|---|
| CNN-GRU | .0026 | 99.46 | 99.52 | 99.43 | 99.90 |
| CNN-GRU-Attention | .0103 | 99.05 | 99.14 | 99.01 | 99.77 |
| CNN-GRU-CNN | .0411 | 99.05 | 99.09 | 99.03 | 99.74 |
| CNN-LSTM-CNN | .0340 | 98.99 | 99.03 | 98.96 | 99.70 |
Consumed Time for the Deep Learning Models per sample
| Time | LSTM [ | ABLSTM [ | CNN-GRU | CNN-GRU-Attention | CNN-GRU-CNN | CNN-LSTM-CNN |
|---|---|---|---|---|---|---|
| Training (sec) | .011 | .042 | .0063 | .0041 | .0047 | .0082 |
| Testing (sec) | .006 | .019 | .0033 | .0019 | .0022 | .0045 |
Models overall comparison in terms of Response Time, Accuracy for all activities, and Total model parameters
| Model | Response time (sec) | Accuracy for all activities (up to)% | Total model parameters |
|---|---|---|---|
| LSTM [ | .0060 | 81 | |
| ABLSTM [ | .0190 | 95 | 3,866,606 |
| CNN-GRU | .0033 | 99.14 | 1,307,526 |
| CNN-GRU-Attention | 914,567 | ||
| CNN-GRU-CNN | .0022 | 98.89 | 2,469,766 |
| CNN-LSTM-CNN | .0045 | 96.43 | 2,568,326 |