| Literature DB >> 29970873 |
Xiaochen Zheng1, Meiqing Wang2, Joaquín Ordieres-Meré3.
Abstract
According to the Industry 4.0 paradigm, all objects in a factory, including people, are equipped with communication capabilities and integrated into cyber-physical systems (CPS). Human activity recognition (HAR) based on wearable sensors provides a method to connect people to CPS. Deep learning has shown surpassing performance in HAR. Data preprocessing is an important part of deep learning projects and takes up a large part of the whole analytical pipeline. Data segmentation and data transformation are two critical steps of data preprocessing. This study analyzes the impact of segmentation methods on deep learning model performance, and compares four data transformation approaches. An experiment with HAR based on acceleration data from multiple wearable devices was conducted. The multichannel method, which treats the data for the three axes as three overlapped color channels, produced the best performance. The highest overall recognition accuracy achieved was 97.20% for eight daily activities, based on the data from seven wearable sensors, which outperformed most of the other machine learning techniques. Moreover, the multichannel approach was applied to three public datasets and produced satisfying results for multi-source acceleration data. The proposed method can help better analyze workers’ activities and help to integrate people into CPS.Entities:
Keywords: Human Activity Recognition (HAR); Industry 4.0; Internet of things (IoT); data preprocessing; deep learning
Mesh:
Year: 2018 PMID: 29970873 PMCID: PMC6068555 DOI: 10.3390/s18072146
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Framework of the proposed method.
Figure 2Raw acceleration plot of time domain: segment length 512 and sampling rate 50 Hz (the dotted lines are added manually for better clarification; image resolution is 512 × 512 pixels).
Figure 3Multichannel RGB color plot on time domain. The segment length is 512 and the sampling rate is 50 Hz (image resolution 512 × 7 pixels).
Figure 4Spectrogram plot of the acceleration data. The segment length is 512 and the sampling rate is 50 Hz, with a short-time Fourier transform (STFT) window length of 64 STFT and an overlap length of 60 STFT (image resolution 336 × 350 pixels).
Shallow features extracted from acceleration data.
| Data | Features |
|---|---|
| Raw signal | max, min, mean, median, variance, kurtosis, skewness, zero-cross, root mean square, standard deviation, interquartile range |
| First derivative | mean, variance, root mean square, standard deviation |
Figure 5Workflow of deep convolutional neural network (CNN) models.
Overall accuracy (%) of the four data transformation methods, based on five segmentation options.
| Segment Length | Raw Plot | Multichannel | Spectrogram | Spectrogram and Shallow Features |
|---|---|---|---|---|
| 64 | 92.44 | 94.60 | 92.86 | 90.39 |
| 128 | 93.05 | 96.14 | 93.37 | 90.42 |
| 256 | 93.45 | 96.58 | 93.94 | 92.02 |
| 512 | 94.97 | 97.19 | 95.56 | 93.58 |
| 1024 | 82.13 | 92.81 | 91.54 | 85.55 |
Variation of overall classification accuracies (%) of 15 subjects based on a segment length of 512 (10.24 s) with four preprocessing methods.
| Subject | Raw Plot | Multichannel | Spectrogram | Spectrogram & Shallow Features |
|---|---|---|---|---|
| Mean | 95.25 | 97.58 | 95.81 | 93.92 |
| Min. | 92.42 | 93.91 | 91.61 | 88.46 |
| Max. | 97.22 | 99.56 | 98.57 | 97.18 |
| Sd. | 1.72 | 2.11 | 2.35 | 2.74 |
Performance of each model based on a segment length of 512 (10.24 s). A1: climbing down; A2: climbing up; A3: jumping; A4: lying; A5: running; A6: sitting; A7: standing; and A8: walking.
| A1 | A2 | A3 | A4 | A5 | A6 | A7 | A8 | ||
|---|---|---|---|---|---|---|---|---|---|
| Raw plot | Precision (%) | 97.16 | 97.99 | 99.61 | 99.59 | 95.18 | 99.15 | 92.06 | 99.49 |
| Recall (%) | 95.41 | 96.89 | 98.78 | 99.18 | 91.40 | 98.53 | 85.24 | 99.08 | |
| Overall Acc. (%) | 94.97 | 95% CI: (0.9434, 0.9556) | |||||||
| Multichannel | Precision (%) | 97.65 | 97.96 | 99.74 | 99.89 | 96.29 | 99.63 | 96.99 | 99.72 |
| Recall (%) | 95.56 | 96.53 | 99.49 | 100.00 | 93.33 | 99.34 | 95.04 | 99.53 | |
| Overall Acc. (%) | 97.19 | 95% CI: (0.9670, 0.9763) | |||||||
| Spectrogram | Precision (%) | 97.65 | 97.23 | 99.92 | 98.60 | 98.84 | 97.47 | 91.18 | 97.76 |
| Recall (%) | 95.65 | 96.05 | 100.00 | 97.56 | 98.96 | 96.55 | 82.73 | 96.08 | |
| Overall Acc. (%) | 94.56 | 95% CI: (0.9251, 0.9618) | |||||||
| Spectrogram & Shallow features | Precision (%) | 94.92 | 98.25 | 91.51 | 98.60 | 95.92 | 96.60 | 93.39 | 95.38 |
| Recall (%) | 91.05 | 98.59 | 83.33 | 97.56 | 93.14 | 93.75 | 88.42 | 91.51 | |
| Overall Acc. (%) | 93.58 | 95% CI: (0.9157, 0.9512) | |||||||
Figure 6Classification accuracy and training time of the four data transformation methods.
Confusion matrix generated by the multichannel model based on a segment length of 512 (A1: climbing down; A2: climbing up; A3: jumping; A4: lying; A5: running; A6: sitting; A7: standing; and A8: walking).
| Original | Prediction | |||||||
|---|---|---|---|---|---|---|---|---|
| A1 | A2 | A3 | A4 | A5 | A6 | A7 | A8 | |
| A1 | 68 | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| A2 | 0 | 78 | 2 | 0 | 1 | 0 | 0 | 1 |
| A3 | 0 | 3 | 22 | 0 | 0 | 0 | 0 | 0 |
| A4 | 0 | 0 | 0 | 81 | 1 | 0 | 0 | 0 |
| A5 | 0 | 6 | 0 | 0 | 98 | 1 | 3 | 0 |
| A6 | 0 | 0 | 0 | 0 | 0 | 92 | 1 | 0 |
| A7 | 0 | 0 | 0 | 1 | 5 | 1 | 86 | 0 |
| A8 | 0 | 1 | 0 | 0 | 0 | 0 | 2 | 100 |
Figure 7Overall classification accuracies of eight activities based on data from seven single positions and two combined positions.
Figure 8Accuracy of different classification methods [8].
Precision (%) and recall (%) obtained by the proposed multichannel (MCT) method and existing study [23] in three public datasets.
| Ravì et al. [ | Prec. | 99.37 | 99.64 | 97.85 | 98.15 | 95.52 | 94.44 |
| Rec. | 99.37 | 99.40 | 98.56 | 97.25 | 95.13 | 95.90 | |
| MCT | Prec. | 98.34 | 98.11 | 100.00 | 100.00 | 96.14 | 98.44 |
| Rec. | 97.31 | 97.53 | 100.00 | 100.00 | 93.10 | 97.67 | |
| Ravì et al. [ | Prec. | 98.01 | 88.65 | 87.32 | 85.00 | 82.05 | 97.17 |
| Rec. | 97.73 | 85.85 | 89.28 | 76.98 | 82.11 | 97.19 | |
| MCT | Prec. | 98.76 | 96.85 | 90.25 | 87.03 | 91.02 | 95.85 |
| Rec. | 97.95 | 94.96 | 82.05 | 75.00 | 85.94 | 94.81 | |
| Ravì et al. [ | Prec. | 96.67 | 97.78 | 89.47 | 91.15 | 100.00 | |
| Rec. | 91.34 | 97.78 | 94.44 | 92.79 | 100.00 | ||
| MCT | Prec. | 100.00 | 99.54 | 100.00 | 100.00 | 80.00 | |
| Rec. | 100.00 | 100.00 | 100.00 | 100.00 | 60.00 | ||
| Ravì et al. [ | Prec. | 88.89 | 92.86 | 98.78 | 100.00 | 93.55 | |
| Rec. | 80.00 | 94.20 | 97.59 | 98.04 | 100.00 | ||
| MCT | Prec. | 99.18 | 100.00 | 100.00 | 100.00 | 94.44 | |
| Rec. | 100.00 | 100.00 | 100.00 | 100.00 | 88.89 | ||