| Literature DB >> 35684649 |
Luís M Martins1,2,3, Nuno Ferrete Ribeiro1,2,3,4, Filipa Soares1,2,3, Cristina P Santos1,2,3.
Abstract
The recognition of Activities of Daily Living (ADL) has been a widely debated topic, with applications in a vast range of fields. ADL recognition can be accomplished by processing data from wearable sensors, specially located at the lower trunk, which appears to be a suitable option in uncontrolled environments. Several authors have addressed ADL recognition using Artificial Intelligence (AI)-based algorithms, obtaining encouraging results. However, the number of ADL recognized by these algorithms is still limited, rarely focusing on transitional activities, and without addressing falls. Furthermore, the small amount of data used and the lack of information regarding validation processes are other drawbacks found in the literature. To overcome these drawbacks, a total of nine public and private datasets were merged in order to gather a large amount of data to improve the robustness of several ADL recognition algorithms. Furthermore, an AI-based framework was developed in this manuscript to perform a comparative analysis of several ADL Machine Learning (ML)-based classifiers. Feature selection algorithms were used to extract only the relevant features from the dataset's lower trunk inertial data. For the recognition of 20 different ADL and falls, results have shown that the best performance was obtained with the K-NN classifier with the first 85 features ranked by Relief-F (98.22% accuracy). However, Ensemble Learning classifier with the first 65 features ranked by Principal Component Analysis (PCA) presented 96.53% overall accuracy while maintaining a lower classification time per window (0.039 ms), showing a higher potential for its usage in real-time scenarios in the future. Deep Learning algorithms were also tested. Despite its outcomes not being as good as in the prior procedure, their potential was also demonstrated (overall accuracy of 92.55% for Bidirectional Long Short-Term Memory (LSTM) Neural Network), indicating that they could be a valid option in the future.Entities:
Keywords: Machine Learning; activity recognition; dataset fusion; deep learning; falls; feature selection
Mesh:
Year: 2022 PMID: 35684649 PMCID: PMC9185447 DOI: 10.3390/s22114028
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Datasets used for the evaluation of the Deep Learning algorithms analyzed in the literature and respective description regarding sensing methods, sample frequency, participants, number of activities (classes) recorded and algorithm performance (accuracy). In this table: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
| DataSet | Work | Sensors | Sample | Participants | N° of | Accuracy |
|---|---|---|---|---|---|---|
| Private | Chung et al. [ | A, G, M | 100 Hz | 5 | 9 | 93% |
| SisFall [ | Wang et al. [ | A, G | 200 Hz | 38 | 2 | <99% |
| PAMAP2 [ | Gil-Martín et al. [ | A, G, M | 9 Hz | 9 | 18 | 96.62% |
| UCI-HAD [ | Altuve et al. [ | A, G, M | 50 Hz | 30 | 6 | [ |
| USC-HAD [ | Murad et al. [ | A, G | 100 Hz | 14 | 12 | 97.8% |
| Opportunity [ | Murad et al. [ | A, G, M | 30 Hz | 4 | 18 1 | 92.5% |
| Daphnet FOG [ | Murad et al. [ | A | 64 Hz | 10 | 2 | 94.1% |
| Skoda [ | Murad et al. [ | A | 98 Hz | 1 | 11 1 | 92.6% |
1 ADL related to hand gestures.
Datasets description regarding sensing methods and location, sample frequency, participants and activities recorded, where: A = Accelerometer, G = Gyroscope, M = Magnetometer, B = Barometer and ADL = Activities of Daily Living.
| DataSet | Availability | Sensors | Location | Sample | Participants | ADL |
|---|---|---|---|---|---|---|
| SisFall [ | Public | A, G | Waist | 200 Hz | 23 subjects <30 years | 19 ADL |
| FALLALLD [ | Public | A, G | Chest, Waist, | 238 Hz | 15 subjects | 44 ADL |
| FARSEEING [ | Public | A, G, M | Waist, Thigh | 20 Hz | 20 subjects 2 | Real falls |
| UCI HAR [ | Public | A, G | Waist | 50 Hz | 30 subjects | 12 ADL |
| Cotechini [ | Public | A, G | Waist | 33, 33 Hz | 8 subjects | 5 ADL |
| UMAFall [ | Public | A, G, M | Waist, Chest, | 20 Hz | 17 subjects | 8 ADL |
| +Sense [ | Private | A, G, M | Waist | 100 Hz | 10 Healthy | 1 ADL |
| SafeWalk [ | Private | A, G, M | Waist, Thighs, | 30 Hz | 12 subjects | 1 ADL |
| InertialLab [ | Private | A, G, M | Waist, Thighs, | 200 Hz | 7 subjects | 5 ADL |
1 Several activities in these datasets were grouped into one single class of basic activities. 2 Only data from 3 subjects were suitable to use.
Figure 1(a) Normalization process steps implemented in order to normalize the public datasets for ADL recognition. (b) Desired sensors orientation. The inertial sensors have to be located on the subjects’ waist. The arrows and letters x, y and z indicate the positive direction of the anteroposterior, mediolateral and longitudinal axes, respectively.
Static postures and locomotion daily activities, postural transitions and fall events selected to be recognized by the Machine and Deep Learning models.
| Periodic Activities and Static Postures | Transitions | Fall Events |
|---|---|---|
| Walking | Lying to Stand | Forwards |
| Standing | Stand to Sit | Backwards |
| Sitting | Sit to Stand | Lateral |
| Lying | Stand to Pick to Stand | Syncope |
| Upstairs | Stand to Lying | |
| Downstairs | Change Position (Lying) | |
| Jumping | Turning | |
| Jogging | Bending |
Figure 2Percentage quantity of windows created of each activity present in the created dataset. The activities are named according to Table 3.
Figure 3Complete process for all the validation, training and evaluation of different Machine Learning and Deep Learning models alongside the best feature set selected by diverse feature selection methods.
Figure 4Signal segmentation example, using the sliding window technique example for feature extraction. A sliding window across time is represented by windows 1 to 3.
List of all extracted features from each window created. AP, V and ML refer to the anteroposterior, vertical and mediolateral axis, respectively.
| Feature Number | Feature Description |
|---|---|
| [1–6] | Acceleration and Angular velocity (AP, V, ML) |
| [7–8] | SumVM of acceleration and Angular velocity |
| [9–24] | Skewness and kurtosis of acceleration, Angular velocity (AP, V, ML) |
| [25–64] | Min, max, mean, variance and Std deviation of acceleration, angular |
| [65–70] | Correlation between V-ML, V-AP and ML-AP axis of acceleration and |
| [71–77] | Slope, Total angular change, Resultant angular acceleration, ASMA, SMA, |
| [78–102] | Peak-to-Peak, Root Mean Square and Ratio Index |
| [103–115] | Resultant angle change, Flutuation frequency, Resultant |
| [116–117] | Resultant of Delta changes of acceleration and Angular velocity |
| [118–133] | Gravity component, Displacement, Displacement range, |
| Slope changes, Zero crossings, Waveform length of acceleration, | |
| [133–189] | Energy, Mean frequency, Peak frequency and magnitude of acceleration, |
| [190–195] | SumVM of resultant angular velocity, average acceleration |
| [196–199] | Acceleration exponential moving average, Rotational angle |
Feature Selection Methods Tested for ADL Recognition.
| Feature Selection Methods (FSM) | FSM Type |
|---|---|
| Infnite Latent Feature Selection (ILFS)
[ | Filtering |
| Unsupervised Feature Selection with Ordinal Locality (UFSOL) [ | Wrapper |
| Feature Selection with Adaptive Structure Learning (FSASL) [ | Wrapper |
| Minimum-Redundancy Maximum-Relevancy (MRMR) [ | Filtering |
| Relief-F [ | Filtering |
| Mutual Information Feature Selection (MutInfFS) [ | Filtering |
| Feature Selection Via Concave Minimization (FSV) [ | Embedded |
| Correlation-Based Feature Selection (CFS) [ | Filtering |
| Least Absolute Shrinkage and Selection Operator (LASSO) [ | Embedded |
| Principal Component Analysis (PCA) [ | Filtering |
Figure 5PCA-based procedure to rank and obtain the most crucial features and to limit the computational cost of comparative analysis.
Machine Learning models evaluated in this work and respective descriptions.
| Model | Reference | Description |
|---|---|---|
| DA | [ | A method that finds combinations of features that separate two or more classes of objects or events, searching for the most variance between classes, and information that maximizes the difference between classes. |
| K-NN | [ | Compares each new instance with all datasets available and the instance closest by distance metrics is used to perform classification. Since every sample of the dataset must be checked for every instance, the time and complexity of the method rises according to the dataset size. |
| Ensemble | [ | Creates multiple instances of traditional ML methods and combines them to evolve a single optimal solution to a problem. This approach is capable of producing better predictive models compared to the traditional approach. |
| DTs | [ | A model that predicts the value of a target variable based on numerous input variables. A decision tree is constituted by an internal node, based on which the tree splits into branches. The end of the branch that does not split any longer is the decision. |
Figure 6Neural Networks architectures used in this manuscript. The input shown represents a single feature window. (a) The CNN identifies correlations from the various features provided. (b) The LSTM network detects crucial temporal features. (c) The BiLSTM network has a similar operating mode as LSTM but with bidirectional LSTM layers. (d) The hybrid CNN-LSTM extracts temporal patterns using convolutional features from the CNN convolutional layer. Based on [53].
Specifications for the use of the Deep Learning models depicted in Figure 6.
| Specification | Value |
|---|---|
| Epoch Number | 100 |
| Hidden Layers | 150 |
| Batch Size | 64 |
| Optimizer | Adam [ |
| Learning Rate | 0.001 (Constant) |
| Loss Function | Cross-Entropy |
Features ranked in descending order per feature selection model.
| FSM | No. of | Ranked Features |
|---|---|---|
| Relief-F | 85 | 66,69,70,68,67,65,110,128,142,143,144,35,31,101,12,9,111,148,15,14,112,114, |
| PCA | 65 | 9,97,188,42,102,43,101,128,144,113,148,110,184,31,142,154,116,83,41,103,111, |
Comparison of the best classification results (ACC, Sens, Spec, Prec, F1S, MCC), attained after the 5-1 and 5-10 k-fold cross-validation steps for the K-NN and Ensemble Learning classifiers.
| ML Model | FSM | CV Step | N° of | ACC (%) | Sens (%) | Spec (%) | Prec (%) | F1S (%) | MCC (%) |
|---|---|---|---|---|---|---|---|---|---|
| K-NN | Relief-F | 85 | 93.63 | 84.17 | 99.64 | 86.80 | 85.43 | 85.10 | |
| PCA | 5 Fold | 85 | 92.99 | 84.08 | 99.60 | 86.01 | 85.01 | 84.63 | |
| FSASL | 70 | 91.49 | 81.39 | 99.51 | 83.66 | 82.48 | 82.02 | ||
|
Ensemble | PCA | 65 | 94.59 | 82.22 | 99.68 | 90.54 | 85.80 | 85.78 | |
| K-NN | Relief-F | 85 | 93.62 ± 0.016 | 84.12 ± 0.066 | 99.64 ± 0.001 | 86.75 ± 0.055 | 85.38 ± 0.056 | 85.05 ± 0.056 | |
| PCA | 5 Fold | 85 | 92.95 ± 0.021 | 83.91 ± 0.094 | 99.60 ± 0.001 | 85.88 ± 0.085 | 84.86 ± 0.085 | 84.48 ± 0.086 | |
| FSASL | 70 | 91.48 ± 0.026 | 81.40 ± 0.063 | 99.51 ± 0.001 | 83.59 ± 0.079 | 82.45 ± 0.066 | 81.99 ± 0.067 | ||
|
Ensemble | PCA | 65 | 94.59 ± 0.015 | 82.18 ± 0.067 | 99.68 ± 0.001 | 90.64 ± 0.073 | 85.79 ± 0.061 | 85.79 ± 0.060 |
Hold-out test results for the Ensemble Learning with the first 65 features ranked by the PCA and for the K-NN classifier with the first 85 features ranked by the Relief-f.
| ML Model | FSM | N° of | ACC (%) | Sens (%) | Spec (%) | Prec (%) | F1S (%) | MCC (%) |
|---|---|---|---|---|---|---|---|---|
| K-NN | Relief-F | 85 | 97.27 | 92.90 | 99.84 | 93.79 | 93.34 | 93.19 |
| Ensemble | PCA | 65 | 95.44 | 85.97 | 99.73 | 91.67 | 88.43 | 88.36 |
Results for the test of the 4 Deep Learning architectures with the 85 first features ranked by Relief-f and 65 first features ranked by PCA.
| FSM | Feature | Architecture | ACC (%) | Sens (%) | Spec (%) | Prec (%) | F1S (%) | MCC (%) |
|---|---|---|---|---|---|---|---|---|
| Relief-F | 85 | CNN | 57.01 | 37.06 | 97.22 | 54.67 | 35.47 | 37.87 |
| LSTM | 92.06 | 79.58 | 99.55 | 84.25 | 81.02 | 81.01 | ||
| CNN-LSTM | 88.84 | 74.48 | 99.36 | 75.24 | 74.53 | 74.06 | ||
| BiLSTM | 92.55 | 81.14 | 99.57 | 85.56 | 83.14 | 82.83 | ||
| PCA | 65 | CNN | 42.67 | 26.46 | 96.15 | 54.49 | 22.27 | 24.90 |
| LSTM | 91.46 | 77.81 | 99.51 | 84.38 | 80.61 | 80.38 | ||
| CNN-LSTM | 88.55 | 74.33 | 99.35 | 75.09 | 74.36 | 73.88 | ||
| BiLSTM | 91.48 | 79.33 | 99.52 | 83.32 | 80.67 | 80.52 |
Window size comparative study results for the K-NN best optimized model with the Relief-F feature selection model.
| ML Model + FSM | Window Size (s) | Window | ACC | Sens | Spec | Precn | F1S | MCC |
|---|---|---|---|---|---|---|---|---|
| K-NN + Relief-f | 0.5 | 80 | 98.22 | 95.20 | 99.90 | 96.04 | 95.62 | 95.52 |
| 1 | 97.27 | 92.90 | 99.84 | 93.79 | 93.34 | 93.19 | ||
| 1.5 | 96.30 | 91.73 | 99.79 | 91.15 | 91.41 | 91.22 | ||
| 2 | 95.33 | 90.53 | 99.74 | 88.51 | 89.44 | 89.22 | ||
| Ensemble + PCA | 0.5 | 96.53 | 88.94 | 99.79 | 94.09 | 91.29 | 91.21 | |
| 1 | 95.44 | 85.97 | 99.73 | 91.67 | 88.43 | 88.36 | ||
| 1.5 | 95.01 | 85.60 | 99.71 | 90.76 | 87.64 | 87.62 | ||
| 2 | 94.51 | 85.21 | 99.68 | 89.37 | 86.92 | 86.79 |
Classification time for the training and testing of the two best combinations of Machine Learning (ML) model and Feature Selection Method (FSM), for each of the selected windows for the window size study.
| ML Model + FSM | Window | Window | Test | Train | Test | Test Time |
|---|---|---|---|---|---|---|
| K-NN + Relief-f | 0.5 | 80 | 409,740 | 4.36 | 213,588.88 | 0.521 |
| 1 | 199,997 | 4.18 | 66,782.58 | 0.334 | ||
| 1.5 | 130,421 | 4.70 | 12,633.08 | 0.097 | ||
| 2 | 95,482 | 4.09 | 6752.47 | 0.071 | ||
| Ensemble + PCA | 0.5 | 409,740 | 829.55 | 15.99 |
| |
| 1 | 199,997 | 279.03 | 8.54 |
| ||
| 1.5 | 130,421 | 145.21 | 5.68 |
| ||
| 2 | 95,482 | 100.23 | 3.94 |
|