| Literature DB >> 35361804 |
Hritam Basak1, Rohit Kundu1, Pawan Kumar Singh2, Muhammad Fazal Ijaz3, Marcin Woźniak4, Ram Sarkar5.
Abstract
Human Action Recognition (HAR) is a popular area of research in computer vision due to its wide range of applications such as surveillance, health care, and gaming, etc. Action recognition based on 3D skeleton data allows simplistic, cost-efficient models to be formed making it a widely used method. In this work, we propose DSwarm-Net, a framework that employs deep learning and swarm intelligence-based metaheuristic for HAR that uses 3D skeleton data for action classification. We extract four different types of features from the skeletal data namely: Distance, Distance Velocity, Angle, and Angle Velocity, which capture complementary information from the skeleton joints for encoding them into images. Encoding the skeleton data features into images is an alternative to the traditional video-processing approach and it helps in making the classification task less complex. The Distance and Distance Velocity encoded images have been stacked depth-wise and fed into a Convolutional Neural Network model which is a modified version of Inception-ResNet. Similarly, the Angle and Angle Velocity encoded images have been stacked depth-wise and fed into the same network. After training these models, deep features have been extracted from the pre-final layer of the networks, and the obtained feature representation is optimized by a nature-inspired metaheuristic, called Ant Lion Optimizer, to eliminate the non-informative or misleading features and to reduce the dimensionality of the feature set. DSwarm-Net has been evaluated on three publicly available HAR datasets, namely UTD-MHAD, HDM05, and NTU RGB+D 60 achieving competitive results, thus confirming the superiority of the proposed model compared to state-of-the-art models.Entities:
Mesh:
Year: 2022 PMID: 35361804 PMCID: PMC8971421 DOI: 10.1038/s41598-022-09293-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Figure 1Overall workflow of the proposed DSwarm-Net model for solving 3D HAR problem.
Figure 2Encoded gray-scale images generated from four different features: (a) distance (b) distance velocity (c) angle, and (d) angle velocity features obtained from the UTD-MHAD dataset. Angle features contain more expressive information as compared to distance features.
Figure 3Architecture of the proposed CNN used in this study.
Figure 4Flowchart of the ALO algorithm used in our proposed DSwarm-Net model.
Specifications of the different datasets used in the present work.
| Dataset | Number of video sequence | Action classes | Subjects | Keyjoints |
|---|---|---|---|---|
| UTD-MHAD[ | 861 | 27 | 8 | 20 |
| HDM05[ | 2337 | 130 | 5 | 31 |
| NTU RGB+D 60[ | 56880 | 60 | 40 | 25 |
Classification performance of the proposed DSwarm-Net model on three benchmark HAR datasets. Acc: Accuracy, F1: F1-Score.
| Encoding Used | UTD MHAD | NTU RGB+D 60 | HDM05 | |||||
|---|---|---|---|---|---|---|---|---|
| Cross-subject | Cross-view | |||||||
| Acc (%) | F1 (%) | Acc (%) | F1 (%) | Acc (%) | F1 (%) | Acc (%) | F1 (%) | |
| Distance encoded | 95.62 | 96 | 84.49 | 85 | 87.24 | 88 | 88.45 | 89 |
| Angle encoded | 96.81 | 97 | 84.92 | 85 | 88.39 | 89 | 89.17 | 89 |
| Distance velocity encoded | 90.23 | 91 | 83.44 | 84 | 87.15 | 88 | 88.34 | 89 |
| Angle velocity encoded | 94.55 | 95 | 84.63 | 85 | 88.15 | 89 | 89.09 | 90 |
| Compact distance encoded | 97.56 | 98 | 84.81 | 85 | 88.92 | 89 | 89.46 | 90 |
| Compact angle encoded | 97.97 | 98 | 84.97 | 85 | 88.66 | 88 | 89.88 | 90 |
Comparison in terms of the number of parameters and the classification accuracy of the proposed model with other CNN models on UTD-MHAD dataset.
| Model | Compact distance encoded features | Compact angle encoded features | ||
|---|---|---|---|---|
| No. of parameter | Accuracy (%) | No. of parameter | Accuracy (%) | |
| VGG19 | 57.9 M | 83.45 | 105.7 M | 88.45 |
| VGG16 | 52.6 M | 85.78 | 97.8 M | 91.21 |
| ResNet101 | 42.7 M | 88.94 | 42.7 M | 94.32 |
| Inception v3 | 24.3 M | 93.28 | 24.3 M | 95.77 |
| DenseNet201 | 18.4 M | 92.64 | 18.4 M | 95.24 |
Parameter settings for the comparative meta-heuristic optimization algorithms. OA denotes Optimization Algorithms.
| OA | Parameter(s) | Value(s) |
|---|---|---|
| MVO[ | Minimum wormhole existence probability ( Maximum wormhole existence probability ( |
|
| PSO[ | Inertia weight ( Acceleration Coefficients ( |
|
| BAT[ | Initial loudness ( Pulse rate ( Minimum frequency ( Maximum frequency ( |
|
| GWO[ | Convergence operator ( Exploration Parameter ( | |
| WOA[ | Random number ( Spiral Updating Probability ( Random Search Ability ( Random Encircling ability ( |
|
| FFA[ | Randomization Parameter, Attractiveness at r=0, Absorption coefficient, | |
| MFO[ | Convergence parameter ( Shape of logarithmic spiral ( Closeness Parameter ( |
|
Comparison of results (accuracies in %) obtained by the different optimization algorithms (OAs) for feature selection (FS) in our DSwarm-Net model and that obtained without any FS, by end-to-end CNN model.
| OAs | UTD-MHAD | NTU RGB+D 60 | HDM05 | |
|---|---|---|---|---|
| Cross subject | Cross view | |||
| Without FS | 89.34 | 79.19 | 80.63 | 81.53 |
| MVO[ | 96.26 | 84.67 | 89.16 | 90.41 |
| PSO[ | 96.47 | 83.85 | 87.58 | 88.56 |
| BAT[ | 95.74 | 84.01 | 88.52 | 87.42 |
| GWO[ | 95.63 | 84.43 | 87.51 | 88.13 |
| WOA[ | 95.26 | 84.22 | 88.13 | 86.87 |
| FFA[ | 95.84 | 84.11 | 87.98 | 86.55 |
| MFO[ | 95.53 | 83.89 | 89.51 | 89.93 |
|
| ||||
Comparison of our DSwarm-Net model with some recent models on the UTD-MHAD dataset by cross-subject analysis.
| Method | Year | Accuracy (%) | Skeleton data | RGB data | Inertial data |
|---|---|---|---|---|---|
| Action machine[ | 2018 | 92.50 | Yes | Yes | No |
| PEM[ | 2018 | 94.51 | Yes | Yes | No |
| BHDM[ | 2019 | 92.80 | Yes | No | No |
| Correlation Congruence[ | 2019 | 94.87 | Yes | Yes | Yes |
| Gimme DSE[ | 2020 | 93.30 | Yes | No | Yes |
| Fuzzy CNN fusion[ | 2020 | 97.91 | Yes | No | No |
| SAKDN[ | 2021 | 98.04 | Yes | Yes | Yes |
| Edge Motion[ | 2021 | 95.59 | Yes | No | No |
| AMGC[ | 2021 | 95.11 | Yes | No | No |
Comparison of our proposed DSwarm-Net model with existing methods on the HDM05 dataset using 10 random split-mean protocol.
| Method | Year | Classification accuracy (%) |
|---|---|---|
| HCN[ | 2018 | 86.51 |
| PB-GCN[ | 2018 | 88.2 |
| Deep STGC[ | 2019 | 85.29 |
| 2S-AGCN[ | 2019 | 88.5 |
| PGCN-TCA[ | 2020 | 86.71 |
| SGCN[ | 2021 | 85.45 |
| Di-StddNet[ | 2021 | 82.32 |
Comparison of our DSwarm-Net model with some existing methods on the NTU RGB+D 60 dataset.
| Method | Year | Cross-subject accuracy (%) | Cross-view accuracy (%) |
|---|---|---|---|
| STVA LSTM[ | 2019 | 82.40 | 89.10 |
| Deep STGC[ | 2019 | 86.45 | 84.65 |
| PC Net[ | 2019 | 85.25 | 91.37 |
| Shift GCN[ | 2020 | 90.70 | 96.5 |
| DS LSTM[ | 2020 | 77.79 | 87.44 |
| AGC-LSTM[ | 2020 | 89.20 | 95.00 |
| PA-ResGCN-B19[ | 2020 | 90.90 | 96.00 |
| MV-IGNet[ | 2020 | 89.2 | 96.3 |
| VIDA[ | 2020 | 79.40 | 84.10 |
| MS-G3D[ | 2020 | 91.5 | 96.2 |
| CTR-GCN[ | 2021 | 92.4 | 96.8 |
| EfficientGCN-B4[ | 2021 | 91.7 | 95.7 |
| ST-TR[ | 2021 | 89.91 | 93.1 |
Results (based on calculated p-values) obtained by performing the McNemar’s test between the ALO algorithm used in this paper and the other popular metaheuristics used for comparison.
| McNemar’s test | UTD-MHAD | NTU RGB+D 60 | HDM05 | |
|---|---|---|---|---|
| Cross subject | Cross view | |||
| ALO vs. MVO | 5.79E−03 | 4.10E−02 | 3.55E−02 | 1.96E−04 |
| ALO vs. PSO | 3.31E−02 | 3.43E−02 | 9.67E−03 | 4.31E−02 |
| ALO vs. BAT | 1.83E−02 | 3.98E−03 | 4.80E−02 | 4.78E−02 |
| ALO vs. GWO | 1.01E−02 | 1.19E−02 | 3.49E−02 | 2.45E−02 |
| ALO vs. WOA | 2.79E−02 | 6.31E−03 | 8.20E−03 | 1.85E−02 |
| ALO vs. FFA | 3.02E−04 | 3.76E−02 | 1.75E−02 | 2.62E−02 |
| ALO vs. MFO | 8.48E−03 | 1.76E−03 | 7.58E−03 | 4.83E−02 |
Comparison of results obtained by performing stability test on selected features from ALO algorithm along with other popular metaheuristics used for comparison.
| OA | UTD-MHAD | NTU RGB+D 60 | HDM05 | |
|---|---|---|---|---|
| Cross subject | Cross view | |||
| MVO | 0.308 | 0.310 | 0.245 | 0.309 |
| PSO | 0.325 | 0.407 | 0.376 | 0.428 |
| BAT | 0.290 | 0.232 | 0.445 | 0.324 |
| GWO | 0.341 | 0.444 | 0.330 | 0.315 |
| WOA | 0.439 | 0.453 | 0.409 | 0.378 |
| FFA | 0.391 | 0.244 | 0.281 | 0.236 |
| MFO | 0.382 | 0.259 | 0.227 | 0.282 |