| Literature DB >> 35154378 |
Muhammad Junaid Ibrahim1, Jaweria Kainat2, Hussain AlSalman3, Syed Sajid Ullah4, Suheer Al-Hadhrami5, Saddam Hussain6.
Abstract
Recent advances in image processing and machine learning methods have greatly enhanced the ability of object classification from images and videos in different applications. Classification of human activities is one of the emerging research areas in the field of computer vision. It can be used in several applications including medical informatics, surveillance, human computer interaction, and task monitoring. In the medical and healthcare field, the classification of patients' activities is important for providing the required information to doctors and physicians for medication reactions and diagnosis. Nowadays, some research approaches to recognize human activity from videos and images have been proposed using machine learning (ML) and soft computational algorithms. However, advanced computer vision methods are still considered promising development directions for developing human activity classification approach from a sequence of video frames. This paper proposes an effective automated approach using feature fusion and ML methods. It consists of five steps, which are the preprocessing, feature extraction, feature selection, feature fusion, and classification steps. Two available public benchmark datasets are utilized to train, validate, and test ML classifiers of the developed approach. The experimental results of this research work show that the accuracies achieved are 99.5% and 99.9% on the first and second datasets, respectively. Compared with many existing related approaches, the proposed approach attained high performance results in terms of sensitivity, accuracy, precision, and specificity evaluation metric.Entities:
Year: 2022 PMID: 35154378 PMCID: PMC8828325 DOI: 10.1155/2022/7931729
Source DB: PubMed Journal: Appl Bionics Biomech ISSN: 1176-2322 Impact factor: 1.781
Figure 1Detailed description of proposed model based on the machine learning methods.
Figure 2Preprocessing stages: (a) original images; (b) background subtraction images; (c) image enhancement; (d) object detection; (e) binary to RGB conversion; (f) image cropping.
Figure 3Visualization of histogram of oriented gradient features: (a) original image; (b) HOG features.
Figure 4Gabor feature visualization: (a) original image; (b) Gabor features.
Figure 5Feature vector selection, fusion, and classification.
Figure 6Sample images of Weizmann dataset.
Figure 7Sample images of KTH dataset.
Summary of experiments setting for Weizmann and KTH datasets.
| Exp no. | No. of classes | Shape | Texture | Color | Folds | |
|---|---|---|---|---|---|---|
| KTH | Weizmann | |||||
| 1 | 6 | 5 | 100 | 60 | 9 | 5 |
| 2 | 6 | 5 | 300 | 60 | 9 | 10 |
| 3 | 6 | 5 | 500 | 60 | 9 | 8 |
| 4 | 6 | 5 | 800 | 58 | 9 | 5 |
| 5 | 6 | 5 | 1100 | 55 | 9 | 7 |
Classification results of experiment 1 with all possible values.
| Weizmann | KTH | |||||||
|---|---|---|---|---|---|---|---|---|
| Method | SEN (%) | SPE (%) | PRE (%) | ACU (%) | SEN (%) | SPE (%) | PRE (%) | ACU (%) |
| Linear-SVM | 98.8 | 99.7 | 98.52 | 98.8 | 99.85 | 99.95 | 99.03 | 99.8 |
| Cubic-SVM | 98.84 | 99.81 | 98.98 | 99.3 | 99.8 | 99.94 | 99.09 | 99.7 |
| Complex tree | 85.92 | 97.33 | 86.06 | 89.0 | 98.26 | 99.67 | 97.77 | 98.4 |
| Fine-KNN | 98.99 | 99.78 | 99.23 | 99.0 | 99.77 | 99.91 | 99.75 | 99.6 |
| Subspace-KNN | 90.3 | 98.30 | 91.7 | 93.3 | 99.86 | 99.96 | 99.74 | 99.8 |
Classification results of experiment 2 along with the sensitivity and other measures.
| Weizmann | KTH | |||||||
|---|---|---|---|---|---|---|---|---|
| Method | SEN (%) | SPE (%) | PRE (%) | ACU (%) | SEN (%) | SPE (%) | PRE (%) | ACU (%) |
| Linear-SVM | 98.83 | 99.84 | 99.0 | 99.3 | 99.89 | 99.95 | 99.89 | 99.8 |
| Cubic-SVM | 99.34 | 99.89 | 99.5 | 99.5 | 98.81 | 99.94 | 98.60 | 99.7 |
| Complex tree | 85.40 | 97.15 | 85.0 | 88.3 | 98.32 | 99.67 | 97.37 | 98.5 |
| Fine-KNN | 87.26 | 97.21 | 95.4 | 91.1 | 99.42 | 99.84 | 99.38 | 99.2 |
| Subspace-KNN | 90.25 | 98.3 | 91.9 | 93.9 | 99.97 | 99.99 | 99.95 | 99.9 |
Confusion matrix of KTH dataset of experiment 2 on subspace-KNN.
| Classification classes | Total images | Clapping | Jogging | Hand waving | Running | Walking | Boxing |
|---|---|---|---|---|---|---|---|
| Clapping | 312 | 312 | |||||
| Jogging | 191 | 191 | |||||
| Hand waving | 581 | 1 | 580 | ||||
| Running | 109 | 109 | |||||
| Walking | 27 | 27 | |||||
| Boxing | 408 | 408 |
Confusion matrix of Weizmann dataset of experiment 2 on cubic-SVM.
| Classification classes | Total images | Hand waving | Running | Jumping | Walking | Bending |
|---|---|---|---|---|---|---|
| Hand waving | 624 | 624 | ||||
| Running | 206 | 201 | 2 | 3 | ||
| Jumping | 421 | 1 | 420 | 1 | ||
| Walking | 271 | 2 | 2 | 267 | ||
| Bending | 375 | 375 |
Classification results of experiment 3 using the linear-SVM method and others.
| Weizmann | KTH | |||||||
|---|---|---|---|---|---|---|---|---|
| Method | SEN (%) | SPE (%) | PRE (%) | ACU (%) | SEN (%) | SPE (%) | PRE (%) | ACU (%) |
| Linear-SVM | 97.90 | 99.69 | 98.39 | 98.7 | 100 | 99.72 | 98.70 | 98.7 |
| Cubic-SVM | 97.86 | 99.67 | 98.544 | 98.7 | 91.51 | 99.21 | 96.39 | 96.5 |
| Complex tree | 85.43 | 97.212 | 85.53 | 88.6 | 98.34 | 99.69 | 97.93 | 98.5 |
| Fine-KNN | 63.7 | 91.63 | 90.79 | 71.9 | 95.30 | 99.32 | 98.02 | 97.0 |
| Subspace-KNN | 89.99 | 98.18 | 92.11 | 93.0 | 99.97 | 99.99 | 99.95 | 99.9 |
Classification results of experiment 4.
| Weizmann | KTH | |||||||
|---|---|---|---|---|---|---|---|---|
| Method | SEN (%) | SPE (%) | PRE (%) | ACU (%) | SEN (%) | SPE (%) | PRE (%) | ACU (%) |
| Linear-SVM | 93.38 | 98.96 | 96.09 | 95.9 | 59.25 | 96.7 | 90.90 | 85.3 |
| Cubic-SVM | 91.56 | 98.5 | 95.63 | 94.5 | 57.56 | 96.45 | 74.49 | 84.3 |
| Complex tree | 80.56 | 97.18 | 85.57 | 88.5 | 97.74 | 99.63 | 97.51 | 98.2 |
| Fine-KNN | 46.29 | 87.54 | 88.81 | 58.2 | 70.14 | 96.30 | 90.90 | 84.0 |
| Subspace-KNN | 89.21 | 98.37 | 91.44 | 92.5 | 99.97 | 99.99 | 99.45 | 99.9 |
Classification results of experiment 5.
| Weizmann | KTH | |||||||
|---|---|---|---|---|---|---|---|---|
| Method | SEN (%) | SPE (%) | PRE (%) | ACU (%) | SEN (%) | SPE (%) | PRE (%) | ACU (%) |
| Linear-SVM | 80.0 | 96.68 | 90.0 | 87.5 | 50.83 | 95.49 | 57.16 | 80.2 |
| Cubic-SVM | 80.0 | 96.20 | 90.31 | 86.3 | 50.42 | 95.37 | 55.03 | 79.5 |
| Complex tree | 85.32 | 97.8 | 85.31 | 88.4 | 98.49 | 99.69 | 97.93 | 98.6 |
| Fine-KNN | 33.42 | 84.1 | 87.63 | 46.7 | 39.47 | 93.6 | 84.92 | 57.6 |
| Subspace-KNN | 88.72 | 97.96 | 90.81 | 92.0 | 99.97 | 99.99 | 99.95 | 99.9 |
AUC results.
| Method | Classes | Experiment 1 | Experiment 2 | Experiment 3 | Experiment 4 | Experiment 5 | |||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| Weizmann | KTH | Weizmann | KTH | Weizmann | KTH | Weizmann | KTH | Weizmann | KTH | ||
| Linear-SVM | C1 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 |
| Cubic-SVM | C1 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 |
| Complex tree | C1 | 0.98 | 1.00 | 0.99 | 1.00 | 0.98 | 1.00 | 0.98 | 0.99 | 0.98 | 0.99 |
| Fine-KNN | C1 | 1.00 | 0.98 | 0.93 | 0.97 | 0.79 | 0.98 | 0.69 | 0.95 | 0.60 | 0.75 |
| Subspace-KNN | C1 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 | 1.00 |
Comparison of action recognition results.
| Reference | Year | Recognition (%) |
|---|---|---|
|
| ||
| [ | 2013 | 95.45 |
| [ | 2014 | 95.56 |
| [ | 2015 | 95.10 |
| [ | 2016 | 88.10 |
| [ | 2017 | 95.80 |
| [ | 2019 | 99.0 |
| [ | 2020 | 96.0 |
| Proposed | 2018 | 99.5 |
|
| ||
| [ | 2014 | 95.0 |
| [ | 2015 | 95.21 |
| [ | 2015 | 96.50 |
| [ | 2016 | 97.10 |
| [ | 2017 | 94.92 |
| [ | 2017 | 99.30 |
| [ | 2020 | 94.83 |
| [ | 2019 | 91.67 |
| Proposed | 2018 | 99.90 |