| Literature DB >> 27399696 |
Hiram Ponce1, María de Lourdes Martínez-Villaseñor2, Luis Miralles-Pechuán3.
Abstract
Human activity recognition has gained more interest in several research communities given that understanding user activities and behavior helps to deliver proactive and personalized services. There are many examples of health systems improved by human activity recognition. Nevertheless, the human activity recognition classification process is not an easy task. Different types of noise in wearable sensors data frequently hamper the human activity recognition classification process. In order to develop a successful activity recognition system, it is necessary to use stable and robust machine learning techniques capable of dealing with noisy data. In this paper, we presented the artificial hydrocarbon networks (AHN) technique to the human activity recognition community. Our artificial hydrocarbon networks novel approach is suitable for physical activity recognition, noise tolerance of corrupted data sensors and robust in terms of different issues on data sensors. We proved that the AHN classifier is very competitive for physical activity recognition and is very robust in comparison with other well-known machine learning methods.Entities:
Keywords: artificial hydrocarbon networks; artificial organic networks; noise tolerance; robust human activity recognition; supervised machine learning; wearable sensors
Year: 2016 PMID: 27399696 PMCID: PMC4970082 DOI: 10.3390/s16071033
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Framework of artificial organic networks.
| Framework Level | Description |
|---|---|
| implementation | |
| mathematical model | |
| chemical heuristic rules | |
| interactions | |
| types of components |
Figure 1Structure of an artificial hydrocarbon network using saturated and linear chains of molecules [29]. For this work, the topology of the proposed classifier considers just one hydrocarbon compound (see Section 4).
Figure 2Diagram of the proposed artificial hydrocarbon network-based classifier (AHN classifier). First, data from sensors and activity labeling are used for training the AHN-model, then it is used as the AHN classifier in the testing step.
Figure 3Location of the three wearable sensors used in the dataset (hand, chest and ankle), adapted from [21].
Physical activities identified in this case study, adapted from [13].
| No. | Performed Activities | Activity Description |
|---|---|---|
| 1 | Lying | This movement is lying flat, slightly changing position or stretching a little bit. |
| 2 | Sitting | Refers to sitting in a chair in any posture. It also includes more comfortable positions as leaning or crossing your legs. |
| 3 | Standing | This position includes the natural movements of a person who is standing, swaying slightly, gesturing or talking. |
| 4 | Walking | This activity is a stroll down the street at a moderate speed of approximately 5 km/h. |
| 5 | Running | The people who made this activity ran at a moderate speed; taking into account non-high level athletes. |
| 6 | Cycling | A bicycle was used for this movement, and people pedaled as on a quiet ride. An activity requiring great effort was not requested. |
| 7 | Nordic walking | For this activity, it was required that persons that were inexperienced walked on asphalt using pads. |
| 8 | Watching TV | This position includes the typical movements of someone who is watching TV and changes the channel, lying on one side or stretching his or her legs. |
| 9 | Computer work | The typical movements of someone who works with a computer: mouse movement, movement of neck, etc. |
| 10 | Car driving | All movements necessary to move from the office to the house for testing sensors. |
| 11 | Ascending stairs | During this activity, the necessary movements up to a distance of five floors were recorded; from the ground floor to the fifth floor. |
| 12 | Descending stairs | This movement is the opposite of the former. Instead of climbing the stairs, the activity of descending them was recorded. |
| 13 | Vacuum cleaning | Refers to all of the activities necessary to clean a floor of the office. It also includes moving objects, such as rugs, chairs and wardrobes. |
| 14 | Ironing | It covers the necessary movements to iron a shirt or a t-shirt. |
| 15 | Folding laundry | It consists of folding clothes, such as shirts, pants and socks. |
| 16 | House cleaning | These are the movements that a person makes while cleaning a house; such as moving chairs to clean the floor, throwing things away, bending over to pick up something, etc. |
| 17 | Playing soccer | In this activity, individuals are negotiating, running the ball, shooting a goal or trying to stop the ball from the goal. |
| 18 | Rope jumping | There are people who prefer to jump with both feet together, and there are others who prefer to move one foot first and then the other. |
Configuration parameters of supervised models employed with the caret package in R.
| No. | Method Name | Parameters | Configurations | ||
|---|---|---|---|---|---|
| 1 | AdaBoost | size, decay, bag | (150, 3, 3) | (150, 3, 3) | 27 |
| 2 | Artificial Hydrocarbon Networks | molecules, eta, epsilon | (18, 0.1, 0.0001) | (18, 0.1, 0.0001) | 1 |
| 3 | C4.5 Decision Trees | c | (0.25) | (0.25) | 1 |
| 4 | kmax, distance, kernel | (9, 2, 1) | (9, 2, 1) | 3 | |
| 5 | Linear Discriminant Analysis | – | – | – | 1 |
| 6 | Mixtures Discriminant Analysis | subclasses | (4) | (4) | 3 |
| 7 | Multivariate Adaptive Regression Splines | degree | (1) | (1) | 1 |
| 8 | Naive Bayes | fl, use_kernel | (0, true) | (0, true) | 2 |
| 9 | Nearest Shrunken Centroids | threshold | (2.512) | (1.363) | 3 |
| 10 | Artificial Neural Networks | size, decay | (5, 0) | (5, 0) | 9 |
| 11 | Random Forest | mtry | (26) | (6) | 3 |
| 12 | Rule-Based Classifier | threshold, pruned | (0.25, true) | (0.25, true) | 1 |
| 13 | Stochastic Gradient Boosting | n.trees, depth, shrinkage | (150, 3, 0.1) | (150, 2, 0.1) | 9 |
| 14 | SVM with Linear Kernel | c | (1) | (1) | 1 |
| 15 | SVM with Radial Basis Function Kernel | sigma, c | (0.0179, 1) | (0.0748, 1) | 3 |
Retained features using the recursive feature elimination (RFE) technique.
| Sensor | Features Selected | Feature Number |
|---|---|---|
| temperature | 4 | |
| temperature | 21 | |
| z-axis 3D-accelerometer 16 g | 24 | |
| z-axis 3D-accelerometer 6 g | 27 | |
| y-axis 3D-magnetometer | 32 | |
| z-axis 3D-magnetometer | 33 | |
| temperature | 38 | |
| x-axis 3D-magnetometer | 48 | |
| z-axis 3D-magnetometer | 50 | |
| first orientation | 51 |
Figure 4Average accuracy of models with respect to the number of features selected using RFE.
Figure 5Methodology of experimentation.
Summary of activities performed by each subject for at least 30 s.
| Activities | Sub 1 | Sub 2 | Sub 3 | Sub 4 | Sub 5 | Sub 6 | Sub 7 | Sub 8 | Sub 9 | Total |
|---|---|---|---|---|---|---|---|---|---|---|
| Lying | - | - | - | - | - | - | - | - | 8 | |
| Sitting | - | - | - | - | - | - | - | - | 8 | |
| Standing | - | - | - | - | - | - | - | - | 8 | |
| Walking | - | - | - | - | - | - | - | - | 8 | |
| Running | - | - | - | - | - | 5 | ||||
| Cycling | - | - | - | - | - | - | - | 7 | ||
| Nordic walking | - | - | - | - | - | - | - | 7 | ||
| Watching TV | - | 1 | ||||||||
| Computer work | - | - | - | - | 4 | |||||
| Car driving | - | 1 | ||||||||
| Ascending stairs | - | - | - | - | - | - | - | - | 8 | |
| Descending stairs | - | - | - | - | - | - | - | - | 8 | |
| Vacuum cleaning | - | - | - | - | - | - | - | - | 8 | |
| Ironing | - | - | - | - | - | - | - | - | 8 | |
| Folding laundry | - | - | - | - | 4 | |||||
| House cleaning | - | - | - | - | - | 5 | ||||
| - | - | 2 | ||||||||
| Rope jumping | - | - | - | - | - | 5 |
Comparison of the supervised classifiers using the complete dataset of the case study.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 1 | C4.5 Decision Trees | 0.9880 | 0.9994 | 0.9893 | 0.9943 | 0.9886 | 0.9885 | 0.9938 |
| 2 | Rule-Based Classifier | 0.9876 | 0.9993 | 0.9890 | 0.9942 | 0.9883 | 0.9882 | 0.9937 |
| 4 | SVM with Linear Kernel | 0.9827 | 0.9989 | 0.9828 | 0.9908 | 0.9833 | 0.9834 | 0.9911 |
| 5 | SVM with Radial Basis Function Kernel | 0.9745 | 0.9985 | 0.9752 | 0.9867 | 0.9746 | 0.9747 | 0.9865 |
| 6 | Random Forest | 0.9727 | 0.9986 | 0.9819 | 0.9902 | 0.9743 | 0.9797 | 0.9890 |
| 7 | Stochastic Gradient Boosting | 0.9725 | 0.9725 | 0.9805 | 0.9893 | 0.9740 | 0.9815 | 0.9899 |
| 8 | k-Nearest Neighbors | 0.9718 | 0.9984 | 0.9724 | 0.9852 | 0.9711 | 0.9717 | 0.9849 |
| 9 | Mixture Discriminant Analysis | 0.9714 | 0.9982 | 0.9717 | 0.9848 | 0.9724 | 0.9729 | 0.9854 |
| 10 | AdaBoost | 0.9710 | 0.9982 | 0.9774 | 0.9877 | 0.9726 | 0.9780 | 0.9880 |
| 11 | Multivariate Adaptive Regression Splines | 0.9553 | 0.9974 | 0.9621 | 0.9794 | 0.9572 | 0.9609 | 0.9788 |
| 12 | Linear Discriminant Analysis | 0.9382 | 0.9962 | 0.9398 | 0.9672 | 0.9399 | 0.9418 | 0.9683 |
| 13 | Naive Bayes | 0.9327 | 0.9962 | 0.9459 | 0.9704 | 0.9350 | 0.9438 | 0.9692 |
| 14 | Artificial Neural Networks | 0.8976 | 0.9939 | 0.9019 | 0.9457 | 0.8984 | 0.9020 | 0.9458 |
| 15 | Nearest Shrunken Centroids | 0.7031 | 0.9820 | 0.7006 | 0.8178 | 0.7073 | 0.7063 | 0.8218 |
Comparison of the supervised classifiers using the reduced dataset of the case study.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 1 | Stochastic Gradient Boosting | 0.9898 | 0.9898 | 0.9907 | 0.9950 | 0.9900 | 0.9900 | 0.9947 |
| 3 | AdaBoost | 0.9657 | 0.9979 | 0.9674 | 0.9824 | 0.9675 | 0.9686 | 0.9831 |
| 4 | Random Forest | 0.9655 | 0.9982 | 0.9768 | 0.9874 | 0.9673 | 0.9744 | 0.9861 |
| 5 | Rule-Based Classifier | 0.9604 | 0.9979 | 0.9716 | 0.9846 | 0.9617 | 0.9693 | 0.9833 |
| 6 | C4.5 Decision Trees | 0.9571 | 0.9976 | 0.9647 | 0.9809 | 0.9586 | 0.9629 | 0.9799 |
| 7 | k-Nearest Neighbors | 0.7222 | 0.9834 | 0.7217 | 0.8325 | 0.7291 | 0.7255 | 0.8351 |
| 8 | Multivariate Adaptive Regression Splines | 0.7061 | 0.9825 | 0.7357 | 0.8413 | 0.7159 | 0.7392 | 0.8437 |
| 9 | SVM with Radial Basis Function Kernel | 0.6549 | 0.9791 | 0.6526 | 0.7832 | 0.6619 | 0.6601 | 0.7887 |
| 10 | Naive Bayes | 0.6069 | 0.9758 | 0.6506 | 0.7807 | 0.6125 | 0.6615 | 0.7888 |
| 11 | Mixture Discriminant Analysis | 0.5853 | 0.9750 | 0.5684 | 0.7181 | 0.5929 | 0.5746 | 0.7232 |
| 12 | SVM with Linear Kernel | 0.5386 | 0.9720 | 0.5196 | 0.6772 | 0.5487 | 0.5296 | 0.6858 |
| 13 | Artificial Neural Networks | 0.5092 | 0.9703 | 0.4631 | 0.6270 | 0.5190 | 0.4716 | 0.6349 |
| 14 | Linear Discriminant Analysis | 0.4792 | 0.9683 | 0.4676 | 0.6307 | 0.4853 | 0.4773 | 0.6396 |
| 15 | Nearest Shrunken Centroids | 0.4216 | 0.9645 | 0.4131 | 0.5784 | 0.4191 | 0.4209 | 0.5863 |
Comparison of the supervised classifiers using the 7% noisy dataset of the case study.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 1 | Random Forest | 0.9825 | 0.9990 | 0.9830 | 0.9909 | 0.9833 | 0.9830 | 0.9909 |
| 2 | Stochastic Gradient Boosting | 0.9722 | 0.9722 | 0.9754 | 0.9868 | 0.9733 | 0.9747 | 0.9864 |
| 4 | Rule-Based Classifier | 0.9484 | 0.9970 | 0.9532 | 0.9747 | 0.9497 | 0.9521 | 0.9740 |
| 5 | C4.5 Decision Trees | 0.9459 | 0.9969 | 0.9519 | 0.9739 | 0.9478 | 0.9507 | 0.9732 |
| 6 | SVM with Radial Basis Function Kernel | 0.9276 | 0.9957 | 0.9281 | 0.9607 | 0.9278 | 0.9285 | 0.9609 |
| 7 | AdaBoost | 0.9151 | 0.9949 | 0.9242 | 0.9582 | 0.9181 | 0.9256 | 0.9590 |
| 8 | 0.9059 | 0.9944 | 0.9066 | 0.9484 | 0.9060 | 0.9075 | 0.9490 | |
| 9 | Mixture Discriminant Analysis | 0.8816 | 0.9928 | 0.8830 | 0.9347 | 0.8831 | 0.8856 | 0.9362 |
| 10 | SVM with Linear Kernel | 0.8796 | 0.9928 | 0.8786 | 0.9322 | 0.8800 | 0.8792 | 0.9326 |
| 11 | Multivariate Adaptive Regression Splines | 0.8755 | 0.9924 | 0.9055 | 0.9469 | 0.8759 | 0.9083 | 0.9486 |
| 12 | Linear Discriminant Analysis | 0.8339 | 0.9899 | 0.8367 | 0.9069 | 0.8353 | 0.8398 | 0.9088 |
| 13 | Artificial Neural Networks | 0.7820 | 0.9870 | 0.7764 | 0.8691 | 0.7842 | 0.7782 | 0.8703 |
| 14 | Naive Bayes | 0.7769 | 0.9868 | 0.7957 | 0.8810 | 0.7811 | 0.7964 | 0.8815 |
| 15 | Nearest Shrunken Centroids | 0.6202 | 0.9771 | 0.6076 | 0.7493 | 0.6235 | 0.6125 | 0.7531 |
Comparison of the supervised classifiers using the 15% noisy dataset of the case study.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 2 | 0.9425 | 0.9425 | 0.9433 | 0.9692 | 0.9418 | 0.9433 | 0.9692 | |
| 3 | SVM with Radial Basis Function Kernel | 0.9249 | 0.9249 | 0.9308 | 0.9621 | 0.9238 | 0.9314 | 0.9624 |
| 4 | Naive Bayes | 0.9182 | 0.9182 | 0.9343 | 0.9638 | 0.9209 | 0.9328 | 0.9630 |
| 5 | Stochastic Gradient Boosting | 0.8402 | 0.8402 | 0.8729 | 0.9278 | 0.8387 | 0.8767 | 0.9302 |
| 6 | SVM with Linear Kernel | 0.8400 | 0.8400 | 0.8550 | 0.9177 | 0.8394 | 0.8579 | 0.9195 |
| 7 | Mixture Discriminant Analysis | 0.7975 | 0.7975 | 0.8113 | 0.8909 | 0.7986 | 0.8148 | 0.8931 |
| 8 | AdaBoost | 0.7755 | 0.7755 | 0.8209 | 0.8960 | 0.7724 | 0.8252 | 0.8988 |
| 9 | Random Forest | 0.7498 | 0.7498 | 0.8263 | 0.8986 | 0.7477 | 0.8313 | 0.9017 |
| 10 | Linear Discriminant Analysis | 0.7478 | 0.7478 | 0.7668 | 0.8622 | 0.7489 | 0.7710 | 0.8650 |
| 11 | Rule-Based Classifier | 0.7414 | 0.7414 | 0.8039 | 0.8851 | 0.7370 | 0.8043 | 0.8855 |
| 12 | C4.5 Decision Trees | 0.7335 | 0.7335 | 0.7984 | 0.8815 | 0.7290 | 0.8014 | 0.8835 |
| 13 | Multivariate Adaptive Regression Splines | 0.6824 | 0.6824 | 0.7610 | 0.8569 | 0.6791 | 0.7684 | 0.8619 |
| 14 | Nearest Shrunken Centroids | 0.6761 | 0.6761 | 0.6820 | 0.8044 | 0.6809 | 0.6884 | 0.8090 |
| 15 | Artificial Neural Networks | 0.5214 | 0.5214 | 0.5339 | 0.6892 | 0.5253 | 0.5307 | 0.6865 |
Comparison of the supervised classifiers using the 30% noisy dataset of the case study.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 1 | Naive Bayes | 0.8798 | 0.8798 | 0.9090 | 0.9491 | 0.8816 | 0.9087 | 0.9489 |
| 3 | 0.8608 | 0.8608 | 0.8681 | 0.9257 | 0.8604 | 0.8708 | 0.9273 | |
| 4 | SVM with Radial Basis Function Kernel | 0.8043 | 0.8043 | 0.8349 | 0.9051 | 0.8024 | 0.8374 | 0.9067 |
| 5 | Stochastic Gradient Boosting | 0.6967 | 0.6967 | 0.7822 | 0.8705 | 0.6950 | 0.7905 | 0.8759 |
| 6 | SVM with Linear Kernel | 0.6747 | 0.6747 | 0.7365 | 0.8410 | 0.6750 | 0.7419 | 0.8448 |
| 7 | Nearest Shrunken Centroids | 0.6302 | 0.6302 | 0.6605 | 0.7883 | 0.6354 | 0.6675 | 0.7935 |
| 8 | AdaBoost | 0.6302 | 0.6302 | 0.7315 | 0.8367 | 0.6249 | 0.7392 | 0.8421 |
| 9 | Mixture Discriminant Analysis | 0.6098 | 0.6098 | 0.6525 | 0.7823 | 0.6116 | 0.6579 | 0.7863 |
| 10 | Random Forest | 0.5514 | 0.5514 | 0.7378 | 0.8390 | 0.5474 | 0.7463 | 0.8449 |
| 11 | Rule-Based Classifier | 0.5496 | 0.5496 | 0.6913 | 0.8082 | 0.5446 | 0.6926 | 0.8093 |
| 12 | Linear Discriminant Analysis | 0.5449 | 0.5449 | 0.5988 | 0.7412 | 0.5483 | 0.6050 | 0.7461 |
| 13 | C4.5 Decision Trees | 0.5308 | 0.5308 | 0.6693 | 0.7925 | 0.5255 | 0.6760 | 0.7976 |
| 14 | Multivariate Adaptive Regression Splines | 0.4688 | 0.4688 | 0.6439 | 0.7731 | 0.4634 | 0.6548 | 0.7814 |
| 15 | Artificial Neural Networks | 0.4559 | 0.4559 | 0.5271 | 0.6824 | 0.4563 | 0.5295 | 0.6845 |
Comparison of the supervised classifiers using the first 30 s of each activity carried out by all subjects.
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 1 | Stochastic Gradient Boosting | 0.9952 | 0.9952 | 0.9996 | 0.9974 | 0.9963 | 0.9962 | 0.9980 |
| 2 | Random Forest | 0.9951 | 0.9951 | 0.9996 | 0.9974 | 0.9963 | 0.9959 | 0.9978 |
| 3 | Rule-Based Classifier | 0.9866 | 0.9866 | 0.9990 | 0.9928 | 0.9894 | 0.9857 | 0.9924 |
| 4 | ||||||||
| 5 | SVM with Radial Basis Function Kernel | 0.9635 | 0.9635 | 0.9973 | 0.9804 | 0.9698 | 0.9662 | 0.9818 |
| 6 | C4.5 Decision Trees | 0.9630 | 0.9630 | 0.9973 | 0.9797 | 0.9709 | 0.9652 | 0.9812 |
| 7 | Multivariate Adaptive Regression Splines | 0.9574 | 0.9574 | 0.9966 | 0.9778 | 0.9669 | 0.9673 | 0.9822 |
| 8 | AdaBoost | 0.9474 | 0.9474 | 0.9959 | 0.9736 | 0.9498 | 0.9621 | 0.9792 |
| 9 | SVM with Linear Kernel | 0.9441 | 0.9441 | 0.9958 | 0.9690 | 0.9550 | 0.9508 | 0.9732 |
| 10 | 0.9141 | 0.9140 | 0.9942 | 0.9530 | 0.9183 | 0.9122 | 0.9518 | |
| 11 | Naive Bayes | 0.9093 | 0.9093 | 0.9939 | 0.9518 | 0.9208 | 0.9150 | 0.9532 |
| 12 | Artificial Neural Networks | 0.8879 | 0.8879 | 0.9923 | 0.9388 | 0.9032 | 0.8909 | 0.9393 |
| 13 | Mixture Discriminant Analysis | 0.8868 | 0.8868 | 0.9922 | 0.9379 | 0.9026 | 0.8914 | 0.9396 |
| 14 | Linear Discriminant Analysis | 0.7748 | 0.7748 | 0.9847 | 0.8700 | 0.7941 | 0.7797 | 0.8711 |
| 15 | Nearest Shrunken Centroids | 0.6199 | 0.6199 | 0.9727 | 0.7350 | 0.6285 | 0.5972 | 0.7414 |
Confusion matrix of the AHN classifier using the first 30 s of each activity carried out by all subjects.
| Actual Values | |||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Lying | Sitting | Standing | Walking | Running | Cycling | Nordic walking | Watching TV | Computer Work | Car Driving | Ascending Stairs | Descending Stairs | Vacuum Cleaning | Ironing | Folding laundry | House Cleaning | Playing Soccer | Rope Jumping | ||
| Lying | 19 | 1 | 22 | 30 | 7 | 13 | 3 | 0 | 6 | 0 | 10 | 0 | 0 | 0 | 0 | 3 | 0 | ||
| Sitting | 13 | 6 | 22 | 8 | 1 | 1 | 0 | 0 | 0 | 0 | 4 | 0 | 0 | 0 | 0 | 0 | 1 | ||
| Standing | 0 | 20 | 30 | 10 | 5 | 11 | 0 | 0 | 0 | 1 | 4 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Walking | 2 | 109 | 48 | 19 | 16 | 38 | 0 | 0 | 0 | 7 | 8 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Running | 0 | 133 | 47 | 70 | 14 | 78 | 0 | 0 | 0 | 7 | 11 | 1 | 0 | 0 | 0 | 1 | 0 | ||
| Cycling | 0 | 47 | 75 | 85 | 19 | 66 | 0 | 0 | 0 | 10 | 21 | 1 | 0 | 0 | 0 | 1 | 1 | ||
| Nordic walking | 0 | 74 | 129 | 134 | 32 | 120 | 0 | 65 | 3 | 43 | 41 | 14 | 2 | 0 | 7 | 3 | 1 | ||
| Watching TV | 0 | 4 | 69 | 31 | 26 | 73 | 35 | 35 | 10 | 27 | 29 | 17 | 2 | 0 | 19 | 1 | 1 | ||
| Computer work | 0 | 4 | 18 | 17 | 21 | 51 | 21 | 33 | 7 | 37 | 31 | 31 | 9 | 3 | 18 | 2 | 0 | ||
| Car driving | 0 | 7 | 12 | 6 | 22 | 37 | 14 | 0 | 37 | 49 | 50 | 41 | 20 | 1 | 27 | 4 | 2 | ||
| Ascending stairs | 0 | 1 | 6 | 1 | 10 | 21 | 19 | 0 | 36 | 4 | 55 | 66 | 42 | 9 | 11 | 6 | 2 | ||
| Predicted values | Descending stairs | 0 | 0 | 2 | 3 | 10 | 10 | 10 | 0 | 17 | 1 | 64 | 72 | 72 | 12 | 33 | 8 | 8 | |
| Vacuum cleaning | 0 | 0 | 17 | 3 | 17 | 12 | 13 | 0 | 20 | 2 | 90 | 103 | 190 | 70 | 90 | 30 | 40 | ||
| Ironing | 0 | 7 | 26 | 0 | 4 | 0 | 0 | 0 | 0 | 0 | 7 | 16 | 30 | 34 | 34 | 8 | 20 | ||
| Folding laundry | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 6 | 20 | 24 | 41 | 27 | 15 | 15 | ||
| House cleaning | 0 | 0 | 0 | 0 | 3 | 0 | 0 | 0 | 0 | 0 | 2 | 8 | 14 | 19 | 22 | 11 | 14 | ||
| Playing soccer | 0 | 0 | 0 | 0 | 4 | 0 | 0 | 0 | 0 | 0 | 6 | 12 | 5 | 9 | 28 | 32 | 69 | ||
| Rope jumping | 0 | 8 | 2 | 8 | 4 | 17 | 5 | 3 | 0 | 10 | 6 | 8 | 0 | 2 | 1 | 12 | 24 | ||
Comparison of the supervised classifiers using a majority voting across windows-based approach (-s window size).
| No. | Method Name | Accuracy | Sensitivity | Precision | Sensitivity | Precision | ||
|---|---|---|---|---|---|---|---|---|
| 2 | Rule-Based Classifier | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
| 3 | C4.5 Decision Trees | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
| 4 | Random Forest | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
| 5 | Stochastic Gradient Boosting | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 |
| 6 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | |
| 7 | SVM with Radial Basis Function Kernel | 0.9841 | 0.9841 | 0.9846 | 0.9916 | 0.9873 | 0.9886 | 0.9938 |
| 8 | SVM with Linear Kernel | 0.9786 | 0.9786 | 0.9791 | 0.9887 | 0.9843 | 0.9821 | 0.9904 |
| 9 | Multivariate Adaptive Regression Splines | 0.9770 | 0.9770 | 0.9785 | 0.9882 | 0.9832 | 0.9844 | 0.9914 |
| 10 | AdaBoost | 0.9627 | 0.9627 | 0.9651 | 0.9808 | 0.9681 | 0.9732 | 0.9853 |
| 11 | Naive Bayes | 0.9571 | 0.9571 | 0.9604 | 0.9785 | 0.9651 | 0.9603 | 0.9785 |
| 12 | Artificial Neural Networks | 0.9460 | 0.9460 | 0.9478 | 0.9714 | 0.9570 | 0.9481 | 0.9718 |
| 13 | Mixture Discriminant Analysis | 0.9365 | 0.9365 | 0.9412 | 0.9677 | 0.9500 | 0.9421 | 0.9684 |
| 14 | Linear Discriminant Analysis | 0.8444 | 0.8444 | 0.8508 | 0.9148 | 0.8583 | 0.8528 | 0.9166 |
| 15 | Nearest Shrunken Centroids | 0.6857 | 0.6857 | 0.6593 | 0.7872 | 0.7018 | 0.6773 | 0.8014 |
Confusion matrix of the artificial hydrocarbon networks (AHN) classifier using a majority voting across windows-based approach (-s window size).
| Actual Values | |||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Lying | Sitting | Standing | Walking | Running | Cycling | Nordic Walking | Watching TV | Computer Work | Car Driving | Ascending Stairs | Descending Stairs | Vacuum Cleaning | Ironing | Folding Laundry | House Cleaning | Playing Soccer | Rope Jumping | ||
| Lying | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Sitting | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Standing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Walking | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Running | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Cycling | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Nordic walking | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Watching TV | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Computer work | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Car driving | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Ascending stairs | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Predicted values | Descending stairs | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | |
| Vacuum cleaning | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Ironing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Folding laundry | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| House cleaning | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Playing soccer | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Rope jumping | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
Figure 6Learning curve of the AHN classifier for the windowing-based approach.
Confusion matrix of the AHN classifier using the 7% noisy dataset.
| Actual Values | |||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Lying | Sitting | Standing | Walking | Running | Cycling | Nordic Walking | Watching TV | Computer Work | Car Driving | Ascending Stairs | Descending Stairs | Vacuum Cleaning | Ironing | Folding Laundry | House Cleaning | Playing Soccer | Rope Jumping | ||
| Lying | 3 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | ||
| Sitting | 0 | 1 | 0 | 2 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Standing | 0 | 3 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Walking | 1 | 3 | 0 | 2 | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | ||
| Running | 0 | 3 | 1 | 0 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | ||
| Cycling | 0 | 1 | 0 | 2 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | ||
| Nordic walking | 0 | 2 | 3 | 3 | 2 | 4 | 3 | 3 | 1 | 5 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | ||
| Watching TV | 0 | 0 | 0 | 0 | 0 | 1 | 4 | 4 | 2 | 0 | 1 | 0 | 0 | 0 | 1 | 0 | 0 | ||
| Computer work | 0 | 1 | 2 | 1 | 1 | 0 | 0 | 1 | 3 | 1 | 2 | 0 | 0 | 1 | 1 | 0 | 0 | ||
| Car driving | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 2 | 2 | 0 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | ||
| Ascending stairs | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 1 | 1 | 0 | 2 | 1 | 0 | 0 | 0 | ||
| Predicted values | Descending stairs | 0 | 0 | 2 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | |
| Vacuum cleaning | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 2 | 1 | 2 | 1 | 3 | 2 | 0 | ||
| Ironing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | ||
| Folding laundry | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 2 | 0 | 0 | ||
| House cleaning | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 2 | 4 | 1 | ||
| Playing soccer | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 1 | 3 | 3 | ||
| Rope jumping | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 2 | ||
Confusion matrix of the AHN classifier using the 15% noisy dataset.
| Actual Values | |||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Lying | Sitting | Standing | Walking | Running | Cycling | Nordic Walking | Watching TV | Computer Work | Car Driving | Ascending Stairs | Descending Stairs | Vacuum Cleaning | Ironing | Folding Laundry | House Cleaning | Playing Soccer | Rope Jumping | ||
| Lying | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | ||
| Sitting | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Standing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Walking | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Running | 0 | 0 | 0 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | 0 | ||
| Cycling | 0 | 0 | 2 | 1 | 1 | 1 | 8 | 0 | 0 | 1 | 2 | 0 | 2 | 0 | 2 | 0 | 0 | ||
| Nordic walking | 6 | 12 | 4 | 9 | 6 | 7 | 6 | 3 | 1 | 6 | 11 | 5 | 12 | 2 | 5 | 0 | 0 | ||
| Watching TV | 0 | 0 | 1 | 1 | 2 | 3 | 0 | 6 | 0 | 2 | 0 | 2 | 0 | 3 | 1 | 0 | 1 | ||
| Computer work | 0 | 1 | 3 | 1 | 2 | 0 | 2 | 0 | 4 | 1 | 1 | 0 | 1 | 3 | 1 | 0 | 0 | ||
| Car driving | 4 | 0 | 2 | 3 | 0 | 2 | 3 | 0 | 2 | 3 | 0 | 2 | 1 | 2 | 2 | 0 | 0 | ||
| Ascending stairs | 1 | 1 | 3 | 0 | 3 | 1 | 0 | 0 | 1 | 1 | 3 | 2 | 0 | 2 | 2 | 0 | 0 | ||
| Predicted values | Descending stairs | 0 | 0 | 2 | 0 | 2 | 1 | 0 | 0 | 0 | 2 | 0 | 0 | 0 | 2 | 1 | 0 | 0 | |
| Vacuum cleaning | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | ||
| Ironing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 1 | ||
| Folding laundry | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 0 | ||
| House cleaning | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Playing soccer | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 5 | ||
| Rope jumping | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 5 | ||
Confusion matrix of the AHN classifier using the 30% noisy dataset.
| Actual Values | |||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Lying | Sitting | Standing | Walking | Running | Cycling | Nordic Walking | Watching TV | Computer Work | Car Driving | Ascending Stairs | Descending Stairs | Vacuum Cleaning | Ironing | Folding Laundry | House Cleaning | Playing Soccer | Rope Jumping | ||
| Lying | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 2 | 0 | ||
| Sitting | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Standing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Walking | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Running | 0 | 2 | 0 | 0 | 0 | 0 | 0 | 2 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 0 | ||
| Cycling | 3 | 4 | 1 | 1 | 1 | 2 | 9 | 1 | 3 | 2 | 0 | 1 | 5 | 3 | 0 | 1 | 0 | ||
| Nordic walking | 17 | 22 | 20 | 26 | 19 | 16 | 20 | 6 | 3 | 16 | 17 | 19 | 31 | 8 | 10 | 1 | 0 | ||
| Watching TV | 1 | 0 | 13 | 1 | 5 | 5 | 8 | 2 | 0 | 2 | 2 | 5 | 4 | 2 | 3 | 0 | 0 | ||
| Computer work | 1 | 0 | 4 | 2 | 2 | 5 | 5 | 0 | 4 | 3 | 2 | 9 | 0 | 5 | 3 | 0 | 0 | ||
| Car driving | 7 | 2 | 3 | 10 | 5 | 6 | 3 | 0 | 9 | 4 | 6 | 4 | 2 | 9 | 2 | 1 | 1 | ||
| Ascending stairs | 5 | 1 | 13 | 5 | 9 | 10 | 4 | 0 | 4 | 9 | 5 | 7 | 1 | 5 | 6 | 1 | 0 | ||
| Predicted values | Descending stairs | 2 | 1 | 2 | 1 | 4 | 3 | 0 | 0 | 3 | 0 | 1 | 1 | 0 | 4 | 1 | 1 | 0 | |
| Vacuum cleaning | 1 | 0 | 0 | 1 | 1 | 2 | 0 | 0 | 2 | 0 | 3 | 2 | 1 | 1 | 0 | 7 | 1 | ||
| Ironing | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | ||
| Folding laundry | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 1 | 2 | ||
| House cleaning | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 2 | 1 | ||
| Playing soccer | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 6 | ||
| Rope jumping | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 0 | 8 | ||
Overall performance of supervised classifiers during the experiments with respect to the accuracy metric.
| No. | Method Name | Complete Dataset | 7% Noisy Dataset | Reduced Dataset | ||
|---|---|---|---|---|---|---|
| 1 | Stochastic Gradient Boosting | 0.9725 | 0.9722 | 0.9898 | 0.9782 | 0.0082 |
| 3 | Random Forest | 0.9727 | 0.9825 | 0.9655 | 0.9736 | 0.0070 |
| 4 | Rule-Based Classifier | 0.9876 | 0.9484 | 0.9604 | 0.9655 | 0.0164 |
| 5 | C4.5 Decision Trees | 0.9880 | 0.9459 | 0.9571 | 0.9637 | 0.0178 |
| 6 | Artificial Neural Networks | 0.8976 | 0.9151 | 0.9657 | 0.9261 | 0.0289 |
| 7 | 0.9718 | 0.9059 | 0.7222 | 0.8666 | 0.1056 | |
| 8 | SVM with Radial Basis Function Kernel | 0.9745 | 0.9276 | 0.6549 | 0.8524 | 0.1409 |
| 9 | Multivariate Adaptive Regression Splines | 0.9553 | 0.8755 | 0.7061 | 0.8456 | 0.1039 |
| 10 | Mixture Discriminant Analysis | 0.9714 | 0.8816 | 0.5853 | 0.8127 | 0.1650 |
| 11 | SVM with Linear Kernel | 0.9827 | 0.8796 | 0.5386 | 0.8003 | 0.1898 |
| 12 | Naive Bayes | 0.9327 | 0.7769 | 0.6069 | 0.7722 | 0.1331 |
| 13 | AdaBoost | 0.9710 | 0.7820 | 0.5092 | 0.7541 | 0.1895 |
| 14 | Linear Discriminant Analysis | 0.9382 | 0.8339 | 0.4792 | 0.7505 | 0.1965 |
| 15 | Nearest Shrunken Centroids | 0.7031 | 0.6202 | 0.4216 | 0.5816 | 0.1181 |
Overall performance of supervised classifiers during the experiments with respect to the - metric.
| No. | Method Name | Complete Dataset | 7% Noisy Dataset | Reduced Dataset | ||
|---|---|---|---|---|---|---|
| 1 | Random Forest | 0.9902 | 0.9909 | 0.9874 | 0.9895 | 0.0015 |
| 3 | Rule-Based Classifier | 0.9942 | 0.9747 | 0.9846 | 0.9845 | 0.0080 |
| 4 | C4.5 Decision Trees | 0.9943 | 0.9739 | 0.9809 | 0.9830 | 0.0085 |
| 5 | Stochastic Gradient Boosting | 0.9725 | 0.9722 | 0.9898 | 0.9782 | 0.0082 |
| 6 | Artificial Neural Networks | 0.9457 | 0.9582 | 0.9824 | 0.9621 | 0.0152 |
| 7 | Multivariate Adaptive Regression Splines | 0.9794 | 0.9469 | 0.8413 | 0.9226 | 0.0590 |
| 8 | 0.9852 | 0.9484 | 0.8325 | 0.9220 | 0.0651 | |
| 9 | SVM with Radial Basis Function Kernel | 0.9867 | 0.9607 | 0.7832 | 0.9102 | 0.0905 |
| 10 | Mixture Discriminant Analysis | 0.9848 | 0.9347 | 0.7181 | 0.8792 | 0.1157 |
| 11 | Naive Bayes | 0.9704 | 0.8810 | 0.7807 | 0.8774 | 0.0775 |
| 12 | SVM with Linear Kernel | 0.9908 | 0.9322 | 0.6772 | 0.8667 | 0.1362 |
| 13 | Linear Discriminant Analysis | 0.9672 | 0.9069 | 0.6307 | 0.8349 | 0.1465 |
| 14 | AdaBoost | 0.9877 | 0.8691 | 0.6270 | 0.8279 | 0.1501 |
| 15 | Nearest Shrunken Centroids | 0.8178 | 0.7493 | 0.5784 | 0.7152 | 0.1006 |
Figure 7Overall variability of supervised classifiers present in the experiments with respect to the accuracy metric.
Figure 8Overall variability of supervised classifiers present in the experiments with respect to the - metric.
Overall performance of supervised classifiers during the experiments with 7%, 15% and 30% noisy datasets.
| No. | Method Name | 7% (acc) | 15% (acc) | 30% (acc) | 7% ( | 15% ( | 30% ( | ||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| 2 | 0.9031 | 0.0334 | 0.9059 | 0.9425 | 0.8608 | 0.9478 | 0.0178 | 0.9484 | 0.9692 | 0.9257 | |
| 3 | SVM with Radial Basis Function Kernel | 0.8856 | 0.0575 | 0.9276 | 0.9249 | 0.8043 | 0.9426 | 0.0266 | 0.9607 | 0.9621 | 0.9051 |
| 4 | Naive Bayes | 0.8583 | 0.0597 | 0.7769 | 0.9182 | 0.8798 | 0.9313 | 0.0361 | 0.8810 | 0.9638 | 0.9491 |
| 5 | Stochastic Gradient Boosting | 0.8363 | 0.1125 | 0.9722 | 0.8402 | 0.6967 | 0.9284 | 0.0475 | 0.9868 | 0.9278 | 0.8705 |
| 6 | SVM with Linear Kernel | 0.7981 | 0.0887 | 0.8796 | 0.8400 | 0.6747 | 0.8970 | 0.0400 | 0.9322 | 0.9177 | 0.8410 |
| 7 | AdaBoost | 0.7736 | 0.1163 | 0.9151 | 0.7755 | 0.6302 | 0.8970 | 0.0496 | 0.9582 | 0.8960 | 0.8367 |
| 8 | Mixture Discriminant Analysis | 0.7629 | 0.1136 | 0.8816 | 0.7975 | 0.6098 | 0.8693 | 0.0641 | 0.9347 | 0.8909 | 0.7823 |
| 9 | Random Forest | 0.7612 | 0.1762 | 0.9825 | 0.7498 | 0.5514 | 0.9095 | 0.0625 | 0.9909 | 0.8986 | 0.8390 |
| 10 | Rule-Based Classifier | 0.7465 | 0.1629 | 0.9484 | 0.7414 | 0.5496 | 0.8893 | 0.0680 | 0.9747 | 0.8851 | 0.8082 |
| 11 | C4.5 Decision Trees | 0.7367 | 0.1695 | 0.9459 | 0.7335 | 0.5308 | 0.8826 | 0.0741 | 0.9739 | 0.8815 | 0.7925 |
| 12 | Linear Discriminant Analysis | 0.7089 | 0.1212 | 0.8339 | 0.7478 | 0.5449 | 0.8368 | 0.0700 | 0.9069 | 0.8622 | 0.7412 |
| 13 | Multivariate Adaptive Regression Splines | 0.6756 | 0.1661 | 0.8755 | 0.6824 | 0.4688 | 0.8590 | 0.0710 | 0.9469 | 0.8569 | 0.7731 |
| 14 | Nearest Shrunken Centroids | 0.6422 | 0.0243 | 0.6202 | 0.6761 | 0.6302 | 0.7807 | 0.0231 | 0.7493 | 0.8044 | 0.7883 |
| 15 | Artificial Neural Networks | 0.5864 | 0.1408 | 0.7820 | 0.5214 | 0.4559 | 0.7469 | 0.0864 | 0.8691 | 0.6892 | 0.6824 |
Training and testing times of the supervised classifiers in the complete and the reduced datasets.
| No. | Method Name | Training Time (s) | Testing Time (ms) | ||
|---|---|---|---|---|---|
| Complete Dataset | Reduced Dataset | Complete Dataset | Reduced Dataset | ||
| 1 | AdaBoost | 20.39 | 6.19 | 0.55 | 0.60 |
| 3 | C4.5 Decision Trees | 2.23 | 0.91 | 0.03 | 0.02 |
| 4 | 6.87 | 3.26 | 0.60 | 0.21 | |
| 5 | Linear Discriminant Analysis | 10.23 | 0.13 | 0.04 | 0.01 |
| 6 | Mixture Discriminant Analysis | 7.23 | 5.02 | 0.21 | 0.15 |
| 7 | Multivariate Adaptive Regression Splines | 40.26 | 7.72 | 0.07 | 0.03 |
| 8 | Naive Bayes | 29.30 | 5.76 | 56.55 | 11.04 |
| 9 | Nearest Shrunken Centroids | 0.08 | 0.08 | 0.01 | 0.01 |
| 10 | Artificial Neural Networks | 18.29 | 12.14 | 0.02 | 0.01 |
| 11 | Random Forest | 24.73 | 8.97 | 0.03 | 0.04 |
| 12 | Rule-Based Classifier | 3.62 | 1.19 | 0.03 | 0.02 |
| 13 | Stochastic Gradient Boosting | 16.54 | 5.48 | 0.07 | 0.07 |
| 14 | SVM with Linear Kernel | 3.51 | 3.91 | 0.10 | 0.07 |
| 15 | SVM with Radial Basis Function Kernel | 26.20 | 36.38 | 1.90 | 3.02 |
Dataset used in the numerical example.
| No. Sample | No. Sample | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| 1 | 4.3213 | 3.6221 | 5.0002 | 1 | 11 | 6.9039 | 4.9760 | 3.3688 | 2 | |
| 2 | 4.2141 | 2.3912 | 6.8321 | 1 | 12 | 7.3675 | 2.3451 | 3.5782 | 2 | |
| 3 | 4.3803 | 6.7115 | 6.8618 | 1 | 13 | 7.4727 | 4.6783 | −4.3858 | 2 | |
| 4 | 3.6646 | 3.5900 | 6.4806 | 1 | 14 | 7.7866 | 5.0108 | 3.4829 | 2 | |
| 5 | 3.2687 | 6.6652 | 6.3268 | 1 | 15 | 6.9858 | 6.8510 | 0.7049 | 2 | |
| 6 | 3.6156 | 4.5413 | 6.6895 | 1 | 16 | 6.8830 | 15.7787 | −3.0057 | 3 | |
| 7 | 3.3436 | 5.9303 | 6.4620 | 1 | 17 | 7.5271 | 17.8117 | −3.1938 | 3 | |
| 8 | 4.4932 | 3.3761 | 5.4943 | 1 | 18 | 7.7992 | 16.9416 | −2.3485 | 3 | |
| 9 | 4.8778 | 2.0918 | 5.5466 | 1 | 19 | 6.9773 | 16.6224 | −2.2307 | 3 | |
| 10 | 4.9102 | 5.1102 | 5.4166 | 1 | 20 | 7.0529 | 17.0954 | 0.4050 | 3 |
Summary of the first update of intermolecular distances.
| 1 | (−0.04, 0.33, −0.01) | (3.27, 3.23, 3.56, 3.54) | (0.11, 0.48, 0.14) |
| 2 | (0.36, 0.53, 0.71) | (3.62, 3.98, 4.51, 5.22) | (0.16, 0.33, 0.51) |
| 3 | (0.06, 0.14, 0.15) | (−4.39, −4.32, −4.19, −4.04) | (0.21, 0.29, 0.30) |
Parameters obtained after training the AHN-model.
| Parameters | Values |
|---|---|
| (0.0, 0.0, 10.94; 0.0, 0.0, 26.99; 0.0, 0.0, 0.0) | |
| (0.0, 0.0; 0.0, 16.02) | |
| (0.0, 0.0, 10.94; 0.0, 0.0, 26.99; 0.0, 0.0, 0.0) | |
| (1, 1, 0) | |
| (3.27, 3.38, 3.86, 3.99) | |
| (3.62, 3.78, 4.11, 4.62) | |
| (−4.39, −4.17, −3.89, −3.59) |
Comparison between the predicted values and the target values y.
| No Sample | ||
|---|---|---|
| 2 | 1 | 1 |
| 3 | 1 | 1 |
| 4 | 1 | 1 |
| 8 | 1 | 1 |
| 9 | 1 | 1 |
| 10 | 1 | 1 |
| 11 | 2 | 2 |
| 12 | 2 | 2 |
| 17 | 3 | 3 |
| 20 | 3 | 3 |