Literature DB >> 27792136

A Flexible Approach for Human Activity Recognition Using Artificial Hydrocarbon Networks.

Hiram Ponce1, Luis Miralles-Pechuán2, María de Lourdes Martínez-Villaseñor3.   

Abstract

Physical activity recognition based on sensors is a growing area of interest given the great advances in wearable sensors. Applications in various domains are taking advantage of the ease of obtaining data to monitor personal activities and behavior in order to deliver proactive and personalized services. Although many activity recognition systems have been developed for more than two decades, there are still open issues to be tackled with new techniques. We address in this paper one of the main challenges of human activity recognition: Flexibility. Our goal in this work is to present artificial hydrocarbon networks as a novel flexible approach in a human activity recognition system. In order to evaluate the performance of artificial hydrocarbon networks based classifier, experimentation was designed for user-independent, and also for user-dependent case scenarios. Our results demonstrate that artificial hydrocarbon networks classifier is flexible enough to be used when building a human activity recognition system with either user-dependent or user-independent approaches.

Entities:  

Keywords:  artificial hydrocarbon networks; artificial organic networks; flexibility; flexible human activity recognition; supervised machine learning; wearable sensors

Mesh:

Substances:

Year:  2016        PMID: 27792136      PMCID: PMC5134431          DOI: 10.3390/s16111715

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

Physical activity recognition based on sensors is a growing area of interest given the great advances in wearable sensors, and the common use of smartphone with powerful embedded sensors. Wearable sensors are getting less obtrusive allowing using sensors for longer periods of time. Applications in various domains are taking advantage of the ease of obtaining data to monitor personal activities and behavior in order to deliver proactive and personalized services. Although many activity recognition systems have been developed for more than two decades, there are still open issues to be tackled with new techniques. Lara et al. [1] envisions six designing challenges for activity recognition: (1) the selection of attributes and sensors; (2) the construction of portable, unobtrusive, and inexpensive data acquisition system; (3) the design of feature extraction and inference methods; (4) data collection under realistic conditions; (5) the flexibility to support new users without the need to re-train the system; and (6) energy consumption. This list of challenges is not exhaustive given that there are other challenges common to various activity recognition scenarios. Recognizing concurrent activities, recognizing interleaved activities, ambiguity of interpretation, and multiple residents are challenges of the nature of human activities defined by [2] needed to be addressed also. We address in this paper one of the main challenges of human activity recognition (HAR) mentioned above: Flexibility. We adopted the flexibility in HAR defined by Lara et al. [1]. In [1], flexibility is contemplated as the ability of the classifier to support new users without the need to collect additional data of the user and re-train the system. Flexibility in activity recognition classifiers can be considered regarding different aspects. For example, flexibility is considered by [3] as the ability of the classifier to recognize different kinds of activities: Common daily activities, activities specific for a certain group of persons, or activities that are rarely performed. Bulling et al. [4] is more interested in generalization ability of the activity recognition classifier. They categorized activity recognition system taking into account if the level of generalization is user independent, user specific and robust to cope with temporal variations. In summary, classifiers must be able to cope with multiple persons performing activities on multiple days, and in multiple runs containing repetitions of the set of activities. Human activity recognition systems generate generic or universal models using training data from several users. These models are then applied to new users without the need to retrain the generic model. Other systems are more focused in specific users that generate personal models that perform the train-test processes only with data of the subject of interest. Recently, personalization of physical activity recognition approaches deal with a new subject from whom data is not available in training phase in a subject- independent system [5]. Nevertheless, personalization approaches only address one aspect of generalization. In previous work, we presented the results of the first tests applying artificial hydrocarbon networks (AHN) for human activity recognition task in [6], using raw sensor data of a public dataset containing five basic and distinctive activity classes (sitting-down, standing-up, standing, walking, and sitting). We compared models generated with ten well-known supervised learning methods against AHN method focusing in the comparison against deep learning method. From this preliminary analysis we concluded that AHN are suitable for activity recognition. Results of a thorough experimental analysis to prove that AHN classifier are very competitive and robust for physical activity recognition was presented in [7]. In that paper we focused in one challenge of HAR: To deal with incomplete noise tolerance. Four experiments were designed using raw data and one window-based approach of another public database with 18 more complex activity classes. Our goal in this work is to present artificial hydrocarbon networks as a flexible approach in a human activity recognition system. We consider flexibility of the approach mainly regarding the ability to support new users (user-independent). We are also concerned the ability to support variations of the same subject in a user-specific approach, and the ability to handle new or irrelevant activities, so we designed some experiments regarding these issues. Since we are using a public dataset for experimentation, real-time variations due to sensors or user behavior are out of the scope of this work. In order to evaluate the performance of artificial hydrocarbon networks based classifier, three kinds of experiments were designed using Attal et al.’s methodology [8]. For each case scenario, the performance of the proposed artificial hydrocarbon networks based classifier was compared with eighteen supervised techniques frequently used in activity recognition systems. Case 1 experiment was designed to assess the performance for all individuals, case 2 experiment assesses the performance of our classifier for user-independent scenario. The first case experiment used cross validation evaluation schema, and the second case used leave-one-subject-out validation. The third experiment (case 3) was designed to test the performance of our classifier for user-dependent scenario. In this case, the classifiers were trained and tested for each individual with her/his own data, and average accuracy and standard deviation was computed. The rest of the paper is as follows. Section 2 discusses related work in flexibility in human activity recognition systems. A brief description of artificial hydrocarbon networks (AHN) technique is presented in Section 3. Our proposed AHN-classifier is presented in Section 4. In Section 5, experimentation is presented, and in Section 6, the results are discussed. Conclusions and directions for future research are described in Section 7.

2. Flexibility in Human Activity Recognition

Every person performs activities in different manner depending on their characteristics such as age, gender, weight, and health condition. Even the same person can change the way of performing an activity depending in the time of the day, emotional and physical state among other thinks. Therefore, the flexibility of a classifier to cope with this diversity of manners of performing the same activity is still one of the main issues of human activity recognition (HAR) [1]. The measures of wearable sensors gathered from a male elderly, a child or a handicapped person doing the same activity present significant differences. One of the main characteristics considered when evaluating human activity recognition systems is flexibility [1]. Although flexibility in a HAR classifier is usually thought regarding the generalization ability in recognizing activities for a new person, it is also considered as the ability of recognizing new activities for one person, or even new runs or sessions to prove robustness over time [4]. In other applications, for example in the video surveillance domain [9] it is more important to consider the classifiers flexibility regarding the ability for adding new activities, namely new and unusual events. Bulling et al. [4] include the characteristic generalization in a HAR system. They identify user-independent systems, user –specific systems, and temporal systems. Lara et al. [1] define the classifier flexibility level as user-specific and monolithic (for user-independent). In the user-specific approach, the system is designed to work with a certain user and to self-adapt to his/her characteristics. Specific approaches are mainly recommended when the users are elderly people, patients with some health problems or disabled. User-dependent systems are frequently used in assisted-living domain given that elderly and people with health problems present differences in their main characteristics that hamper the performance of a generic classifier [8,10,11]. Recently, Capela et al. [12] compared activity recognition with able-bodies and stroke participants, proving that their classifiers performed worse for stroke participants. Regarding their performance, as expected, user-specific models have better performance, but are not generalizable. The main drawback of this approach is that a new model must be done for each user; the system must be retrained. Unlike the specific approach, user-independent systems need to be flexible enough to work with different users [1]. It is important for this kind of systems to be able to keep good performance if new users arrive. Generic or universal models are created from time series dataset of measured attributes from a small or large set of individuals performing each different activity. Depending on case use scenarios, new users may arrive to the activity recognition process. Too many or new activities can also be performed by individuals, making it difficult for one model to cope with all those differences. One way to solve this problem is to create groups with similar characteristics and/or similar activities performed. Some systems like [13], carry out subject-dependent and subject-independent analysis to prove that their classification technique is able to cope with multiple persons, but is also well fitted to build a specific oriented model. Recently, personalization of physical activity recognition has gained interest. Personalization approaches try to deal with the fact that training for activity recognition is usually done on a large number of subjects, and then applied with a new subject from whom data is not available in training phase [5]. Each person has different characteristics that ultimately cause high variance in the activity recognition performance for each subject. Personalization approaches try to cope with these differences adapting the model created with large number of subjects for its application with new users [3,5,14]. In [14], the authors create a model for basic activity recognition based on a decision tree technique, and they change the thresholds of the decision nodes afterwards based on labeled data of each new user. Berchtold et al. [3] present a modular classifier approach based on Recurrent Fuzzy Inference Systems (RFIS). In the later approach the best classifier module is selected from a set of classifiers, and it is adapted to work with new users. In works [3,14], parameters are changed of a general model in order to adapt this universal model to new users. The drawback of this approach is that the general model is either to simple to cope with challenging activity tasks and variety of users, or the model is to complex and therefore entails great computational costs. Reiss [5] presents a different method of personalization in which the general model consists in a set of classifiers weighted the same. A strategy based on weighted majority voting is applied to increase the performance of the model for new users. Instead of retaining classifiers, the method retains only weights reducing the computational complexity. Personalization of physical activity recognition applications is a valid approach to deal with new subject from whom data is not available in training phase in a subject-independent system. Nevertheless, personalization approaches only address one aspect of generalization [15]. A number of researches have explored transfer learning for activity recognition [16]. Transfer learning is the ability to extend what has been learned in one context to new context [17,18]. This approach allows reusing the knowledge previously obtained in a source to a new target population. Roggen et al. [19] defined a run-time adaptive activity recognition chain (adARC) to deal with variations due to placement of sensors, behavior of the user over time, and sensing infrastructure. This architecture allows adaptation according to the recognition of new conditions of the system. The smartphone-based framework of self-learning schema presented in Guo et al. [20] is able to recognize unpredictable activities without any knowledge in the training dataset. They also support variations in smartphone orientation. Li et al. [21] proposed a generic framework for human motion recognition based on smartphones. They presented features to deal with variations due to sensor position and orientation, and user motion patterns. Regarding experimentation design, feature selection can help or hinder the flexibility performance of a HAR classifier. Given the great variability in the performance of activities between different subjects, and even in the same subject at different time, features derived from wearable sensors can lead to great variability. “A good feature set should show little variation between repetitions of the same movements and across different subjects but should vary considerably between different activities” [22]. It is very important to find the best subset of features that combined deliver the best predictors. Regarding the classifier evaluation scheme, subject-dependent and subject-independent methods of evaluation analysis have been used [13]. Preece et al. [22] commented that cross-validation can be done in evaluations between different subjects and within-subject. In user–independent (or between-subject) oriented systems, training is made with almost every subject and test with leave one or a few subjects out. The train-test process is repeated until all subjects have been tested. For the within-subject case, train-test process is made only with the data of a subject, and this process is repeated for data of all subjects available. Average accuracy must be calculated from the results of train-test repetitions in both cases. Lara et al. [1] describe similar evaluation schemes in order to assess the flexibility power of a classifier for each kind of generalization. They state that cross validation or leave-one-out validation schemes are used in user-independent analysis. “Leave-one-person-out is used to assess generalization to an unseen user for a user-independent recognition system” [5].

3. Artificial Hydrocarbon Networks

Artificial hydrocarbon networks (AHN) is a supervised learning method inspired in organic chemistry in order to simulate the chemical rules involved within organic molecules, representing the structure and behavior of data [23,24]. Currently, this method inherits from a general framework of learning algorithms so-called artificial organic networks that proposes two representations of artificial organic molecules: A graph structure related to their physical properties, and a mathematical model behavior related to their chemical properties. The main characteristic of artificial organic networks is packaging information in modules called molecules. These packages are then organized and optimized using heuristic mechanisms based on chemical energy. For readability, Table 1 summarizes the description of chemical-based terms of the artificial organic networks framework and their meanings in the computational AHN technique described below [23].
Table 1

Description of the chemical terms used in artificial hydrocarbon networks.

Chemical TerminologySymbolsMeaning
environmentx(features) data inputs
behaviory(target) data outputs, solution of mixtures
atomsHi, σ(parameters) basic structural units or properties
moleculesφ(x)(functions) basic units of information
compoundsψ(x)(composite functions) complex units of information made of molecules
mixturesS(x)(linear combinations) combination of compounds
stoichiometric coefficientαi(weights) definite ratios in mixtures
intermolecular distancesrj(distances) length between two adjacent molecules
boundsL0,Lj(parameters) lower and upper delimiters, in the inputs, of molecules
energyE0,Ej(loss function) value of the error between real and estimated values
To this end, artificial organic networks, as well as artificial hydrocarbon networks, allow [23,25]: Modularity and organization of information, inheritance of packaging information, and structural stability of data packages. A detailed description of the artificial organic networks framework can be found in [23].

3.1. Description of the AHN-Algorithm

Artificial hydrocarbon networks algorithm (see Figure 1) is inspired in chemical hydrocarbon compounds; thus, this algorithm is only composed of hydrogen and carbon elements that can be linked together with at most one and four atoms, respectively. In this algorithm, linking them in a specific way forms molecules which they are primitive units of information so-called CH-molecules [23]. In fact, these molecules define a mathematical function φ representing the behavior of the CH-molecule, or , as expressed in (1); where, is called the carbon value, is the i-th hydrogen atom attached to the carbon atom, k represents the number of hydrogen atoms in the CH-molecule, and is the input vector with p features.
Figure 1

Structure of an artificial hydrocarbon network using saturated and linear chains of molecules [26]. Throughout this work, the topology of the proposed classifier considers one hydrocarbon compound. Reprinted from Publication Expert Systems with Applications, 42 (22), Hiram Ponce, Pedro Ponce, Héctor Bastida, Arturo Molina, A novel robust liquid level controller for coupled tanks systems using artificial hydrocarbon networks, 8858–8867, Copyright (2015), with permission from Elsevier.

Two or more unsaturated molecules, i.e., , can be joined together in order to form artificial hydrocarbon compounds. Different compounds have been defined in literature [23], and the simplest of those is the saturated and linear chain of molecules like in (2); where, the line symbol represents a simple bond between two molecules. In fact, if there are n CH-molecules, then the compound will have two and molecules [25,26]. Then, a function is associated to the behavior of the artificial hydrocarbon compound, e.g., the piecewise function [23,27], as expressed in (3); where, represents the t-th bound that limits the action of a CH-molecule over the input space by transforming the bounds into centers . In that sense, if the input domain is in the interval , then and , and the j-th CH-molecule is centered at , for all [23]. In addition, bounds are computed using the distance , as (4), between two adjacent molecules, i.e., with . A gradient descent method based on the energy of the adjacent molecules ( and ) is used to calculate the distances as in (5); where, is the learning rate parameter [23,25]. For implementability, the energy of molecules is computed using a loss function [23,25]. In this work, the least squares estimates (LSE) was used to compute the energy of molecules. Several artificial hydrocarbon compounds can interact among them in definite ratios, so-called stoichiometric coefficients, forming a mixture . To this end, a mixture is represented as shown in (6); where, c represents the number of compounds in the mixture and is a set of stoichiometric coefficients [23]. Formally, an artificial hydrocarbon network is a mixture of artificial hydrocarbon compounds (see Figure 1) each one computed using a chemical-based heuristic rule, expressed in the so-called AHN-algorithm [23,25]. Throughout this work, an artificial hydrocarbon network considers one compound, such that and . As noted, the AHN-algorithm is reduced to Algorithm 1 that uses saturated and linear hydrocarbon compounds. At first, the AHN-algorithm initializes an empty compound . Then, a new compound C with n CH-molecules is created as well as a set of random distances . While the difference between real and estimated values are greater than a tolerance value , the data set is partitioned into n subsets using the set of bounds generated with the intermolecular distances. With each subset, the hydrogen and carbon values of the molecular behavior are computed using the LSE method. Then, the compound behavior is assembled and the distances are updated using the error values computed in the LSE method. When the difference between real and estimated values fulfills the tolerance value, the compound is updated with C and its behavior ψ such that . A detailed description of the AHN-algorithm can be found in [23,25]. Also, Appendix A shows a numerical example of training and testing artificial hydrocarbon networks.

3.2. Properties of Artificial Hydrocarbon Networks

The artificial hydrocarbon networks algorithm is characterized by several properties that are very useful when considering regression and classification problems, such as [7,23,26]: Stability, robustness, packaging data and parameter interpretability. Particularly, stability implies that the AHN-algorithm minimizes the changes in its output response when inputs change slightly [7,23], promoting the usage of the artificial hydrocarbon networks as a supervised learning method. In addition, robustness considers that the AHN-algorithm can deal with uncertain and noisy data which implies that it behaves as a filtering information system. For example, it has been used in audio filtering [23,27], and ensembles of artificial hydrocarbon networks with fuzzy inference systems have been successfully employed as intelligent control systems [24,26]. Packaging data is another property of the AHN-algorithm. In fact, this characteristic enables to compute molecular structures into the algorithm in the sense that similar data with similar capabilities are clustered together [23]. In fact, this property intuitively reveals that data is not only packaged by its features, but also by its tendency. Lastly, parameter interpretability refers to that bounds, intermolecular distances and hydrogen values can be useful as metadata to partially understand underlying information or to extract features. For example, the AHN-algorithm has been used in facial recognition approaches when using its parameters as metadata information [23]. Furthermore, the artificial hydrocarbon networks algorithm can be contrasted with other learning models. For instance, it is a supervised, parametric, nondeterministic and multivariate learning algorithm. It means that backpropagation-based multilayer artificial neural networks and support vector machines are close related to artificial hydrocarbon networks in terms of supervised learning and non-probabilistic models used for regression and classification problems. In fact, in [23] authors analyze the location of the AHN-algorithm in the space of learning models, concluding that it is located between regression algorithms, e.g., linear regression and general regression based-learners, and clustering algorithms like k-nearest neighbors, k-means algorithm and fuzzy clustering means. Also, like-smoothers models are not far away from the AHN-algorithm, supporting the robustness property of the latter. To this end, random forest and decision trees models are probabilistic algorithms differing from the artificial hydrocarbon networks algorithm. A detailed comparison of the AHN-algorithm with other learning models can be seen in [23].

4. Description of the Artificial Hydrocarbon Networks Based Classifier

This work considers training and using an AHN-classifier as a flexible approach in human activity recognition systems. In fact, this AHN-classifier is computed and employed in two steps: Training-and-testing and implementation, as shown in Figure 2. Previous work in this direction can be found in [6,7].
Figure 2

Diagram of the proposed artificial hydrocarbon network based classifier (AHN-classifier). First, reduced feature set is used to train the AHN-model, then it is used as AHN-classifier in the testing step.

Currently, the AHN-classifier considers that sensor data has already processed in N features for all , and has organized in Q samples, each one associated to its proper label representing the jth activity in the set of all possible activities Y for ; where, J is the number of different activities in the data set. Thus, samples are composed of features and labels as -tuples of the form for all . Considering that there is a dataset of Q samples of the form defined above, then the AHN-classifier is built and trained using the AHN-algorithm shown in Algorithm 1. It should be noted that this proposal is using a simplified version of artificial hydrocarbon networks. Thus, the AHN-classifier is composed of one saturated and linear hydrocarbon compound, i.e., no mixtures were considered (see Figure 1 for a hydrocarbon compound reference). In that sense, the inputs of the AHN-algorithm are the following: The training dataset Σ is a subset of R samples, from the original dataset, as (7), the number of molecules n in the hydrocarbon compound is proposed to be the number of different activities (), and the learning rate and the tolerance value ϵ are positive numbers selected manually. Notice that the number of molecules in the compound is an empirical value, thus no pairing between classes and molecules occurs. At last, the AHN-algorithm will compute all parameters in the AHN-classifier: Hydrogen and carbon values, as well as the bounds of molecules. For testing and validating the AHN-classifier, the remaining samples P from the original data set (i.e., such that ) conforms the testing data set. Then, the testing data set is introduced to the AHN-classifier, previously computed, and the output response is rounded in order to obtain whole numbers as labels. If output values were out the permitted labels, they were considered as the nearest defining label. Lastly, validation of the classifier is calculated using some metrics. Moreover, new sample data can be also used in the AHN-classifier for recognizing and monitoring a human activity based on the corresponding features.

5. Experimentation

A case study of human activity recognition was implemented using a public dataset in order to measure how well the proposed AHN-classifier performs as a flexible approach in HAR systems. We adopted the activity recognition chain (ARC) approach described by Bulling et al. [4] and we also added an unknown-activity detection module in order to discriminate possible new or irrelevant activities that might lead in misclassification. Our approach performs the following stages: (i) data acquisition; (ii) signal preprocessing and segmentation, e.g., windowing; (iii) feature extraction; (iv) feature reduction; (v) building an unknown-activity detector; (vi) building activity models; and (vii) classification or activity evaluation. Figure 3 shows the methodology of the HAR system of this case study.
Figure 3

Methodology implemented in the case study for HAR systems.

5.1. Dataset Description

This case study employs a dataset provided by the Bilkent University from Ankara, Turkey [28]. It consists on a set of 45 raw signals from five inertial measurement units (IMUs) placed in the body of eight different subjects, performing nineteen different activities. In fact, each IMU is composed of three 3-axes sensors: An accelerometer, a gyroscope, and a magnetometer. In addition, Figure 4 shows the position of the IMUs: One at the torse, two at the arms and two at the legs.
Figure 4

Location of the five wearable IMUs used in the dataset.

The nineteen activities carried out by the subjects are [28]: (1) sitting; (2) standing; (3) lying on back; (4) lying on right side; (5) ascending stairs; (6) descending stairs; (7) standing in an elevator still; (8) moving around in an elevator; (9) walking in a parking lot; (10) walking on a treadmill with a speed of 4 km/h in flat; (11) walking on a treadmill with a speed of 4 km/h and 15 degree inclined positions; (12) running on a treadmill with a speed of 8 km/h; (13) exercising on a stepper; (14) exercising on a cross trainer; (15) cycling on an exercise bike in horizontal positions; (16) cycling on an exercise bike in vertical positions; (17) rowing; (18) jumping; and (19) playing basketball. We used the public dataset [28] given that each activity was performed by the subjects in their own style. This allows inter-subject variability. It is also correctly labeled and segmented by subject and by activity. These segmentations permit to easily design different experimental datasets. The limitation of this dataset is that it does not include intra-subject variability.

5.2. Windowing and Feature Extraction

We apply a windowing approach to the entire dataset of raw signals. In particular, we select windows of 5 s in size without overlapping. Then, we extract 18 features for each channel based on literature: 12 features in time domain as shown in Table 2, and 6 features in frequency domain as shown in Table 3. Currently, each window is composed of 125 raw samples, and there are 1140 windows per subject. Considering that each activity is performed during 5 min by each subject, then there are 60 windows per activity.
Table 2

Features extracted in time domain.

FeaturesReferences
mean[4,29,30,31,32,33,34]
standard deviation[30,31,34]
root mean square[29]
maximal amplitude[31,33]
minimal amplitude[31,33]
median[32,34]
number of zero-crossing[29,31]
skewness[33]
kurtosis[4,33]
first quartile[32,34]
third quartile[32,34]
autocorrelation[31,33]
Table 3

Features extracted in frequency domain.

FeaturesReferences
mean frequency[29,31]
median frequency[29]
entropy[30,32]
energy[4,30,32]
principal frequency[32,33,34]
spectral centroid[31,32]

5.3. Feature Reduction

Considering that there are 45 channels of raw signals and 18 features per channel, then the total number of features extracted is 810. Due to the fact that the latter demands high computational resources, a feature reduction procedure was applied using the well-known principal components analysis (PCA) [35]. Currently, PCA transforms a high dimensional domain into a lower dimensional domain by applying a linear combination of weighted features. In that sense, we applied PCA to the feature set and we obtain a reduced feature set of so-called components [35]. In order to select the optimal number of components, we chose the eigenvalue criterion or the Kaiser criterion [36], one of the most commonly used criteria for solving the number of components problem in PCA that consists of retaining any component with a variance value greater than 1. Thus, the components were sorted in descending order, finding that the first 91 components have variance value greater than one (representing the of the feature set), as shown in Figure 5. To this end, the reduced feature set of the first 91 components was employed in this case study to build the activity models, as described below.
Figure 5

A subset of the first one-hundred components calculated by the PCA method: Variance values shown in straight line, and cumulative variance shown in dashed line.

5.4. Unknown-Activity Detection Module

We developed a module to detect new and/or irrelevant activities inspired in the methodology of Guo et al. [20]. This module performs a rough classification of reduced feature vectors in known and unknown activities using an AHN-based classifier. If an instance is considered unknown, it will be stored for future manual tagging. Otherwise, the instance is processed normally. It should be noted that this module is a first and independent classifier that roughly determines if a reduced feature vector would be an already known activity in order to let it continue in the workflow. In order to validate this module, we selected five different activities (sitting, lying on back, ascending stairs, walking in a parking lot and exercising on a stepper) coming from all the subjects in the dataset avoiding user-specific training. Then, we used of them to build the AHN-classifier. From (3), it can be seen that each molecule has an associated parameter referring to its center . Then, these centers can be used as the centers, namely for all , of these clusters/activities. Then, we measured the distance of each training sample to the nearest center, and we computed the mean m and standard deviation σ of these distances. After that, the unknown activity detection module was developed using the heuristic as expressed in (8); where, x is the input (i.e., the reduced feature vector representing the testing sample), d is the -norm distance, and is the j-th center computed before when training the AHN-classifier. In a nutshell, determines if the input x is near at least to one of the clusters defined by the training activities () and then is a known activity. If not (), then the input x is an unknown activity. Four unknown activities were selected as part of the testing set (i.e., cycling on an exercise bike in horizontal positions, jumping, walking on a treadmill with a speed of 4 km/h in flat and lying on right side) as well as the remaining of the known activities. Table 4 shows the accuracy of this module for detecting known and unknown activities. In terms of the known activities, this module recognizes it with a mean accuracy of . However, cycling on an exercise bike and walking on a treadmill with a speed of 4 km/h in flat activities were misclassified. In the first activity, the AHN-classifier got confused between exercising on a stepper and cycling on an exercise bike; while the latter can be explained since it is very similar to the known activity walking in a parking lot. Lying on right side activity was well classified.
Table 4

Accuracy of the unknown-activity detection module.

ActivityTargetAccuracy of AHNAccuracy of k-Means
sittingknown0.95830.9653
lying on backknown0.91660.8958
ascending stairsknown0.93060.8472
walking in a parking lotknown0.60420.9444
exercising on a stepperknown0.95830.9653
cycling on an exercise bikeunknown0.16660.9583
jumpingunknown0.79171.0000
walking on a treadmillunknown0.02080.1528
lying on right sideunknown1.00001.0000
For comparison purposes, we also designed a similar classifier based on the k-means method since it calculates centers of known clusters/activities. Table 4 summarizes its results. It can be seen that the module can classify known activities with in average. In terms of the unknown activities, most of them were classified with in average except the activity walking on a treadmill with a speed of 4 km/h in flat. This misclassification can be explained since the latter is very similar to the known activity walking in a parking lot. As noted, both classifiers obtain similar performance accuracy on known activities. In terms of unknown activities, there is a similar tendency, except on the cycling on an exercise bike activity. Since the activities are well classified and similar unknown activities are recognized as the known-like activities by the module using AHN or k-means, this methodology is proposed to be used before more accurate human activity classifier models. Finally, this experiment opens the possibility to use the same AHN-classifier for both human activity recognition (using the output response of artificial hydrocarbon networks) and unknown-activity detection (using the parameter interpretability of the center of molecules).

5.5. Building Supervised Activity Models

To compare our proposed AHN-classifier, we choose eighteen supervised methods aiming to evaluate the performance of artificial hydrocarbon networks as classifier over HAR systems in both user-independent and user-dependent approaches. The following supervised learning methods were selected, i.e., supported in reviewed literature [1,22,37,38], to build activity models: stochastic gradient boosting (SGB), AdaBoost (AB), C4.5 decision trees (DT4), C5.0 decision trees (DT5), rule-based classifier (RBC), single rule classification (SRC), support vector machines with basis function kernel (SVM-BF), random forest (RF), k-nearest neighbors (KNN), penalized discriminant analysis (PDA), mixture discriminant analysis (MDA), shrinkage discriminant analysis (SDA), multivariate adaptive regression splines (MARS), naive Bayes (NB), multilayer feedforward artificial neural networks (ANN), model averaged artificial neural networks (MA-ANN), nearest shrunken centroids (NSC), and deep learning (DL) using a deep neural networks (DNN) approach. The caret package and other libraries in R were employed to build suitable activity models. Table 5 summarizes the configuration parameters of these models. For reproducibility, we set a seed value, , when building the models.
Table 5

Configuration parameters for building suitable activity models using the caret package in R. Other parameters of the method marked with (*) are: Activation_function = hyperbolic tangent, hidden_layers = (200, 250, 200), balance_classes = true.

NoMethodConfigurationsParametersValuesTraining Time (s)Testing Time (ms)
1AdaBoost27(mfinal, maxdepth, coeflearn)(150, 3, 3)22.9601.213
2Artificial Hydrocarbon Networks1(n_molecules, learning_rate, tolerance)(19, 0.5, 0.1)1709.1200.028
3C4.5-Decision Trees1(C)(0.25)3.4260.069
4C5.0-Decision Trees12(trials, model, winnow)(20, 1, TRUE)10.5090.545
5Deep Learning *1(rate annealing, epochs, rate)(0.001, 300, 0.01)21.5800.970
6k-Nearest Neighbors3(kmax, distance, kernel)(5, 2, 1)5.7770.804
7Mixture Discriminant Analysis3(subclasses)(4)5.8390.197
8Model Averaged Artificial Neural Networks9(size, decay, bag)(5, 0.1, FALSE)12.1140.040
9Multivariate Adaptive Regression Splines1(degree)(1)99.2150.172
10Naive Bayes2(fL, usekernel)(0, TRUE)32.06592.953
11Nearest Shrunken Centroids3(threshold)(3.38)0.0690.022
12Artificial Neural Networks9(size, decay)(5, 0.1)5.9050.022
13Penalized Discriminant Analysis3(lambda)(1)0.3640.022
14Random Forest3(mtry, ntrees)(2, 100)29.4640.077
15Rule-Based Classifier1(threshold, pruned)(0.25, 1)7.2130.077
16Shrinkage Discriminant Analysis3(diagonal, lambda)(FALSE, 0)0.2990.018
17Single Rule Classification1(-)(-)2.9800.062
18Stochastic Gradient Boosting9(n.trees, interaction.depth, shrinkage)(150, 3, 0.1)18.2770.164
19SVM with Radial Basis Function Kernel3(C)(1)25.4793.187
In order to build these activity models, three different cases were considered in order to measure and validate the performance of the AHN-classifier in flexibility, as follows: Case 1: All subjects using cross-validation. This experiment uses 70% of the reduced feature set as the training set and 30% as the testing set, in order to validate how well the AHN-classifier performs for all users. To obtain the best model configuration, we previously used 10-fold cross-validation and 5 repetitions in the training set. Case 2: User-independent performing leave-one subject-out. This experiment is based on the well-known leave-one subject-out technique [1], aiming to prove how well is the AHN-classifier to predict activities in new subjects. In fact, we build eight models by training each model with information from seven subjects and leaving one subject out. Then, the latter not used in the training step is employed to test the performance of the classifier. Then, the overall performance of classifiers is measured as an average of the eight models. Case 3: User-dependent performing cross-validation within a subject. This experiment considers building eight different models from each of the subjects in order to measure the performance of the AHN-classifier in a user-specific approach. For each subject, 70% of the feature set is used as the training set and the 30% of it is used as the testing set. Then, the overall performance of classifiers is measured as an average of the eight models. To this end, the experiments were executed in a computer Intel CoreTMi5-2400 with CPU at 3.10 GHz and 16 GB-RAM over Windows 7 Pro, Service Pack 1 64-bits operating system.

5.6. Metrics

We use different metrics to evaluate the performance of the AHN-classifier in comparison with the other supervised classifiers, such as: , , , and [39]. Notice that the reduced feature set contains the same size samples of each class, so it is balanced. Other metrics are computed as well (Table 5): specifies the training time (in seconds) to build and train a model, and specifies the evaluation time of an input sample (in milliseconds).

6. Results and Discussion

This section presents the results of the comparison between the proposed AHN-classifier and other eighteen supervised methods as a flexible approach for human activity recognition systems. Then, a discussion is also presented.

6.1. Case 1: All Subjects Using Cross-Validation

A cross-validation with 10-folds and 5-repetitions was computed in the training step to obtain a suitable model. Table 6 summarizes the results of this experiment sorted in descending order by accuracy. Additionally, Figure 6 shows the confusion matrix of the proposed AHN-classifier. As noted, the proposed AHN-classifier ranks in second place with an accuracy of such below to the deep learning based classifier. Then, mixture discriminant analysis, C5.0 decision trees, random forest and SVM with radial function are in the top of the list.
Table 6

Results of case 1: Performance for all subjects using cross-validation.

NoMethodAccuracySensitivitySpecificityPrecisionF1-Score
1Deep Learning99.2799.2799.9699.2899.62
2Artificial Hydrocarbon Networks98.7698.7699.9398.7899.35
3Mixture Discriminant Analysis98.3698.3699.9198.4399.16
4C5.0-Decision Trees98.2898.2899.9098.2899.08
5Random Forest98.2598.2599.9098.2799.08
6SVM with Radial Basis Function Kernel98.1098.1099.8998.1799.03
7Stochastic Gradient Boosting97.9997.9999.8998.0398.95
8Artificial Neural Networks97.8897.8899.8897.8798.87
9Multivariate Adaptive Regression Splines97.4897.4899.8697.4398.63
10Penalized Discriminant Analysis97.0097.0099.8397.0798.43
11Shrinkage Discriminant Analysis97.0097.0099.8397.0798.43
12Rule-Based Classifier96.2796.2799.7996.2998.01
13k-Nearest Neighbors95.7695.7699.7695.6797.68
14Naive Bayes95.5895.5899.7595.8697.77
15AdaBoost95.5095.5099.7595.7897.73
16C4.5-Decision Trees95.2595.2599.7495.2597.44
17Nearest Shrunken Centroids93.3193.3199.6393.7096.57
18Model Averaged Artificial Neural Networks91.0591.0599.5092.7095.98
19Single Rule Classification37.8737.8796.5538.0554.59
Average93.3293.3299.6393.4895.82
Figure 6

Results of case 1: Confusion matrix of the AHN-classifier in the performance for all subjects using cross-validation. Numbers represent window counts.

6.2. Case 2: User-Independent Performing Leave-One Subject-Out

This experiment is based on the well-known leave-one subject-out technique [1], aiming to prove how well is the AHN-classifier to predict activities in new subjects. Table 7 shows the overall performance of the supervised models sorted in descending order by accuracy, and Table 8 shows the performance of each model. In addition, Figure 7 shows the average confusion matrix of the AHN-classifier. In this case, the AHN-classifier also ranks in second place with a mean accuracy of just below the deep learning based classifier. Additionally, penalized discriminant analysis, shrinkage discriminant analysis, mixture discriminant analysis and nearest shrunken centroids are also in the top of the list.
Table 7

Results of case 2: Leave-one subject-out overall performance for the user-independent approach. Values with (*) were obtained using only available metrics when they can be performed over results.

NoMethodAccuracySensitivitySpecificityPrecisionF1-Score
1Deep Learning94.0594.0599.6796.0497.82
2Articial Hydrocarbon Networks93.2393.2399.6293.5996.51
3Penalized Discriminant Analysis92.6492.6499.5994.28 *96.87 *
4Shrinkage Discriminant Analysis92.6392.6399.5994.28 *96.87 *
5Mixture Discriminant Analysis90.8090.8099.4992.26 *95.73 *
6Nearest Shrunken Centroids90.4190.4199.4791.39 *95.23 *
7C5.0-Decision Trees87.5787.5799.3189.37 *94.07 *
8Random Forest87.3587.3599.3090.00 *94.39 *
9Stochastic Gradient Boosting87.1187.1199.2890.64 *94.75 *
10AdaBoost86.8086.8099.2788.40 *93.43 *
11Multivariate Adaptive Regression Splines85.1585.1599.1886.69 *92.46 *
12SVM with Radial Basis Function Kernel81.3381.3398.9688.44 *93.36 *
13Rule-Based Classifier81.2381.2398.9684.13 *90.93 *
14C4.5-Decision Trees80.0780.0798.8983.93 *90.79 *
15Naive Bayes79.0679.0698.8470.01 *74.29 *
16Model Averaged Artificial Neural Networks75.0475.0498.6177.45 *86.82 *
17k-Nearest Neighbors74.9174.9198.6182.18 *89.65 *
18Artificial Neural Networks61.3861.3897.8576.32 *86.06 *
19Single Rule Classification29.9229.9296.1130.7946.51
Average80.9280.9298.9484.22 *89.82 *
Table 8

Results of case 2: Leave-one subject-out performance for the user-independent approach for each of the models created.

NoMethodAvg & StdSub 1Sub 2Sub 3Sub 4Sub 5Sub 6Sub 7Sub 8
1Deep Learning94.05 ± 1.6197.3795.3592.5492.5493.6093.9593.4293.60
2Articial Hydrocarbon Networks93.23 ± 1.3794.5695.6192.5492.6392.4692.6394.0491.40
3Penalized Discriminant Analysis92.64 ± 0.8893.4292.4692.4691.9394.0491.1492.8192.89
4Shrinkage Discriminant Analysis92.63 ± 0.8893.4292.4692.4691.9394.0491.1492.7292.89
5Mixture Discriminant Analysis90.8 ± 1.8290.0989.3990.2689.3993.9589.3993.3390.61
6Nearest Shrunken Centroids90.41 ± 3.3885.2693.6093.9591.8493.1690.8888.1686.40
7C5.0-Decision Trees87.57 ± 4.3582.8186.7590.4480.7987.9889.6594.6587.46
8Random Forest87.35 ± 3.9483.2584.9186.8490.7091.9381.4987.9891.67
9Stochastic Gradient Boosting87.11 ± 5.0984.3083.3391.7581.9395.0983.6892.3784.39
10AdaBoost86.8 ± 5.181.1486.5893.9578.7789.8285.2691.5887.28
11Multivariate Adaptive Regression Splines85.15 ± 4.7282.6385.5387.7290.3580.7076.9390.0087.37
12SVM with Radial Basis Function Kernel81.33 ± 3.176.4980.6180.7086.1482.9878.6080.7084.39
13Rule-Based Classifier81.23 ± 6.1775.0982.4685.1869.2181.0586.7587.0283.07
14C4.5-Decision Trees80.07 ± 5.1777.8977.8177.1973.6884.8286.6786.6775.79
15Naive Bayes79.06 ± 4.4173.7776.8484.3081.2382.4672.1179.3982.37
16Model Averaged Artificial Neural Networks76.82 ± 10.4673.6864.0496.3278.3373.6071.6787.6369.30
17k-Nearest Neighbors74.91 ± 6.1374.1275.7972.5477.1185.0063.3378.3373.07
18Artificial Neural Networks73.65 ± 8.9775.5365.1885.7081.3278.9561.1464.3077.11
19Single Rule Classification29.92 ± 4.1430.0924.2129.4728.1632.8937.5425.9631.05
Average81.7 ± 4.4479.3179.8684.6580.8684.1679.4483.7681.58
Figure 7

Results of case 2: Confusion matrix of the AHN-classifier in the leave-one subject-out performance for user-independent. Numbers represent the average of window counts in the eight models.

6.3. Case 3: User-Dependent Performing Cross-Validation within a Subject

This experiment considers building eight different models from each of the subjects in order to measure the performance of the AHN-classifier in a user-specific approach. Table 9 summarizes the overall results of this experiment sorted in descending order by accuracy, and Table 10 shows the performance of each model. Additionally, Figure 8 reports the confusion matrix of the proposed AHN-classifier. The proposed AHN-classifier ranks at the first place with a mean accuracy of . Currently, deep learning, mixture discriminant analysis, shrinkage discriminant analysis and penalized discriminant analysis are also in the top of the list.
Table 9

Results of case 3: Cross-validation within a subject overall performance for the user-dependent approach. Values with (*) were obtained using only available metrics when they can be performed over results.

NoMethodAccuracySensitivitySpecificityPrecisionF1-Score
1Articial Hydrocarbon Networks99.4999.4999.9799.5199.74
2Deep Learning99.2799.2799.9699.3599.66
3Mixture Discriminant Analysis99.2099.2099.9699.2699.61
4Shrinkage Discriminant Analysis99.0599.0599.9599.1299.53
5Penalized Discriminant Analysis99.0199.0199.9599.0899.51
6Model Averaged Artificial Neural Networks98.7998.7999.9398.8499.38
7Random Forest98.7298.7299.9398.7999.36
8Multivariate Adaptive Regression Splines98.4398.4399.9198.5999.25
9C5.0-Decision Trees97.9997.9999.8998.0798.97
10SVM with Radial Basis Function Kernel97.9297.9299.8898.0698.96
11Nearest Shrunken Centroids97.6297.6299.8797.9198.88
12Stochastic Gradient Boosting97.4897.4899.8697.6498.73
13AdaBoost97.4497.4499.8697.9698.90
14Naive Bayes97.2697.2699.8597.7098.76
15C4.5-Decision Trees96.4296.4299.8096.6198.18
16Rule-Based Classifier95.9495.9499.7796.0997.89
17k-Nearest Neighbors95.6195.6199.7695.6797.67
18Artificial Neural Networks92.5892.5899.5948.12 *48.98 *
19Single Rule Classification61.8461.8497.8854.1166.27
Average95.7995.7999.7795.6997.18
Table 10

Results of case 3: Cross-validation within a subject performance for the user-dependent approach for each of the models created.

NoMethodAvg & StdSub 1Sub 2Sub 3Sub 4Sub 5Sub 6Sub 7Sub 8
1Articial Hydrocarbon Networks99.49 ± 0.4499.1299.1299.71100.0098.83100.0099.4299.71
2Deep Learning99.27 ± 0.1699.1299.4299.1299.4299.4299.1299.1299.42
3Mixture Discriminant Analysis99.2 ± 0.4698.5498.8399.4299.4298.8399.4299.12100.00
4Shrinkage Discriminant Analysis99.05 ± 0.398.5499.1299.4299.4298.8399.1298.8399.12
5Penalized Discriminant Analysis99.01 ± 0.3898.2599.1299.4299.4298.8399.1298.8399.12
6Model Averaged Artificial Neural Networks98.79 ± 0.3698.2598.5498.5499.4298.8398.8399.1298.83
7Random Forest98.72 ± 0.6298.2597.9598.5499.4298.2598.5499.1299.71
8Multivariate Adaptive Regression Splines98.43 ± 1.3495.6198.5499.7199.1297.3799.1299.4298.54
9C5.0-Decision Trees97.99 ± 0.5796.7898.2597.6698.5498.2597.9597.9598.54
10SVM with Radial Basis Function Kernel97.92 ± 0.7497.6696.7897.3798.8397.6698.5497.6698.83
11Nearest Shrunken Centroids97.62 ± 1.197.6697.3796.2098.8395.9197.9598.2598.83
12Stochastic Gradient Boosting97.48 ± 1.1295.6197.0897.3799.1298.5496.4997.6697.95
13AdaBoost97.44 ± 1.3295.9197.9598.2598.8397.6695.0397.3798.54
14Naive Bayes97.26 ± 1.0696.7896.2097.9599.4296.2096.7897.3797.37
15C4.5-Decision Trees96.42 ± 0.8495.3296.2097.0897.0897.0895.3297.3795.91
16Rule-Based Classifier95.94 ± 1.2695.0395.3295.3295.9196.4994.1597.9597.37
17k-Nearest Neighbors95.61 ± 1.0496.7893.8696.4995.3295.0394.7496.4996.20
18Neural Network92.58 ± 4.1689.1894.1596.4998.2585.0993.8691.2392.40
19Single Rule Classification61.84 ± 4.163.4565.2063.7462.8767.5457.8958.4855.56
Average95.79 ± 1.1295.0495.7496.296.7795.5195.3795.8395.89
Figure 8

Results of case 3: Confusion matrix of the AHN-classifier in the cross-validation within a subject performance for user-dependent. Numbers represent the average of window counts in the eight models.

6.4. Discussion

As noted above, the proposed AHN-classifier outperformed in the three case experiments. In addition, we conducted a paired t-test analysis to find out if the differences between the accuracy of the AHN-classifier performance and the other supervised model performances are statistically significant. Table 11 summarizes the p-values of this test for cases 2 and 3 using a confidence level. As shown, any p-value greater than (bold values in Table 11) means that the null hypothesis about the equality of accuracy values between model performances is accepted, otherwise accuracy values between model performances are not statistically equal and the hypothesis is denied. In that sense, the AHN-classifier can be fairly compared with those model performances with a p-value less than , concluding that the AHN-classifier is significantly better than mixture discriminant analysis based classifier and those below the seventh position in Table 8 for case 2. In addition, the AHN-classifier is significantly better than those model performances below the third position in Table 10. To this end, it is shown that the AHN-classifier is significantly equivalent to deep learning in both cases 2 and 3. In fact, these experiments and their t-test analysis consider to validate that the AHN-classifier is suitable as a flexible approach for HAR systems based on: The ability to support new users (user-independent), and the ability to build models for a specific user. Furthermore, new and unknown activities need more tests before validating its flexibility. From now, we handle and filter them before the main human activity classification.
Table 11

Results of the t-test analysis reporting the p-values or cases 2 and 3. Bold values represent p-values greater than ( confidence level).

Methodp-Value in Case 2p-Value in Case 3
AdaBoost0.0130.003
C4.5-Decision Trees0.0000.000
C5.00.0100.000
Deep Learning0.1090.244
k-Nearest Neighbors0.0000.000
Mixture Discriminant Analysis0.0250.000
Model Averaged Artificial Neural Networks0.0040.004
Multivariate Adaptive Regression Splines0.0020.002
Naive Bayes0.0000.006
Nearest Shrunken Centroids0.0630.001
Artificial Neural Networks0.0010.002
Penalized Discriminant Analysis0.3240.000
Random Forest0.0110.033
Rule-Based Classifier0.0010.005
Shrinkage Discriminant Analysis0.3170.000
Single Rule Classification0.0000.000
Stochastic Gradient Boosting0.0160.001
SVM with Radial Basis Function Kernel0.0000.031
On one hand, the proposed AHN-classifier reached (case 1) when dealing with a HAR system using all subjects for training and testing. In the user-independent performance (case 2), results computed of accuracy in cross-validation over subjects and leave-one subject-out experiments. Then, analyzing the confusion matrices of both experiments (Figure 6 and Figure 7), it can be seen that false predictions are very close to the diagonal (true positives). In the cross-validation experiment (case 1), we can observed that the activities predicted are very similar to the actual activities. For example, the AHN-classifier predicted lying on back when the actual activity was lying on right side. Likewise, it predicted walking at 4 km/h on flat when the actual activity was walking in a parking lot. In the leave-one subject-out experiment (case 2), the maximum average window counts value has to be 60, the number of window counts per activity, at each element in the chart of Figure 7. Then, it can be observed that true positive counts are very close to the maximum value and the others are close to zero. It means that the AHN-classifier is able to truly classify human activities with a very low misclassification. On the other hand, in user-dependent approach, the proposed AHN-classifier reached of accuracy in cross-validation within a subject. In some cases such as subject-4 and subject-6, the AHN-classifier predicts of the activities carried out by the subject. This is a slight advantage over the other five top supervised models (see Table 9). The same behavior in the confusion matrix of this experiment (Figure 8) was found as in the other cases. It is important to note that deep learning, i.e., DNN, based classifier is the only model that outperforms the AHN-classifier in the first two cases. Conversely, the AHN-classifier kept at the top of the benchmark experiments in comparison with deep learning which dropped to the second place in case 3. In terms of the unknown-activity detection module using the AHN-classifier, the experimentation shows a suggested way in which the AHN-based classifier can discriminate known and unknown activities by themselves, using the centers of molecules. The latter actually reflects the parameter interpretability property of AHN, since the centers of molecules serve as features to find correspondence between training and new data. Thus, a proper training of a single AHN-classifier should deal with both human activity classification and detecting unknown activities at the same time. Particularly to this work, the centers of molecules were not employed in the main experimentation (cases 1, 2 and 3) since the sensor signals are clearly related to known activities in the dataset. From Table 5, some computational issues can be identified in the AHN-classifier. Particularly, the training time of the AHN ( s) exceeded times the maximum training time ( s) performed by the other methods. This step is time-consuming mainly on the splitting procedure at each iteration of Algorithm 1 and the inner building function (1) used for running the LSE method. In the current work, this is not a problem. However, if real-time HAR systems are implemented, an improvement on this computational issue will be handled. For instance, another splitting procedure might be considered. To this end, the AHN-classifier can achieve a flexible approach for user-dependent, user-independent, or both scenarios, in human activity recognition, as described in this work.

7. Conclusions and Future Work

In order to cope with real-world activity recognition challenges, a supervised machine learning technique must be flexible. In this paper, we considered flexibility of the approach regarding: The ability to support new users (user-independent). We were also concerned the ability to support variations of the same subject in a user-specific approach, and the ability to handle new or irrelevant activities. In that sense, we presented a novel supervised machine learning method called artificial hydrocarbon networks as a flexible approach for human activity recognition. The AHN-classifier performance was compared with eighteen commonly used supervised techniques. We also designed an unknown-activity detection module that performs a rough classification to handle new and irrelevant activities. For our user-independent and user-dependent case scenarios, our results showed that AHN-classifier remained at the top of the other classifiers. Our results demonstrated that artificial hydrocarbon networks classifier serves as a flexible approach when building a human activity recognition system with either user-dependent or user-independent approaches. For future research, we must address flexibility regarding the ability to recognize new and complex activities. Also, the parameter interpretability of AHN will be deeply analyzed to determine the conditions and training procedures to perform human activity recognition and unknown-activity detection with a single AHN-based model. Further experimentation is also needed in order to prove flexibility when intra-subject variability occurs. Another challenge to be attended is to demonstrate that our AHN-classifier is well suited for real-time HAR systems, using other sensor configurations and improvements of computational issues at the training step of the method.
Table A1

Data set used in the numerical example.

No Samplex1x2x3yNo Samplex1x2x3y
14.33.65.11116.94.93.82
24.22.46.81127.12.53.72
33.96.26.81137.54.7−4.32
43.83.66.51147.85.13.42
53.35.66.31156.56.50.72
63.74.56.61166.815.7−3.03
73.65.96.41177.517.2−3.23
84.43.65.41187.216.9−2.33
94.52.05.51196.916.3−2.23
104.95.25.31207.017.10.43
Table A2

Training and testing sets for the numerical example.

SetSamples
training{1,6,9,10,13,15,16,18,19,20}
testing{2,3,4,5,7,8,11,12,14,17}
Table A3

Intermolecular distances and bounds for the numerical example at iterations and .

jrj(i=0)Lj(i=0)rj(i=1)
0(3.3,2.0,4.3)
1(0.7,3.2,2.5)(4.0,5.2,1.8)(0.7,3.2,2.5)
2(1.3,8.2,4.2)(5.3,13.4,2.4)(1.31,8.21,4.21)
3(2.2,5.7,7.8)(7.8,17.2,6.8)(2.19,5.69,7.79)
Table A4

Obtained subsets when partitioning the training set at iteration .

SetSamples
Σ1{9,12,13}
Σ2{1,2,3,4,5,6,7,8,10,11,14,15,16}
Σ3{17,18,19,20}
Table A5

Comparison between estimated and target y values for the numerical example.

No SampleyyAHN
211
311
411
511
711
811
1122
1222
1422
1733
  10 in total

1.  Personalization algorithm for real-time activity recognition using PDA, wireless motion bands, and binary decision tree.

Authors:  Juha Pärkkä; Luc Cluitmans; Miikka Ermes
Journal:  IEEE Trans Inf Technol Biomed       Date:  2010-09

Review 2.  Activity identification using body-mounted sensors--a review of classification techniques.

Authors:  Stephen J Preece; John Y Goulermas; Laurence P J Kenney; Dave Howard; Kenneth Meijer; Robin Crompton
Journal:  Physiol Meas       Date:  2009-04-02       Impact factor: 2.833

3.  A comparison of feature extraction methods for the classification of dynamic activities from accelerometer data.

Authors:  Stephen J Preece; John Yannis Goulermas; Laurence P J Kenney; David Howard
Journal:  IEEE Trans Biomed Eng       Date:  2008-10-31       Impact factor: 4.538

4.  Smartphone-Based Patients' Activity Recognition by Using a Self-Learning Scheme for Medical Monitoring.

Authors:  Junqi Guo; Xi Zhou; Yunchuan Sun; Gong Ping; Guoxing Zhao; Zhuorong Li
Journal:  J Med Syst       Date:  2016-04-22       Impact factor: 4.460

5.  The development of an artificial organic networks toolkit for LabVIEW.

Authors:  Hiram Ponce; Pedro Ponce; Arturo Molina
Journal:  J Comput Chem       Date:  2015-01-06       Impact factor: 3.376

6.  Human Activity Recognition and Pattern Discovery.

Authors:  Eunju Kim; Sumi Helal; Diane Cook
Journal:  IEEE Pervasive Comput       Date:  2010       Impact factor: 3.175

7.  Transfer Learning for Activity Recognition: A Survey.

Authors:  Diane Cook; Kyle D Feuz; Narayanan C Krishnan
Journal:  Knowl Inf Syst       Date:  2013-09-01       Impact factor: 2.822

Review 8.  Physical Human Activity Recognition Using Wearable Sensors.

Authors:  Ferhat Attal; Samer Mohammed; Mariam Dedabrishvili; Faicel Chamroukhi; Latifa Oukhellou; Yacine Amirat
Journal:  Sensors (Basel)       Date:  2015-12-11       Impact factor: 3.576

9.  A Novel Wearable Sensor-Based Human Activity Recognition Approach Using Artificial Hydrocarbon Networks.

Authors:  Hiram Ponce; María de Lourdes Martínez-Villaseñor; Luis Miralles-Pechuán
Journal:  Sensors (Basel)       Date:  2016-07-05       Impact factor: 3.576

10.  Evaluation of a smartphone human activity recognition application with able-bodied and stroke participants.

Authors:  N A Capela; E D Lemaire; N Baddour; M Rudolf; N Goljar; H Burger
Journal:  J Neuroeng Rehabil       Date:  2016-01-20       Impact factor: 4.262

  10 in total
  7 in total

Review 1.  Multi-Sensor Fusion for Activity Recognition-A Survey.

Authors:  Antonio A Aguileta; Ramon F Brena; Oscar Mayora; Erik Molino-Minero-Re; Luis A Trejo
Journal:  Sensors (Basel)       Date:  2019-09-03       Impact factor: 3.576

2.  A Comparison Study of Classifier Algorithms for Cross-Person Physical Activity Recognition.

Authors:  Yago Saez; Alejandro Baldominos; Pedro Isasi
Journal:  Sensors (Basel)       Date:  2016-12-30       Impact factor: 3.576

3.  Impact of Sliding Window Length in Indoor Human Motion Modes and Pose Pattern Recognition Based on Smartphone Sensors.

Authors:  Gaojing Wang; Qingquan Li; Lei Wang; Wei Wang; Mengqi Wu; Tao Liu
Journal:  Sensors (Basel)       Date:  2018-06-18       Impact factor: 3.576

4.  Comparison of Different Sets of Features for Human Activity Recognition by Wearable Sensors.

Authors:  Samanta Rosati; Gabriella Balestra; Marco Knaflitz
Journal:  Sensors (Basel)       Date:  2018-11-29       Impact factor: 3.576

5.  Sport-Related Human Activity Detection and Recognition Using a Smartwatch.

Authors:  Zhendong Zhuang; Yang Xue
Journal:  Sensors (Basel)       Date:  2019-11-16       Impact factor: 3.576

6.  Understanding LSTM Network Behaviour of IMU-Based Locomotion Mode Recognition for Applications in Prostheses and Wearables.

Authors:  Freddie Sherratt; Andrew Plummer; Pejman Iravani
Journal:  Sensors (Basel)       Date:  2021-02-10       Impact factor: 3.576

7.  Intelligent Diagnosis Based on Double-Optimized Artificial Hydrocarbon Networks for Mechanical Faults of In-Wheel Motor.

Authors:  Hongtao Xue; Ziwei Song; Meng Wu; Ning Sun; Huaqing Wang
Journal:  Sensors (Basel)       Date:  2022-08-22       Impact factor: 3.847

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.