Literature DB >> 35336522

Multilabel Classification Methods for Human Activity Recognition: A Comparison of Algorithms.

Athanasios Lentzas1, Eleana Dalagdi1, Dimitris Vrakas1.   

Abstract

As the world's population is aging, and since access to ambient sensors has become easier over the past years, activity recognition in smart home installations has gained increased scientific interest. The majority of published papers in the literature focus on single-resident activity recognition. While this is an important area, especially when focusing on elderly people living alone, multi-resident activity recognition has potentially more applications in smart homes. Activity recognition for multiple residents acting concurrently can be treated as a multilabel classification problem (MLC). In this study, an experimental comparison between different MLC algorithms is attempted. Three different techniques were implemented: RAkELd, classifier chains, and binary relevance. These methods are evaluated using the ARAS and CASAS public datasets. Results obtained from experiments have shown that using MLC can recognize activities performed by multiple people with high accuracy. While RAkELd had the best performance, the rest of the methods had on-par results.

Entities:  

Keywords:  activity recognition; ambient sensors; ensemble learning; multilabel classification; smart home

Mesh:

Year:  2022        PMID: 35336522      PMCID: PMC8955852          DOI: 10.3390/s22062353

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

Human activity recognition (HAR) is a research area with increased scientific interest. HAR is the task of correctly recognizing human actions and activities. Its applications vary from athlete and patient monitoring [1,2] to elderly care [3,4]. As IoT devices have become more accessible, solutions that use activity recognition are easier to deploy in smart home installations. The main purpose of an activity recognition system is to correctly recognize human actions and inform any agent interested about those actions. Recognizing performed activities is a process that, in most cases, happens in real time, as the results are used immediately. The majority of published works focus on recognizing the actions of a single individual [5,6]. While recognizing the activity of only one person is important when monitoring elderly people living alone, it is also critical to recognize activities performed by multiple humans operating in the same space. Information about multiple actions and the way humans interact while performing those activities could provide additional insight about the actions’ context. Machine learning approaches are a common way to recognize actions performed by humans [4,6]. In a HAR system, usually, multiple activities are recognized; thus, recognizing multiple activites is addressed as a multiclass classification problem [7]. Given a sensory input, the trained model assigns one class (i.e., activity) to that specific input. Multiclass classification algorithms have increased accuracy [8] when only a single person is present. When more than one individual is acting, the problem must be reformed as a multilabel classification problem. Multilabel classification (MLC) is a classification task where multiple labels may be assigned to each input [9]. Instead of selecting precisely one class out of many, as multiclass classification does, MLC returns one or more classes. It is worth mentioning that there is no constraint on how many classes an input can be assigned to. Treating HAR as a multilabel classification problem will allow activity recognition for residents acting concurrently. Sensory information, containing activity-related data for multiple people acting concurrently, is used as input. The purpose of this paper is to perform a benchmark on multilabel classification algorithms used on HAR. Implemented classifiers were trained to recognize a broad set of activities using multiple ambient sensors (i.e., pressure, contact, etc.). Section 1 is a brief introduction to HAR and multilabel classification. Related work is briefly presented in Section 2, followed by the methodology used in our work in Section 4. Thereafter, the results obtained are presented and discussed. At the end, conclusions are drawn and future work is discussed.

2. Related Work

Human activity recognition in smart house installations with a single resident is a research area with rapid growth. Several machine learning techniques have been employed in order to address the problem [4]. Random forests have been proven an effective solution [10,11]; they provide robust and explainable results and are easy to implement and train. Hidden Markov models (HMMs) are also a common approach [12,13]; HMMs are not only used for activity recognition, but for absence and home visit detection as well [13]. Deep learning models specifically developed for HAR [14,15] have been gaining increased interest over the past years. The popularity of those models is due to their ability to handle sequential data and time series, the main form of data generated by ambient sensors used in smart home installations. Despite the fact that machine learning algorithms can recognize performed actions with impressive accuracy, their performance is sub-par when more than one people interact in the same environment. Benchmarks on multi-resident activity recognition [16,17] provide a baseline for researchers that want to address multi-resident activity recognition. Thus, state-of-the-art algorithms, such as the hidden Markov model (HMM), recurrent neural networks (RNN), k-nearest neighbors (KNN), conditional random field (CRF), random forests (RF), and feed-forward neural networks (FFNN), were evaluated thoroughly in this study. Additionally, sensor placement was evaluated, as the sensor network is important for activity recognition [17]. In order to tackle the multiple residents concurrent activity recognition problem, ensemble techniques have been proposed in the literature. A factorial hidden Markov model (HMM) that models two different chains (one for each resident) was proposed by [18]. An ensemble method based on hidden Markov models was presented, aiming to address multi-resident activity recognition, by [19]; the proposed technique introduced a mixed-dependency model, dealing with the complexity of multiple residents acting concurrently. Combined-label HMMs were also evaluated in [20], where activity labels and observation labels of different residents were combined, generating a sequence of observations, recognized by a conventional HMM. Additionally, a model where activities are linked at each time step was proposed. Both models, combined-label and linked activities, outperformed the baseline HMM models. As smart homes are dynamic environments not limited to their residents, the need to detect visitors, occupants, and even pets is crucial [20,21]. Calculating entropy measures and comparing them with the standard deviation, visitors can be detected and identified [22]. This method can help activity recognition systems to focus only on sensor activations that are relevant to the main residents of the house. Incremental decision trees have also been employed [23]. While the results were promising, authors identified that further work is needed to obtain results that are on par with single-resident activity recognition. Deep learning is also often explored as a technique able to recognize activities performed by multiple residents. Recurrent neural networks (RNNs) were evaluated using not only real-world data, but also synthetic data generated with a generative adversarial network [24]. Multi-task learning, i.e., regarding the activity of each resident as a learning task, was paired with zero-shot learning to recognize previously unseen activities [25]. Multi-label classification, while extensively used on multiple domains, such as text [26], video [27], and image [28] classification, etc., have only recently drawn the attention of researchers to the activity recognition field. A comparative experimental study between two well-known MLC algorithms [29] proved that MLC can address concurrent activity recognition effectively. Interaction between actions performed by different people can be identified by transforming the multi-resident activity recognition problem into an MLC problem. Experiments have shown that random forests, when used as base classifiers on a MLC algorithm, have an impressive performance [30]. An evaluation of multilabel classification was performed by [31]. The authors use frequent item mining, a technique commonly used in data mining, in order to generate frequent itemsets as features for each activity. The generated features were used to evaluate MLC algorithms such as multilabel KNN, labelset, and decision trees. Treating multi-resident HAR as a multi-label classification problem while transforming it to a multi-class problem was investigated by [32]. Decision trees were used as the base classifier and label combination was employed for the problem transformation. Binary relevance has been proposed in the literature for multilabel HAR [33]. The results were competitive with approaches transforming the problem into a classic classification problem. The classifier chains approach considers activity correlation, an important aspect of activity recognition [34]. This proposed approach was able to find underlying correlations between activities and reduce the classification time.

3. Our Contribution

As already discussed in the previous sections, multi-resident human activity recognition is a research area that attracts a lot of attention. Multi-class classification techniques, while they yield good results, are far from providing a robust and reliable solution. On the other hand, multilabel classification approaches seem to better recognize concurrent activities, but they have not been investigated to the same extent as multi-class classification. In our work, we aim to provide a baseline for future research on multilabel classification in multi-resident activity recognition. This will allow future researchers to compare their results with results obtained from three well known-algorithms (considered state-of-the-art). Two datasets were used, ARAS and CASAS, which are extensively used on multi-resident activity recognition. Furthermore, our results are compared with multilabel and multi-class algorithms used on the same problem. This is to identify whether or not the experimentally evaluated algorithms can provide similar or better results. Lastly, has never been evaluated on a multi-resident activity recognition problem before (based on our search); thus, our work reports results not available in the literature.

4. Methodology

MLC can help address the problem of multi-resident activity recognition, as the trained model is able to learn from a diverse set of sensors and assign each example to multiple classes. In our work, several state-of-the-art (, classifier chain, binary relevance) MLC algorithms are evaluated. Algorithms were chosen based on previous benchmarks [29,35]. All algorithms were implemented using Python and Scikit-Learn [36]. The features used for training were the values of each sensor, recorded every second. No sliding windows were applied on the datasets.

4.1. RAkELd

The random k-labelsets algorithm () [37] is an ensemble method for multilabel classification. Each member of the ensemble is constructed by considering a relatively small, random subset of labels. A single label classifier is trained for the prediction of each element in the powerset of the subset. The benefit of applying single label classifiers to sub-tasks with a feasible number of labels is that label correlation is taken into account. Constructing the labelsets can be realized by using two different methods: disjoint and overlapping labelsets. In our work, disjoint labelsets were constructed. The label space was partitioned to equal partitions of size k. For each partition, a label powerset classifier was trained, and the prediction was the sum of all trained classifiers. Labelset size k was set to 3, as, according to authors, it allows the best results. An extensive evaluation was performed on the labelset size [37], and results showed that a smaller number affects the performance of the algorithm positively, with 3 being the value that leads to the best results on multiple datasets. In our work, the labelset was evaluated with different values (k = 2,3,4,5) and the value the authors proposed (k = 3) was the optimal. Two different classifiers were used during experimentation: multilayer perceptron (MLP) and random forests. In order to fine tune the parameters of the two used classifiers, different techniques were used. The random forest classifier was tuned using randomized parameter optimization [38]. For every hyperparameter that required tuning, a set or range of values was given. A random search was then performed on these distributions M times. A model was trained for each combination and the hyperparameters for the best model were returned. MLP’s hyperparameters were chosen using quantum genetic algorithms (QGA) [39]. Similarly to randomized parameter optimization, a range or set of values was given. The search space was then explored using QGA. Each set of parameters was represented as a chromosome consisting of qudits. A model was trained for each set of parameters. The evaluation was based on the accuracy of the trained model. Non-optimal chromosomes had a chance to re-initialize instead of converging to the best solution. This method allows for a fast convergence to the best solution while performing a random search, thereby avoiding the local optimal.

4.2. Binary Relevance

Binary relevance [40] attempts to transform a multilabel classification problem with L labels into L binary classification problems. The same classifier is used on all sub problems (random forests in our experiments). The final prediction of binary relevance is the union of all individual label classifiers. Binary relevance is a rather simple algorithm that yields promising results. The simplicity of the technique allows for low modeling complexity, linear to the number of class labels in the label space.

4.3. Classifier Chain

Another widely used MLC ensemble method is the classifier chain method [41]. A classifier chain constructs a chain of binary classifiers. The number of classifiers constructed is the same as the number of labels in the dataset. As the classifiers are linked in a chain format, a single classifier can utilize the prediction of all previously trained models. Therefore, a classifier chain reconstructs the multilabel problem into a a multi-class problem, where each label combination is a different class and the number of classifiers is equal to the total number of classes. In our experiments, the binary classifier chosen was random forests, tuned with randomized parameter optimization.

4.4. ARAS Dataset

The first dataset used in our work is the ARAS (activity recognition with ambient sensing) human activity recognition dataset [42]. The ARAS dataset contains real-life data from two houses with multiple residents. More specifically, 2 young males (25) were living in house A, and a married couple (35) was living in house B. Several ambient sensors were employed to gather data. Force and pressure sensors were placed under the bed and couches, photocells were placed in the drawers, wardrobes, and refrigerator, contact sensors were placed on the door frames, shower cabin, and cupboards, proximity sensors were placed on the chairs, closet, and taps, sonars were placed on the walls and door frames, temperature sensors were installed near the kitchen and oven, and infrared sensors were placed near the TV. The sensors and the actions associated with them are summarized in Table 1. As a result, 27 different activities were recognized (Table 2).
Table 1

Sensors and their associated actions.

SensorsActions
Force and pressure sensorsSleeping, sitting, napping
Photo sensorOpening drawers and wardrobes
Contact sensorsOpening/closing doors, cupboards
SonarPresence detection
Temperature sensorsCooking
InfraredWatching TV
Table 2

Activities available in the ARAS dataset.

Activities
Going outPreparing breakfastHaving breakfast
Preparing lunchHaving lunchPreparing dinner
Having dinnerWashing dishesMaking snack
SleepingWatching TVStudying
Having showerToiletingNapping
Using internetReading bookLaundry
ShavingBrushing teethTalking on the phone
Listening to musicCleaningHaving conversation
Changing clothesHaving clothesOther
Residents did not follow a specific activity plan during data acquisition, but continued to behave naturally. The only interaction with the system was battery replacement of the sensors and labeling the ground truth manually. Sensors were sampled at 10 Hz (the IR sensor was sampled at 100 Hz). Sensor events were logged every second, resulting in 86,400 data points for each day.

4.5. CASAS Dataset

The CASAS dataset contains data from different participants acting as residents of the same house. Each resident performed 15 unique activities (Table 3). Some activities were performed individually (ind.) while some of them required the cooperation of the participants (co-op). Motion sensors were placed on the ceiling, monitoring movement around the room, and contact switches were placed on objects, registering use events. The apartment was always occupied by two participants and 40 volunteers were recruited.
Table 3

Activities performed by each resident on the CASAS dataset.

Person APerson B
Filling medication dispenser (ind.)Hanging up clothes (ind.)
Moving furniture (co-op)Moving furniture (co-op)
Watering plants (ind.)Reading magazine (ind.)
Playing checkers (co-op)Sweeping floor (ind.)
Preparing dinner (ind.)Playing checkers (co-op)
Reading magazine (ind.)Setting the table (ind.)
Gathering and packing picnic food (ind.)Paying bills (co-op)
-Gathering and packing picnic supplies (co-op)

4.6. Data Preprocessing

Before training the classifiers mentioned, the data were preprocessed. The main purpose was to transform the dataset into a format that could be used to train the MLC models. As the encoding of the ARAS dataset labels is sequential and common for all the participants (i.e., sleeping is encoded as 11, washing dishes as 9, etc.), a transformation was applied to distinguish actions performed by different persons. Activities performed by the second individual were re-labeled with sequential numbers by adding 27 to the already existing value. As a result, a “new” set of 27 activities was introduce andd applied only to the second person, while the initial set was only referenced by the first participant. Class imbalance was also a problem we had to address while preprocessing the dataset. As activities do not occur with the same frequency, i.e., some activities are performed more times than others, the dataset was imbalanced. As seen in Figure 1 and Figure 2, there is a big difference between the major and the minor class in each house. In order to balance the dataset, we upscaled minority classes. To augment our data, the MLSMOTE [43] algorithm (multilabel synthetic minority over-sampling technique), an extension of the SMOTE [44] (synthetic minority over-sampling technique) focused on multilabel datasets, was employed.
Figure 1

House A class imbalance. Number of classes per resident (normalized) and percentage of each class for the first house.

Figure 2

House B class imbalance. Number of classes per resident (normalized) and percentage of each class for the first house.

Further inspecting class imbalance figures (Figure 1 and Figure 2), one can identify the major and minor classes. In both houses, activities 2 (Going out) and 11 (Sleeping) are the activities mostly performed, followed by watching TV (12) and studying (13). One of the residents in both houses had a major class (going out—2) that represents approximately 55% of the dataset. It is worth mentioning that not all classes are present for all participants. While upscaling the minority classes, activities never recorded were ignored. As the CASAS dataset has sensor names and values recorded as strings, e.g., ”MD2” as the sensor ID and ”ON” as the sensor event, the data had to be transformed to a more appropriate form. Following the same approach as [45], each sensor was assigned a numerical ID, while values were converted to ”1” or ”0”, based on whether the sensor was triggered or not.

4.7. Performance Metrics

Two different metrics were used for evaluation: F1 score and Hamming loss. Before defining the metrics, the type of errors and predictions a classifier can make must be defined: False Positives (FP): The classifiers predicts a label that is not correct; False Negatives (FN): The classifier misses a label that exists in the example; True Positives (TP): The classifier correctly predicts the existence of a label; True Negatives (TN): The classifier correctly predicts the non-existence of a label. Using the above information, we can calculate the metrics needed. F1 score is the harmonic mean of Precision (proportion of correct predictions among all predictions) and Recall (proportion of examples of a certain class predicted as members of the class). Equations for each metric can be seen in Equation (1). Hamming loss is the fraction of labels that are incorrectly predicted [9]. While on multi-class classification, the Hamming loss is the Hamming distance between true and predicted labels, in MLC, the Hamming label penalizes only the individual classes. The Hamming loss is calculated using Equation (2), where N is the total number of data samples, L is the total number of available classes, is the target, and is the prediction. The xor operator (⊕) returns zero when the target and prediction are the same, and one otherwise. Since the metric is a loss function, the optimal value is zero and the upper bound is one.

5. Results

The experiments carried out and results are presented in this section. The evaluation was based on 10-fold cross validation. Using the 10-fold cross validation technique is a common practice when evaluating the performance of machine learning models. This technique has numerous advantages, such as reduced bias and variance, etc. The dataset was partitioned into 10 subsets. During each epoch, one subset was used for validation, while the rest (9) were used as training data. The Hamming loss and F1 score were logged after each training epoch.

5.1. ARAS Dataset

All proposed models were trained using a whole month’s data. Similar attempts to address the problem trained a different model for each day and averaged the individual results [29]. The average F1 score and Hamming loss for each classifier and house can be seen in Table 4 and Table 5, respectively. Observing the results, one can see that with MLP classifier had the best F1 score and Hamming loss for both houses. Although had a better performance and lower Hamming loss, the difference from the rest of the classifiers was not significant.
Table 4

Average F1 score per classifier. Inside the parentheses is the confidence interval (96% confidence). The best value can be seen in bold.

ClassifierHouse AHouse A
RAkELd (MLP)0.676 (0.64–0.70)0.909 (0.88–0.93)
Classifier chain0.671 (0.64–0.69)0.903 (0.87–0.92)
Binary relevance0.674 (0.64–0.7)0.904 (0.88–0.92)
RAkELd (Random forest)0.674 (0.64–0.7)0.904 (0.88–0.92)
Table 5

Average Hamming loss per classifier (lower is better). Inside the parentheses is the confidence interval (96% confidence). The best value can be seen in bold.

ClassifierHouse AHouse B
RAkELd (MLP)0.024846898 (0.022–0.024)0.008057439 (0.0059–0.01)
Classifier chain0.026821258 (0.024–0.029)0.008799636 (0.0065–0.0088)
Binary relevance0.02482958 (0.022–0.027)0.00846863 (0.0065–0.0084)
RAkELd (Random forest)0.024929984 (0.022–0.027)0.008469172 (0.0065–0.0084)
As the activity recognition problem is, by its nature, unbalanced, experiments were also performed before applying oversampling. The results, as seen in Table 6 and Table 7, do not differ a lot compared to a balanced dataset. Therefore, multi-label classification can provide robust results on the activity recognition problem, even when the number of examples differs a lot.
Table 6

Average F1 score per classifier trained on an unbalanced dataset. Inside the parentheses is the confidence interval (96% confidence). The best value can be seen in bold.

ClassifierHouse AHouse B
RAkELd (MLP)0.679 (0.64–0.72)0.911 (0.89–0.93)
Classifier chain0.673 (0.64–0.70)0.905 (0.9–0.93)
Binary relevance0.675 (0.64–0.70)0.906 (0.88–0.92)
RAkELd (Random forest)0.676 (0.64–0.71)0.911 (0.89–0.93)
Table 7

Average Hamming loss per classifier (lower is better) trained on an unbalanced dataset. Inside the parentheses is the confidence interval (96% confidence). The best value can be seen in bold.

ClassifierHouse AHouse B
RAkELd (MLP)0.024651636 (0.022–0.028)0.007962516 (0.0059–0.01)
Classifier chain0.026690177 (0.024–0.03)0.008666588 (0.0065–0.01)
Binary relevance0.026660177 (0.024–0.03)0.00837087 (0.0065–0.01)
RAkELd (Random forest)0.02475237 (0.022–0.028)0.008020018 (0.006–0.01)
An important observation is the significant difference between the results of the two houses. Classifiers evaluated on the second house yielded better results. This observation is justified by the difference in activities performed. Residents of the second house performed less distinct activities, compared to residents of the first house (Figure 1 and Figure 2). This is important, as classification performance is related to the number of activities performed. Further analysing the results, the standard deviation of the F1 score was calculated for each algorithm. For both houses, had the lower standard deviation, 0.041 for the first house ( with MLP) and 0.034 for the second house ( with random forest). The results for each epoch for all the classifiers can be seen in Figure 3 for the first house and Figure 4 for the second house.
Figure 3

Classification results for House A. Hamming loss for each classifier per validation epoch and F1 score for each classifier per validation epoch.

Figure 4

Classification results for House B. Hamming loss for each classifier per validation epoch and F1 score for each classifier per validation epoch.

Performing the same experiments on a daily basis, instead of using the data from a whole month, provided the results seen in Table 8 and Table 9. , with both classifiers evaluated (MLP and RF), had the best overall performance. The results were consistent and no major deviations were observed between different days. Furthermore, the results of using MLP are compared with the results obtained on the same dataset by [29] (Table 10). The comparison is based on the only common metric, Hamming loss. By comparing the results, one can identify that performed better for both houses.
Table 8

Average F1 score and Hamming loss per day for House A. Best value per day can be seen in bold.

DayRAkELd (MLP)Classifier ChainBinary RelevanceRAkELd (Random Forest)
F1HammingF1HammingF1HammingF1Hamming
10.8080.0360.7910.04120.6770.0555 0.81 0.0328
2 0.86 0.033 0.8580.03660.8410.03630.8530.0454
3 0.769 0.036 0.6960.04980.70.0495 0.769 0.0375
40.811 0.029 0.7160.05030.7890.0333 0.82 0.0298
50.7950.0250.7320.03990.7230.0357 0.826 0.0219
60.807 0.024 0.7350.03730.7210.0372 0.823 0.0255
70.8110.0360.7490.04450.7330.0475 0.865 0.0253
8 0.848 0.02 0.7640.03270.6920.04570.8370.0216
90.782 0.03 0.6880.04710.7070.0386 0.793 0.035
10 0.837 0.026 0.7260.04260.8290.02580.8350.0269
110.786 0.042 0.7280.05070.8050.0435 0.794 0.042
120.7980.0450.7760.0440.8060.0451 0.817 0.0378
130.7810.0390.7630.04450.7960.0273 0.834 0.0332
140.7910.0640.7520.07130.7810.0706 0.82 0.0515
150.7580.0680.7220.06450.7620.0687 0.8 0.0516
160.8220.0270.730.04060.7870.0339 0.846 0.0242
17 0.819 0.027 0.7110.04870.8140.03250.8060.0312
18 0.826 0.033 0.7740.04130.7830.04460.8130.0366
190.8110.0350.7330.04120.7980.0342 0.828 0.0297
200.786 0.063 0.7290.07070.7770.06 0.809 0.066
21 0.804 0.0470.7490.05730.7470.04410.794 0.0415
220.8280.0390.8430.04090.8220.043 0.87 0.0361
230.780.0390.7150.05560.770.0403 0.786 0.0338
240.8040.0330.7650.0371 0.828 0.0262 0.8170.0282
250.7820.0510.7080.06830.7730.0586 0.815 0.0477
26 0.807 0.039 0.6480.05880.7550.04810.7950.0435
270.8490.030.8660.0246 0.917 0.0191 0.8910.0235
280.8160.040.7330.04880.82 0.0341 0.833 0.0385
290.8380.0430.7750.0573 0.862 0.0377 0.8520.0408
300.8810.0280.8770.02950.8720.0298 0.917 0.0226
Table 9

Average F1 score and Hamming loss per day for House B. Best value per day can be seen in bold.

DayRAkELd (MLP)Classifier ChainBinary RelevanceRAkELd (Random Forest)
F1HammingF1HammingF1HammingF1Hamming
10.8020.0450.7830.04570.7740.0466 0.812 0.044
20.902 0.0216 0.903 0.02230.8810.0220.897 0.0216
30.9580.0110.957 0.0108 0.9520.0113 0.96 0.0111
4 0.858 0.0236 0.8430.02820.8250.02790.8520.0276
5 0.91 0.0248 0.91 0.0248 0.9010.02490.9080.0255
60.934 0.0134 0.940.01350.9290.0136 0.941 0.0134
7 0.903 0.0215 0.9020.0230.8870.02330.8910.0232
80.948 0.0167 0.9480.01750.9390.0176 0.95 0.0175
90.9470.01070.9460.01210.944 0.0106 0.951 0.0113
10 0.924 0.0145 0.8950.01960.8980.01750.9010.0188
110.92 0.0185 0.953 0.01890.9010.01890.92 0.0185
12 0.922 0.0172 0.9180.01750.910.01730.92 0.0172
13 0.96 0.0096 0.9370.01420.9510.01060.9560.0114
140.984 0.0062 0.9760.00740.9810.0065 0.986 0.0074
150.926 0.0194 0.9240.02070.920.0202 0.928 0.0202
160.9080.0173 0.919 0.0164 0.8990.01790.9120.0177
17 0.887 0.02260.8810.0230.87 0.0219 0.8850.0235
180.96 0.0096 0.954 0.0096 0.951 0.0096 0.961 0.0096
19 0.928 0.015 0.9170.01690.9080.01660.920.0165
200.951 0.0166 0.982 0.0170.946 0.0166 0.952 0.0166
21 0.953 0.01530.9290.01990.952 0.0125 0.950.0157
22 0.953 0.0131 0.953 0.0104 0.9410.0130.950.0131
23 0.914 0.0256 0.906 0.0256 0.899 0.0256 0.913 0.0256
24 0.903 0.0174 0.880.0220.8780.01860.8920.0194
25 0.96 0.01210.9540.01220.9510.01240.9570.0124
26 0.952 0.0126 0.9480.01450.9430.01390.9510.0139
27 0.9 0.02140.8880.0231 0.9 0.02080.8970.0216
280.8470.02620.830.0310.846 0.0254 0.86 0.0245
290.932 0.0209 0.930.0217 0.933 0.02050.9290.0235
300.9180.01690.9050.02040.914 0.0163 0.925 0.0163
Table 10

Hamming loss comparison between (MLP) and results presented in [29]. Best value per day can be seen in bold.

DayHouse AHouse B
RAkELd (MLP)[29]RAkELd (MLP)[29]
1 0.036 0.107 0.045 0.058
2 0.033 0.083 0.0216 0.043
3 0.036 0.09 0.011 0.023
4 0.029 0.119 0.0236 0.039
5 0.025 0.123 0.0248 0.031
6 0.024 0.1830.0134 0.009
7 0.036 0.116 0.0215 0.018
8 0.02 0.104 0.0167 0.02
9 0.03 0.192 0.0107 0.014
10 0.026 0.112 0.0145 0.023
11 0.042 0.105 0.0185 0.049
12 0.045 0.101 0.0172 0.021
13 0.039 0.077 0.0096 0.011
14 0.064 0.1070.0062 0.006
15 0.068 0.158 0.0194 0.044
16 0.027 0.149 0.0173 0.054
17 0.027 0.083 0.0226 0.051
18 0.033 0.1020.0096 0.008
19 0.035 0.164 0.015 0.022
20 0.063 0.1340.0166 0.014
21 0.047 0.207 0.0153 0.027
22 0.039 0.145 0.0131 0.028
23 0.039 0.182 0.0256 0.039
24 0.033 0.095 0.0174 0.026
25 0.051 0.124 0.0121 0.018
26 0.039 0.12 0.0126 0.015
27 0.03 0.0860.0214 0.018
28 0.04 0.096 0.0262 0.064
29 0.043 0.079 0.0209 0.025
30 0.028 0.05 0.0169 0.041
Analyzing the results of the experiments, one can identify activities that are more prone to misclassification. For House A, the hardest activities to recognize were having lunch and having conversation for residents 1 and 2, respectively. reparing dinner was the least recognized activity for resident 1 in House B and having conversation was least recognized for resident 2. It is worth mentioning that, except for resident 1 (House A), the rest of the least recognized classes were under-sampled. As a result, there were training epochs where the class was not represented in the training data.

5.2. CASAS Dataset

Experiments on the CASAS dataset were performed similarly to the ARAS dataset. The same classifiers, binary relevance, classifier chain, RAkEL with random forests, and RAkEL with MLP were used. The evaluation of the CASAS dataset was based on 10-fold cross validation in order to reduce bias and variance. Using the whole CASAS dataset, i.e., 26 days, provided the results seen in Table 11. RAkEL with MLP classifier had the best overall performance when the classifiers were trained and evaluated using the entire dataset. The rest of the classifiers were on par regarding performance, making them a viable solution for multi-resident activity recognition. Similarly to the ARAS dataset experiments, each classifier was evaluated using daily data (Table 12). RAkEL had the best performance with, results the being consistent with the ARAS dataset.
Table 11

Average F1 score & Hamming loss per classifier (CASAS dataset). Inside the parentheses is the confidence interval (96% confidence). Best values can be seen in bold.

ClassifierF1Hamming Loss
RAkELd (MLP)0.912 (0.89–0.92)0.0080326 (0.0072–0.01)
Classifier chain0.887 (0.87–0.9)0.009235 (0.0075–0.012
Binary relevance0.89 (0.88–0.91)0.0089681 (0.0073–0.011)
RAkELd (Random forest)0.9 (0.9–0.91)0.008220018 (0.0058–0.01
Table 12

Average F1 score and Hamming loss per day for the CASAS Dataset. Best values per day can be seen in bold.

DayRAkELd (MLP)Classifier ChainBinary RelevanceRAkELd (Random Forest)
F1HammingF1HammingF1HammingF1Hamming
10.860.0330.7320.03990.7930.0355 0.864 0.0301
2 0.886 0.0246 0.8260.0350.7680.03240.8750.0296
30.7925 0.0285 0.7980.0470.7820.0373 0.925 0.0338
4 0.806 0.0214 0.7590.03140.7980.02570.8050.0337
50.8270.0340.7620.03640.7920.0269 0.863 0.0211
60.8090.02640.7630.03910.7680.035 0.89 0.0249
70.7870.0250.7490.02980.7440.0367 0.866 0.0154
8 0.942 0.012 0.7760.02990.8470.02590.8750.0155
90.8280.0340.8560.03750.73 0.0315 0.877 0.0315
10 0.926 0.02050.8410.0250.8050.02570.872 0.014
110.8520.02230.7310.03040.7210.0303 0.871 0.0189
120.823 0.0119 0.7970.0340.8330.0331 0.902 0.0295
13 0.917 0.0213 0.8280.02230.7970.03430.9110.0285
14 0.925 0.0220.8680.02350.7280.03750.844 0.015
150.854 0.0225 0.8080.03030.7870.0385 0.909 0.0306
160.902 0.025 0.7550.03840.880.0221 0.911 0.03
17 0.97 0.013 0.7810.03620.7990.02580.8690.0328
18 0.916 0.023 0.8930.02690.8990.03450.8710.026
19 0.929 0.016 0.840.02870.780.02410.8350.0201
200.8890.0230.8150.03020.730.0224 0.904 0.0152
210.8660.03450.8230.0250.8560.0222 0.89 0.0107
22 0.93 0.0103 0.8550.03780.7350.0350.8990.027
23 0.89 0.0226 0.8870.02950.810.02750.8870.0324
24 0.914 0.0218 0.8530.0350.7960.0220.8670.0227
250.8580.03920.790.03610.80.0371 0.863 0.0232
260.8470.05630.7270.03450.823 0.0203 0.909 0.0336

5.3. Multilabel and Multi-Class Comparison

Multi-resident activity recognition, as already discussed in Section 2, can be addressed both as a multilabel and multi-class classification problem. In order to further support our position that MLC can be used on multi-resident activity recognition without sacrificing accuracy, a comparison between the two approaches is needed. We did not run any experiments with multi-class models, but we compared our results with results already published in the literature [16]. As seen in Table 13, multilabel classification outperformed multi-class classification. The average F1 score was better for MLC approaches with both datasets. The multi-class classifiers with the best overall performance was recurrent neural networks for the CASAS dataset, and random forests for ARAS. It is worth mentioning that on the ARAS dataset, for House A, there was a significant difference between multilabel and multi-class classification. This is important, as the aforementioned house had increased difficulty, mainly because of the diverse set of activities performed.
Table 13

F1 score comparison between multilabel and multi-class classification approaches. Best values are in bold.

MethodARAS House AARAS House BCASAS
Multilabel 0.679 0.911 0.902
Multi-class [16]0.59870.87960.8227

6. Conclusions

In this work, an experimental evaluation of RAkEL, classifier chain, and binary relevance methods for multi-resident human activity recognition was performed. Experiments were performed on the ARAS and CASAS datasets, two activity recognition datasets with two residents acting concurrently. RAkEL had the best overall performance on both datasets, with the rest of the evaluated classifiers yielding on-par results. When a multi-layered perceptron was used as a base classifier on RAkEL, the performance of the algorithm was improved. The ARAS dataset was analyzed and the findings showed that it suffers from the class imbalance problem. In order to address this, the minority classes were upscaled before training the classifiers. Balancing the dataset was an important step during data preprocessing, as the major classes represented approximately 90% of the whole data for both houses. Experiments were performed on both a balanced and unbalanced dataset, without any significant difference between the results. Observing the activity distribution on the dataset, it was observed that residents follow a different activity pattern. Although the dataset has records for 27 different activities, not all of them are present in each house. As house B residents performed less distinct activities, the trained models had a better overall performance. Results were also compared with multi-class classification methods used on the same set of data, as well as multilabel classification methods already published in the literature. The comparison showed that RAkELs performance is on par with, if not better than, different multilabel classification approaches, outperforming multi-class classification techniques. This can provide a baseline for future comparisons of multilabel classification solutions used on the multi-resident activity recognition problem. Future work will be focused on identifying correlations between different activities in multi-resident environments. Some activities could potentially be mutually exclusive (i.e., both residents toileting), or inclusive (i.e., both having a conversation, lunch, etc.). That relationship could be identified and thereby further enhance the performance of classification. As the datasets used in our work do not contain mutually exclusive activities, this aspect is not investigated in this work. Furthermore, more MLC algorithms should be evaluated. While this study extends already published works on multi-resident activity recognition, there are a lot of techniques that have not yet been evaluated, such as ML-KNN [46], CLARE decision trees [47], and deep neural networks adapted for MLC problems. While our work is tested on two different datasets, proving the robustness of the evaluated algorithms, there are more methods available in the literature. Additionally, hierarchical multilabel classification should be explored. As some activities are correlated in a hierarchical order, e.g., preparing dinner always precedes eating dinner, it is a promising idea to create efficient structures that represent that hierarchical order.
  6 in total

1.  Recognizing independent and joint activities among multiple residents in smart environments.

Authors:  Geetika Singla; Diane J Cook; Maureen Schmitter-Edgecombe
Journal:  J Ambient Intell Humaniz Comput       Date:  2010-03-01

2.  Detection of daily activities and sports with wearable sensors in controlled and uncontrolled conditions.

Authors:  M Ermes; J Pärkka; J Mantyjarvi; I Korhonen
Journal:  IEEE Trans Inf Technol Biomed       Date:  2008-01

3.  ML-Net: multi-label classification of biomedical texts with deep neural networks.

Authors:  Jingcheng Du; Qingyu Chen; Yifan Peng; Yang Xiang; Cui Tao; Zhiyong Lu
Journal:  J Am Med Inform Assoc       Date:  2019-11-01       Impact factor: 4.497

Review 4.  Machine Learning Approaches for Activity Recognition and/or Activity Prediction in Locomotion Assistive Devices-A Systematic Review.

Authors:  Floriant Labarrière; Elizabeth Thomas; Laurine Calistri; Virgil Optasanu; Mathieu Gueugnon; Paul Ornetti; Davy Laroche
Journal:  Sensors (Basel)       Date:  2020-11-06       Impact factor: 3.576

5.  Exploring Entropy Measurements to Identify Multi-Occupancy in Activities of Daily Living.

Authors:  Aadel Howedi; Ahmad Lotfi; Amir Pourabdollah
Journal:  Entropy (Basel)       Date:  2019-04-19       Impact factor: 2.524

6.  Multimodal wireless sensor network-based ambient assisted living in real homes with multiple residents.

Authors:  Can Tunca; Hande Alemdar; Halil Ertan; Ozlem Durmaz Incel; Cem Ersoy
Journal:  Sensors (Basel)       Date:  2014-05-30       Impact factor: 3.576

  6 in total
  1 in total

1.  Dynamic Segmentation of Sensor Events for Real-Time Human Activity Recognition in a Smart Home Context.

Authors:  Houda Najeh; Christophe Lohr; Benoit Leduc
Journal:  Sensors (Basel)       Date:  2022-07-21       Impact factor: 3.847

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.