Literature DB >> 31513592

Hierarchical multi-view aggregation network for sensor-based human activity recognition.

Xiheng Zhang1, Yongkang Wong2, Mohan S Kankanhalli2, Weidong Geng1.   

Abstract

Sensor-based human activity recognition aims at detecting various physical activities performed by people with ubiquitous sensors. Different from existing deep learning-based method which mainly extracting black-box features from the raw sensor data, we propose a hierarchical multi-view aggregation network based on multi-view feature spaces. Specifically, we first construct various views of feature spaces for each individual sensor in terms of white-box features and black-box features. Then our model learns a unified representation for multi-view features by aggregating views in a hierarchical context from the aspect of feature level, position level and modality level. We design three aggregation modules corresponding to each level aggregation respectively. Based on the idea of non-local operation and attention, our fusion method is able to capture the correlation between features and leverage the relationship across different sensor position and modality. We comprehensively evaluate our method on 12 human activity benchmark datasets and the resulting accuracy outperforms the state-of-the-art approaches.

Entities:  

Mesh:

Year:  2019        PMID: 31513592      PMCID: PMC6742398          DOI: 10.1371/journal.pone.0221390

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


1 Introduction

Human Activity Recognition (HAR) refers to the automatic detection of various physical activities performed by people in their daily lives [1]. It has been applied to practical scenarios such as smart environment [2], health care [3], and footstep detection [4]. Benefiting from ubiquitous computing and well-protected individual privacy, sensor-based HAR research has received increasing interests recently. In the sensor-based HAR task, the raw data from various modalities is collected and then utilized to infer useful contextual information for classifying activities. And there are two challenges lies in processing raw data. The first one is how to construct discriminative feature spaces from the heterogeneous sensor data. Focusing on this challenge, early methods leveraged human domain knowledge to feature engineering for HAR, and these properly designed white-box features are extracted based on different types of methods [5-9]. Recently, deep learning models have been brought significant impacts to HAR. Neural networks such as convolutional neural networks (CNNs) and recurrent neural networks (RNNs) are applied to learn black-box features from raw data [10-13]. We argue that these different types of features describe the contextual information of raw data from different viewpoints, and could be effectively integrated in a unified framework to take advantage of each other. The second challenge is how to integrate features from different feature spaces effectively, which can lead to more accurate and robust performance. We notice that the sensors are not only of different modalities, but also of different wearing positions on the human body. The heterogeneity in sensor modality and position inspired us to integrate views of feature spaces by a hierarchical context, i.e. from the aspect of feature level, position level and modality level. Therefore, in this paper, we propose a hierarchical multi-view aggregation network for HAR, which can effectively fuse white-box features with black-box features from different feature spaces. Moreover, we design three aggregation modules to construct a unified representation for multi-view features from the aspect of feature level, position level and modality level. In feature level aggregation, we apply a non-local operation augmented with the L2-norm to explores the correlation between different features and fuse them. In the position level aggregation, we take the correlation of different sensor positions into consideration by introducing the correlation feature, which can enhance the representation of each view and effectively improves the resulting accuracy. Finally, in the modality level aggregation, we conduct a soft attention mechanism to quantify the discrimination of each view and fuse them for classification. The main contribution of this paper is two-fold: We propose a Hierarchical Multi-View Aggregation Network (HMVAN) for HAR task, which targets at integrating features from various feature spaces. Specifically, our method constructs multi-view features for individual sensor from the point of view of white-box features and black-box features. These views are then embedded into a shared feature space and aggregated into a unified hierarchical representation. Compared with existing deep learning-based methods which mainly extracting black-box features from the raw data of different sensors, our model is much more effective in representing the discriminative information of human activities. We design three aggregation modules to integrate these views into a unified representation of multi-view features from the aspect of feature level, position level and modality level. Based on the idea of non-local operation and attention, our method is capable of capturing the correlation between features, and leveraging the relationship across different sensor position and modality, which effectively improves to the resulting recognition accuracy.

2 Related works

Sensor-based HAR task is usually formulated as a time-series segment classification problem. Segmentation by sliding windows, extraction of features followed by a classification constitute the standard pipeline for recognizing human activities. Previous works can be roughly divided into two categories: traditional feature-based methods and deep learning-based method. Conventional methods tend to using engineered features obtained by statistical process. Plötz et al. [5] calculated mean, standard deviation, energy and entropy for each source channel of accelerometers. And in [14], the authors extracted zero crossing, root mean square value, spectral energy, mean, variance, standard deviation, median and the sum of FFT (Fast Fourier Transformation) coefficients from accelerometers, gyroscopes and magnetometers. Chen and Shen [15] chose the feature set that includes mean, standard deviation, max, min, interquartile range, dynamic time warping distance, FFT coefficients and wavelet energy. Kwon et al. [16] explored how temporal structure can be add into distribution-based feature extraction schemes for improving performance of HAR applications. Although extracting and identifying relevant features is time-consuming, these approaches work relatively well, even when the data is scare and highly unbalanced [17]. Recently, deep learning methods are widely adopted for sensor-based HAR task for the capacity of extracting features without human domain knowledge. Convolutional neural network (CNN) has been widely used in many HAR works. The works such as [18] and [10], in which each axis of the sensor was treated as an independent channel and CNN was performed on every channel separately. But with 1D kernel, CNN can only capture local dependency over time. Therefore, in [19], different sensors were grouped by their position and 2D CNN was raised to capture both local dependency over time and spatial dependency over the sensors. Later in their work [20], they improved the networks by employing weight-sharing mechanism, which enabled the networks to learn modality-specific characteristics across multi-model sensor data. Jiang and Yin [21] assembled signal sequence of accelerometer and gyroscopes into a novel “activity image”, which enabled CNN to learn optimal features from the “activity image” for image classification task. In [22], pressure sensor data was converted into pressure distribution images, then CNN was used for transfer learning on the converted imagery data. Ravi et al. [23] introduced a deep learning method which combines convolutional features learned from inertial sensor data together with complementary information from a set of shallow features to enable accurate activity classification. Rueda et al. [24] presented three convolutional architecture to search for attributes that represent favorable signal segments for recognizing human activities. In addition, recurrent neural network (RNN) shows competitive results when applied to HAR task. In [2, 25, 26], LSTM (Long Short-Term Memory) cell is the most used in RNN-based architecture. Edel et al. [27] proposed a binarized-BLSTM-RNN model, of which the input, weight parameters and intermediate hidden layer output were all binary-valued. Vu et al. [28] developed a self-gated recurrent neural network, which reduced resource memory usage and computational cost. In [29], different convolutional and recurrent models were evaluated and bi-directional LSTMs outperformed the others. To better interpret the recurrent networks to gain insight into the models’ behavior, Zeng et al. [30] introduced temporal attention and sensor attention into RNN, adaptively focusing on important signals and sensor modalities. Moreover, there are some hybrid deep architectures which combines different models. An Early work is [31], which recommended a deep neural network for modeling the emission distribution of hidden Markov models. A good example that combines CNN and RNN was offered by [11]. It was proven that the performance of the proposed deepConvLSTM was better than single CNN. Yao et al. [32] proposed DeepSense, which integrates convolutional and recurrent neural networks to exploit local interactions of different sensory modalities. Zheng et al. [33] combined CNN with stacked auto-encoder while Liu et al. [34] combined CNN with restricted Boltzmann machine. The method proposed in [23] is most related to ours, where they combined a set of shallow features with those obtained from deep learning. Our method has two key differences. Firstly, we construct different views of feature spaces on multi-modal sensor data rather than simply directly extracting single view features. The “multi-view” fashion can better represent the specific property of each type of sensor data. Secondly, our method introduces three aggregation modules to integrate these views into a unified representation of multi-view features from the aspect of feature level, position level and modality level. Comparing with simply concatenating different features, our approach is capable of capturing the correlation between features, and leveraging the relationship across different sensor position and modality, which effectively improves to the resulting recognition accuracy.

3 Problem formulation

Sensor-based HAR task is usually formulated as a time-series segment classification problem. Given a HAR dataset S, we denote the samples generated by sliding window procedure as , in which , and represent the data of accelerometers, gyroscope and magnetometer respectively. In addition, , , and where P is the amount of positions of accelerometers, gyroscopes and magnetometers. Each sensor has 3 channels. The number of modalities V = 3. For each sample I, we aim to map the raw data into a unified representation for multi-view features h by the following function: where C(⋅) is the construction of multiple views. G, G and G is the feature level, position level and modality level aggregation, respectively. Given the unified representation h, we simultaneously optimize the network by minimizing a loss function L to shorten the distance between the predicted label and ground truth. In the following sections, we will present details of the construction of multi-view features, the design and optimization of the aggregation module of each layer.

4 Hierarchical multi-view aggregation network

Our proposed hierarchical multi-view aggregation network is a three-layer multi-view framework, as illustrated in Fig 1. First of all, on the bottom layer, four views of features, including time domain, frequency domain, time-frequency and visual domain features, are extracted to characterize the measurement of a sensor in a certain position. Secondly, views of features are aggregated by feature level aggregation to form the views of position in the middle layer. Then position level module aggregates the views of positions into a view of modality, and the unified representation of multi-view features is obtained by aggregating views of modality. Finally, cross-entropy losses for each view of modality and the unified representation are computed to optimize the network. The detail of each component is explained in the following subsections.
Fig 1

An overview of the hierarchical multi-view aggregation network.

We first construct four views of feature spaces for each individual sensor in the bottom layer. Then we designed three aggregation modules to integrate views step-by-step into a multi-view unified representation.

An overview of the hierarchical multi-view aggregation network.

We first construct four views of feature spaces for each individual sensor in the bottom layer. Then we designed three aggregation modules to integrate views step-by-step into a multi-view unified representation.

4.1 Multi-view construction

The white-box features extracted for HAR task is obtained by the statistical process, and can be divided into three main categories namely as time domain features, frequency domain features and time-frequency domain features [8]. Time domain features are extracted from the time series of the raw data, while frequency domain features are obtained from the frequency representation of the data. And when a wavelet transformation is applied to the raw data, the extracted features are in the time-frequency domain. These features from different feature spaces represent the statistical property of the data’s time domain, frequency domain and time-frequency domain. Recently, [21] and [22] proposed to transform non-interpretable sensor data into the image domain and applied deep neural networks to extract features. Specifically, the segments of sensor data are transformed into a visually interpretable image and CNN was trained to discriminate the images of different activities. Therefore, we add a view of visual domain features as the black-box feature, complementing to the white-box features. In our method, we use three views of white-box features and one view of black-box feature to from the views of features. We summarize the white-box features [5, 8, 14, 15] in Table 1. In our case, we calculate resultant acceleration and angular velocity from the data of accelerometer and gyroscope respectively and treat them as channels. And we apply FFT and wavelet transformation to transform the time-domain data into the frequency domain and the time-frequency domain. For visual domain feature, we follow the method in [21]. The segment of sensor data is transformed into an activity image, and CNNs is trained to extract black-box features from the images. Specifically, raw signals are first stacked row by row into a signal image by an algorithm. In the signal image, every signal sequence has the chance to be adjacent to every other sequence, which enables CNN to extract hidden correlation between neighboring signals. Then 2D discrete Fourier transformation is applied to the signal image and its magnitude is chosen as activity image. The output dimension of CNN is 120. After multi-view construction, we obtain a multi-view representation of the raw data.
Table 1

White-box features extracted from the sensor data.

DomainFeatures
TimeInterquartile Range, Max, Min, Mean, Median
Amplitude, Mean Crossing, Signal Magnitude Area
Standard Deviation, Skewness, Kurtosis, Zero Crossing
FrequencyLargest Frequency Component, Energy
Skewness, Kurtosis, Weighted average
Sum of the First 5 FFT Coefficients
Time-frequencyStandard Deviation, Max, Min, Mean
Median Crossing Rate, Wavelet Energy

4.2 Feature level aggregation

In this subsection, a fusion module is proposed to aggregate the views of features, as shown in Fig 2. Given the xtime, xfre, xt-f and xvisual from a single sensor in a certain position. We first embed xtime, xfre, xt-f and xvisual into a shared feature space, i.e., where MLP stands for a multilayer perceptron and d is the dimensionality of the shared feature spaces.
Fig 2

Overview of feature level aggregation.

Inspired by non-local neural networks [35], we then concatenate all those features into a matrix and transform it into a fused multi-feature embedding z by non-local operation: where and W2 are learnable parameters for linear transformation. Since the non-local operation does not change the dimension of the input vector, the dimension of the multi-feature embedding z could be relatively large and result in over-fitting. Motivated by [36], we adopt L2-norm to select subsets of z for avoiding over-fitting. Firstly, a vector was computed to measure the importance of each row in z: where z means the i-th row in z. Then we select the rows in z through a threshold. The key idea is to make the vector p (the norm of the embedding at each location) “compete” against a threshold τ. We put both part into the competition by selecting those elements of z whose softmax of p values exceed the threshold, i.e.: Although τ could be chosen through the hyper-parameter selection, we follow the determination of the threshold τ ≔ 1/d in [36]. Such value for τ has an interesting interpretation. If each location of the input were equally important, we would sample the locations from a uniform probability distribution p(⋅) ≔ τ = 1/d. This is equal to a probability distribution induced by the vector p of a neural network with uniformly distributed representation, i.e. τ = softmax(puniform), and hence the trained network with the vector p has to “win” against the puniform of the random network in order to select right input features by shifting the probability mass accordingly [36]. Finally, z is the aggregated feature obtained from feature level aggregation module, and it is further fed into position level aggregation module.

4.3 Position level aggregation

In this subsection, we introduce the position level aggregation module to fuse the features from the feature level aggregation module according to the position of sensors. Nowadays, the exploration of correlation has been proven to be successful in achieving good performance in computer vision tasks such as visual question answering [37] and object detection [38]. We believe that the correlations between the sensors in different position can also help improve performance. Intuitively, in a certain type activity, the correlation between some sensors is bigger than the others. For example, when people walk, the state of the accelerometer on the left leg is similar to the state of the accelerometer on the right leg, but it is different from the state of the accelerometer on the back [39]. Therefore, we take the correlation between the sensors into the position level aggregation process. Inspired by [38], we define a correlation feature f(n) for the n-th view of position, as shown in Fig 3, which is computed by the following equation:
Fig 3

Overview of correlation feature computing.

In fact, the correlation feature for the n-th view is a weighted sum of the features from the other views. The weights α measures the correlation between the n-th view and the m-th view, which can be calculated as follows: where d is the dimension of features of each view. When the value of d is relatively big, the dot products grow large in magnitude, pushing the softmax function into regions where it has extremely small gradients [40]. To counteract this effect, we scale the dot product by . Then we add the correlation feature to the view’s original features. For the n-th view, that is: As a result, each view was enhanced by the correlation feature, which is an aggregation of the other views by the correlation between them. At last, the result of position level aggregation z is obtained by a view pooling operation, i.e. element-wise max-pooling, to all the views.

4.4 Modality level aggregation

Given the features after position level aggregation, the objective here is to conduct a modality level aggregation towards a final representation for classification. We apply a soft attention mechanism to compute an aggregated vector h over the input z: The weight β for each modality is computed by: In this way, the view containing more discriminative information contribute more to the final representation. By using the three levels of aggregation in our hierarchical multi-view framework, a unified representation h for multi-view features can be constructed, and the correlation between features and position of sensors is leveraged to help get a more effective representation of the discriminative information of human activities.

4.5 Loss function

Given the final representation h of a sample I, we can compute the probability of a sample belonging to each class: And we use cross-entropy to calculate the classification loss: To further strengthen the representation capability of our model, we add auxiliary losses to the network. For each view of modality, we calculate a cross entropy loss: where y represents the ground truth of sample I. Therefore, the total loss of our model is the weighted sum of and for each view of modality:

5 Experiments

5.1 Benchmark datasets and experiment setup

To demonstrate the adaptability of our method, we conduct a number of experiments on 12 public human activity datasets. The details about the datasets are shown in Table 2. For these datasets, we divide them into four categories:
Table 2

Public human activity datasets for evaluation.

Dataset#SubjectSample Rate#Activity#SampleSensor#PositionReference
OPPORTUNITY432Hz16191564A, G, M5[41]
PAMAP29100Hz1864173A, G, M3[42]
DSA825Hz1975998A, G, M5[43]
MHEALTH1050Hz1240522A, G, M3[3]
HHAR9100-200Hz6366038A, G3[39]
Skoda196Hz1022000A4[44]
Daphnet Gait1064Hz249942A3[45]
UCI Smartphone3050Hz610299A, G1[46]
USC-HAD14100Hz1241998A, G1[47]
SHO1050Hz720998A, G, M1[14]
WISDM v1.12920Hz691515A1[48]
WISDM v2.03620Hz6248653A1[49]

A = accelerometer, G = gyroscope, M = magnetometer

Datasets which contain multiple modalities and positions, including OPPORTUNITY, PAMAP2, DSA, MHEALTH and HHAR. Datasets which contain multiple positions, but only one single modality, including Skoda and Daphnet Gait. Datasets which contain multiple modalities, but only one single position, including UCI Smartphone, USC-HAD, SHO. Datasets which contain only one single modality and one single position, including WISDM v1.1 and WISDM v2.0. A = accelerometer, G = gyroscope, M = magnetometer For the first type of datasets, we can apply the proposed method without any changes. For the second type, since there is only one sensor modality, we can only construct a two-layer multi-view model without views of modality. Similarly, we construct a two-layer multi-view model without views of position. And for the last type, there is only one single modality and one single position. Therefore, our method is simplified to a basic one-layer multi-view model. On the OPPORTUNITY dataset, we use the same training set, validation set and test set employed in the OPPORTUNITY challenge to train and test our models, as all the other papers did. On the other dataset, we conduct a 5-fold cross validation to study the performance of our model, since there is no agreed division on the datasets.

5.2 Implementation details

Sensor data are preprocessed to fill in missing values using linear interpolation and to do a per channel normalization to interval [0, 1]. The length of sliding window is 1.2 second and the overlap is 50%. We initialize the CNN for extracting visual features after training on ImageNet, and finetune it on respective dataset, following the architecture of CNN in [21]. The aggregation networks are trained on each dataset, and finetuned along with the CNN. The output size of multilayer perceptron is 128, and dropout is applied to the output with probability of 0.2. The weight decay of the whole network is configured to 0.0001. During the training, we apply a mini-batch size of 128 samples and an Adam optimizer [50] with a learning rate of 0.001 to train the networks. These hyper-parameters are same for all HAR datasets. The whole pipeline is implemented using TensorFlow [51].

5.3 Evaluation metrics

We adopt the accuracy and the weight F1 score as our evaluation metrics for fair comparison with the state-of-the-arts. Since the classes of the datasets may be highly unbalanced, the weight F1 score can also measure the performance of the model appropriately. Specifically, the weight F1 score is defined as follows: where i is the class index and w = n/N is the proportion of samples of class i, with n being the number of samples of the i-th class and N being the total number of samples.

5.4 Overall performance

In previous methods, benchmark datasets used for evaluation are quite different. In order to make a fair comparison, we implement five state-of-the-art methods and conduct comprehensive experiments on 12 datasets. In experiments, our HMVAN is compared to [21], [11], [29], [23] and [24]. In addition, we provided the results of the simplified version of our HMVAN on the Datasets which contain multiple modalities or positions. The overall comparison results are presented in Table 3. It is illustrated that our method is capable of achieving higher performance on these datasets. First of all, on the datasets which have multiple modalities and positions, such as the OPPORTUNITY dataset, our method improves the F1 score from 0.929 to 0.933 for gesture recognition task, and from 0.900 to 0.917 for locomotion recognition task, when compared with [24]. And on the second type of datasets which only have one single modality, such as Skoda dataset, our method improves the state-of-the-art [11] from 0.958 to 0.965. Then on the third type of datasets which only have one single position, such as UCI Smartphone dataset, our method improves the accuracy from 0.9518 to 0.955 when compared with [21]. On the last type of dataset, which only have one single modality and one single position, such as WISDM dataset v1.1, our method can still outperform the other methods, and brings 0.4% improvement to the result of [23]. Moverover, it is shown that on the dataset with multiple modalities or positions, our simplified HMVAN can still get competitive results with those state-of-the-art methods. Note that our method outperforms the method proposed in [23], which is most related to ours, which demonstrated that our model is benefited from integrating black-box features with white-box features in a hierarchical multi-view structure and get better performance.
Table 3

Comparison of the proposed model against the state-of-the-art methods on various human activity benchmark datasets.

Results from each method
Jiang et al. [21]DeepConvLSTM [11]Hammerla et al. [29]Ravi et al. [23]attrCNN-IMU [24]HMVANSimplified HMVAN
DatasetsAcc.F1Acc.F1Acc.F1Acc.F1Acc.F1Acc.F1Acc.F1
OPP-Gesture0.9130.912-0.915*-0.927*0.9220.921-0.929*0.9340.9330.9300.929
OPP-Locomotion0.8890.889-0.895*0.8920.8910.8970.896-0.900*0.9180.9170.9020.902
PAMAP20.9110.9100.9270.926-0.937*0.9300.929-0.9088*0.9440.9430.9320.932
DSA0.8630.8620.8770.8720.8920.8910.8840.8830.8650.8650.9050.9040.8850.885
HHAR0.9540.9540.9770.9760.9590.9580.9400.9400.9470.9470.9780.9770.9530.952
MHEALTH0.9330.9320.9210.9200.9460.9450.9500.9490.9420.9410.9680.9670.9510.950
Skoda0.9440.943-0.958*0.9500.9480.9530.9520.9590.9580.9660.9650.9540.954
Daphnet Gait0.9010.8990.9420.941-0.760*0.958*-0.9330.9320.9660.9650.9600.959
UCI Smartphone0.9518*-0.9440.9440.9310.9300.9450.9430.9500.9500.9550.9540.9470.947
USC-HAD0.9701*-0.9570.9570.9540.9530.9610.9590.9670.9650.9750.9730.9640.963
SHO0.9993*-0.9870.9860.9890.9890.9940.9940.9970.9970.99950.99870.99580.9958
WISDM v1.10.9550.9540.9480.9470.9330.9330.986*-0.9660.965--0.9900.989
WISDM v2.00.8970.8960.9060.9050.9110.9110.927*-0.9200.919--0.9310.930

Results marked with ‘*’ are obtained from the papers.

Results marked with ‘*’ are obtained from the papers.

5.5 Ablation study

In this section, we conduct ablation experiments to study the effectiveness of individual component in our model on the validation dataset of the OPPORTUNITY dataset for locomotion recognition task.

5.5.1 On the hierarchical multi-view structure

The primary contribution to be investigated is the effectiveness of the hierarchical structure multi-view structure for human activity recognition. We compare the HMVAN with three baselines, which removes the feature aggregation layer, position aggregation layer and modality aggregation layer respectively. Moreover, we also drop the of the HMVAN as a baseline to explore the improvement by the auxiliary losses for each view of modality. As shown in Table 4, the F1-score declines 1.6%, 3.3% and 2.5% when removing the feature aggregation layer, the position aggregation layer and the modality aggregation layer from HMVAN respectively. Such decrease in performance shows the hierarchical multi-view structure indeed plays an important role in our proposed method. In addition, adding auxiliary losses to HMVAN brings an improvement of 1.2%, which demonstrates its ability to strengthen the representation capability of our HMVAN. From the results, we can infer that, on the dataset which contain multiple modalities and positions, the model is benefited more from the position aggregation layers than feature aggregation layer. The result declines 3.3% when not using position aggregation layer, while the result declines 1.6% when not using feature aggregation layer, which is smaller than the former.
Table 4

Results of our model with different structure.

MethodsAccuracy
HMVAN w/o position aggregation layer0.874
HMVAN w/o modality layer0.882
HMVAN w/o feature aggregation layer0.891
HMVAN w/o auxiliary losses0.895
HMVAN0.907

5.5.2 On different views

We illustrate the effectiveness of each view of features in HMVAN. We exclude each view of features in each layer from HMVAN one by one, and the result is reported in Table 5. From the result, we can see that each view of features brings more or less improvement to HMVAN. The view of frequency domain features contributes most to our model, which brings 2.6% improvements. The view of time domain features and the view of time-frequency domain features contribute 2.2% and 1.7% improvements to the HMVAN respectively. In addition, if we exclude the view of visual domain features, the performance of our model decreases 2.1%. Therefore, it can be inferred that the view of visual domain features is effective to improve the performance of our model.
Table 5

Results of the proposed model with different views.

MethodsAccuracy
HMVAN w/o time domain features view0.885
HMVAN w/o frequency domain features view0.881
HMVAN w/o time-frequency domain features view0.890
HMVAN w/o visual domain features view0.886
HMVAN0.907

5.5.3 On the feature level aggregation

We further compared different methods in the feature level aggregation module. The non-local operations are replaced by four different operations. And we remove the L2-norm to prove its effectiveness of preventing over-fitting. As shown in Table 6, the non-local operation achieves better performance than the other feature aggregation methods, which proves that our model can gain benefit from the feature correlation learned from the non-local operation. In addition, the L2-norm plays an important role to regularize the network, since it brings 0.7% improvements to the model only using non-local operation.
Table 6

Results of different feature level aggregation methods.

MethodsAccuracy
element-wise adding0.885
element-wise multiplying0.881
element-wise mean0.890
element-wise maximum0.886
non-local0.900
L2-norm0.863
non-local + L2-norm0.907

5.5.4 On the position level aggregation

In position level aggregation, the correlation feature plays an important role. We drop the correlation feature to prove its contribution to our model. As is illustrated in Table 7, simple view pooling cannot effectively find the correlation between different position and only keeps a part of discriminative information. By adding correlation feature, each view is augmented with the correlation between position, which leads to a better performance in accuracy.
Table 7

Results of different position level aggregation methods.

MethodsAccuracy
simple view pooling0.875
correlation feature + view pooling0.907

5.5.5 On the modality level aggregation

For modality level aggregation, we compared our method with five different approaches: MLP, element-wise adding, element-wise multiplying, element-wise mean and element-wise maximum. It can be seen in Table 8 that our attention aggregation method performs better than the other methods. It is able to fully explore the contribution of each modality to the final representation.
Table 8

Results of different modality level aggregation methods.

MethodsAccuracy
MLP0.894
element-wise adding0.899
element-wise multiplying0.881
element-wise mean0.887
element-wise maximum0.901
attention fusion0.907

5.5.6 On the attention mechanisms

In the proposed framework, we use three types of attention mechanisms in different levels of aggregation. To explore the effectiveness of each type of attention mechanism in the particular layer, we conduct experiments to interchange them in each layer. We define the attention mechanism in feature, position and modality aggregation layer as attention type #1, type #2 and type #3 respectively. The results are presented in Table 9. It is shown that when we interchange the different attention mechanisms the results declined heavily. Since the goal of each layer is different, we properly design each aggregation layer to enhance the performance of the whole model. Firstly, on the feature aggregation layer, the model aims to learn the relationship between the different views of feature spaces. Since the dimension of each feature space is relatively large, we use the non-local operation to effectively compute the interact of features. And L2-norm regularization is added to the output for preventing over-fitting. Secondly, on the position aggregation layer, our goal is to learn a correlation features which represent the dependency of each position with the others. Then the correlation features are added to the original layer to enhance the features of each view, therefore each view actually contains holistic information of the whole position. At last, on the modality aggregation layer, our method intends to combine the features of each view of modality according to their contribution to the final representations, and simple soft attention is used to calculate the weight of each view. Although it is feasible to interchange these attention mechanisms in the programming, the performance of the model will decline when we use an improper attention mechanism.
Table 9

Results of different attention mechanisms in each aggregation layer.

Feature layerPosition layerModality layerAccuracy
type #1type #1type #10.837
type #2type #2type #20.802
type #3type #3type #30.849
type #1type #2type #10.875
type #1type #3type #10.844
type #1type #1type #20.852
type #1type #1type #30.850
type #1type #2type #20.822
type #2type #2type #20.784
type #3type #2type #20.802
type #2type #1type #20.801
type #2type #3type #20.823
type #2type #2type #10.795
type #2type #2type #30.809
type #1type #3type #30.867
type #2type #3type #30.871
type #3type #3type #30.880
type #3type #1type #30.865
type #3type #2type #30.883
type #3type #3type #10.870
type #3type #3type #20.842
type #1type #2type #30.907
type #3type #2type #10.821
type #2type #1type #30.812

5.5.7 On the number of training samples

To further compare the performance of our method with the “pure” deep learning methods, we conduct experiments on HHAR dataset with different percentage of training samples to see how the performance changes. Specifically, we provide 10%, 50%, 80% and 100% of training samples to train each model and evaluate them on the test dataset. The comparison of our method with the other state-of-the-art methods are in Fig 4. It is shown that the performance our method still performs better than the other state-of-the-art methods when only using a small percentage of the training samples.
Fig 4

Comparison of the proposed model against the state-of-the-art methods with different percentage of training samples in HHAR dataset.

6 Discussion

The main findings from the direct comparison of our HMVAN against the other methods which treat sensor data as single view is that: (1) HMVAN reaches higher performance of both accuracy and F1 score; (2) it is significantly better able to disambiguate closely-related activities; (3) it is applicable even when the dataset is relatively small. Firstly, these findings support the hypothesis that white-box features and black-box features could be effectively integrated in a unified framework to take advantage of each other. Note that even when the dataset only contains one single modality and one single position, our method can still get a better performance than the other methods. The results demonstrate that the model is benefited from the construction of different views of feature spaces on multi-modal sensor data. Secondly, it has been proved that capturing the feature dependency is important to fine-grained action classification [11]. In our method, we introduce three attention mechanisms to capture the correlation between features, different sensor position and different modality, which effectively improves to the resulting recognition accuracy. At last, previous study have shown that shallow features can perform relatively well when the dataset is scale and unbalanced [17]. The results of our model with different percentage of training samples shows that our method outperforms those “pure” deep learning method when the dataset is insufficient, which demonstrate different views of shallow feature indeed help our model to gain competitive results on small dataset. The confusion matrices on the OPPORTUNITY dataset for the gesture recognition task are illustrated in Table 10 for our HMVAN and Table 11 for the method in [23]. The confusion matrices contain information about actual and predicted gesture classifications, to identify the nature of the classification errors, as well as their quantities. Each cell in the confusion matrix represents the number of times that the gesture in the row is classified as the gesture in the column. It is shown that our HMVAN performs better than the method in [23].
Table 10

Confusion matrix for OPPORTUNITY dataset using our HMVAN.

Predicted
NullOpen Door 1Open Door 2Close Door 1Close Door 2Open FridgeClose FridgeOpen Draw WasherClose Draw WasherOpen Draw 1Close Draw 1Open Draw 2Close Draw 2Open Draw 3Close Draw 3Clean TableDrink From CupToggle Switch
ActualNull13832655324155210135422397589
Open Door 1107601000000000000000
Open Door 270155020000000000000
Close Door 181507800000000000000
Close Door 2100001300000000000000
Open Fridge11100002532220000000001
Close Fridge4100001921001000000000
Open Draw Washer61000060994100000000
Close Draw Washer430000201079000100000
Open Draw 110000000313860131001
Close Draw 120000010008460000000
Open Draw 2130000000118229610000
Close Draw 25000000021542503000
Open Draw 314000000000080883000
Close Draw 36000000000029580000
Clean Table88000000000000008100
Drink From Cup1430000001100000003970
Toggle Switch570000000022000000122
Table 11

Confusion matrix for OPPORTUNITY dataset using the method in [23].

Predicted
NullOpen Door 1Open Door 2Close Door 1Close Door 2Open FridgeClose FridgeOpen Draw WasherClose Draw WasherOpen Draw 1Close Draw 1Open Draw 2Close Draw 2Open Draw 3Close Draw 3Clean TableDrink From CupToggle Switch
ActualNull137525865391814292011402021148
Open Door 1175102800000000000000
Open Door 21501110380000000000000
Close Door 1102206900000000000000
Close Door 290701240000000000000
Open Fridge13000002203441000000000
Close Fridge4900007614600000000000
Open Draw Washer1080000404514000000000
Close Draw Washer750000403026000000000
Open Draw 131000000002750020001
Close Draw 1400000000019160000000
Open Draw 236000000009118160000
Close Draw 214000000003113590000
Open Draw 3290000000000005628000
Close Draw 390000000000005142000
Clean Table98000000000000007300
Drink From Cup1940000000000000003490
Toggle Switch99000000002000000082
The auxiliary loss plays an important role in our network regularization. For strengthening the representation capability of each view, we calculate cross entropy losses to help regularize the network the learn the most discriminative features in each view of feature spaces. We also evaluate different weights on the main cross entropy loss and the auxiliary loss. The results of different weights seem almost same as the results of weight them equally. We infer that, since the type of main loss and the auxiliary loss are same and the task of them are both helping extract discriminative features, the performance of the model will not be influenced by different weights. In terms of training time, there is not such a significant difference between our models and the other deep learning models, despite the complex attention mechanism included in the aggregation layer. Training HMVAN on the OPPORTUNITY dataset requires 342.5 minutes to converge while deepConvLSTM [11] requires 340.3 minutes. And the inference time of the whole dataset is about 7 seconds while deepConvLSTM [11] takes 6.68 seconds. The GPU used to train the model is NVIDIA GTX TITAN X. Recently, high-end mobile platforms already contain GPUs that can be used for general purpose processing [52]. A mobile processor, such as Qualcomm Snapdragon 855, comprises Adreno 640 GPU running at a maximum of 585 MHz and support OpenCL profiles for general purpose GPU computing. While cores differ in capabilities, the available computational power may well be sufficient for real-time recognition in upcoming mobile devices.

7 Conclusions and future works

In this paper, we propose a Hierarchical Multi-View Aggregation Network (HMVAN) for sensor-based human activity recognition (HAR). Compared with existing deep learning-based method, our method constructs multi-view feature spaces for each individual sensor from the point of white-box features and black-box features, and aggregate them into a unified representation by a hierarchical context. In addition, we propose three aggregation modules from the point of feature level, position level and modality level respectively. For feature level aggregation, we apply a non-local operation augmented with the L2-norm to explores the correlation between different features and aggregate them. Then, in the position level aggregation, we take the correlation of different sensor position into consideration by introducing the correlation feature, which can enhance the representation of each view. Finally, in the modality level aggregation, we conduct a soft attention mechanism to measure the discrimination of each view and fuse them for classification. Ablation study demonstrates the effectiveness of hierarchical multi-view architecture and the view aggregation modules. We extensively evaluated our method on 12 benchmark datasets, and our method is capable of achieving state-of-the-art performance on each dataset. At present, the sensor-based HAR method still relies heavily on labeled training samples. In future work, we plan to explore unsupervised or semi-supervised learning method for sensor-based HAR.
  6 in total

1.  A Deep Learning Approach to on-Node Sensor Data Analytics for Mobile or Wearable Devices.

Authors:  Daniele Ravi; Charence Wong; Benny Lo; Guang-Zhong Yang
Journal:  IEEE J Biomed Health Inform       Date:  2016-12-23       Impact factor: 5.772

2.  Wearable assistant for Parkinson's disease patients with the freezing of gait symptom.

Authors:  Marc Bächlin; Meir Plotnik; Daniel Roggen; Inbal Maidan; Jeffrey M Hausdorff; Nir Giladi; Gerhard Tröster
Journal:  IEEE Trans Inf Technol Biomed       Date:  2009-11-10

3.  Feature selection for wearable smartphone-based human activity recognition with able bodied, elderly, and stroke patients.

Authors:  Nicole A Capela; Edward D Lemaire; Natalie Baddour
Journal:  PLoS One       Date:  2015-04-17       Impact factor: 3.240

4.  Step detection and activity recognition accuracy of seven physical activity monitors.

Authors:  Fabio A Storm; Ben W Heller; Claudia Mazzà
Journal:  PLoS One       Date:  2015-03-19       Impact factor: 3.240

5.  Fusion of smartphone motion sensors for physical activity recognition.

Authors:  Muhammad Shoaib; Stephan Bosch; Ozlem Durmaz Incel; Hans Scholten; Paul J M Havinga
Journal:  Sensors (Basel)       Date:  2014-06-10       Impact factor: 3.576

6.  Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition.

Authors:  Francisco Javier Ordóñez; Daniel Roggen
Journal:  Sensors (Basel)       Date:  2016-01-18       Impact factor: 3.576

  6 in total
  2 in total

1.  An improved multi-view attention network inspired by coupled P system for node classification.

Authors:  Qian Liu; Xiyu Liu
Journal:  PLoS One       Date:  2022-04-28       Impact factor: 3.752

2.  PathCNN: interpretable convolutional neural networks for survival prediction and pathway analysis applied to glioblastoma.

Authors:  Jung Hun Oh; Wookjin Choi; Euiseong Ko; Mingon Kang; Allen Tannenbaum; Joseph O Deasy
Journal:  Bioinformatics       Date:  2021-07-12       Impact factor: 6.937

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.