| Literature DB >> 34945941 |
Leyuan Liu1, Jian He1,2, Keyan Ren1,2, Jonathan Lungu1, Yibin Hou1,2, Ruihai Dong3.
Abstract
Wearable sensor-based HAR (human activity recognition) is a popular human activity perception method. However, due to the lack of a unified human activity model, the number and positions of sensors in the existing wearable HAR systems are not the same, which affects the promotion and application. In this paper, an information gain-based human activity model is established, and an attention-based recurrent neural network (namely Attention-RNN) for human activity recognition is designed. Besides, the attention-RNN, which combines bidirectional long short-term memory (BiLSTM) with attention mechanism, was tested on the UCI opportunity challenge dataset. Experiments prove that the proposed human activity model provides guidance for the deployment location of sensors and provides a basis for the selection of the number of sensors, which can reduce the number of sensors used to achieve the same classification effect. In addition, experiments show that the proposed Attention-RNN achieves F1 scores of 0.898 and 0.911 in the ML (Modes of Locomotion) task and GR (Gesture Recognition) task, respectively.Entities:
Keywords: Attention-RNN; attention mechanism; human activity recognition; information gain
Year: 2021 PMID: 34945941 PMCID: PMC8700115 DOI: 10.3390/e23121635
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Information gain-based human activity model. (a) Human skeleton. (b) Positions of sensors can be fixed. (c) Cartesian coordinate system.
Figure 2Network architecture of the Attention-RNN.
Figure 3Sensors placement of the dataset.
Composition of the dataset intercepted by the sliding window.
| Task | Activity Name | # of Training Instances | # of Testing Instances |
|---|---|---|---|
| GR | Open_Door1 | 864 | 58 |
| Open_Door2 | 887 | 95 | |
| Close_Door1 | 806 | 60 | |
| Close_Door2 | 846 | 83 | |
| Open_Fridge | 921 | 228 | |
| Close_Fridge | 850 | 160 | |
| Open_Dishwasher | 666 | 100 | |
| Close_Dishwasher | 628 | 77 | |
| Open_Drawer1 | 490 | 39 | |
| Close_Drawer1 | 413 | 42 | |
| Open_Drawer2 | 457 | 40 | |
| Close_Drawer2 | 416 | 26 | |
| Open_Drawer3 | 566 | 67 | |
| Close_Drawer3 | 564 | 61 | |
| Clean_Table | 904 | 99 | |
| Drink_Cup | 3246 | 317 | |
| Toggle_Switch | 623 | 105 | |
| Null | 32348 | 8237 | |
| ML | Stand | 19321 | 3101 |
| Walk | 10875 | 2272 | |
| Sit | 7410 | 2016 | |
| Lie | 1209 | 463 | |
| Null | 7680 | 2042 |
F1 comparison of different classification algorithms.
| Method | F1 (ML Task) | F1 (GR Task) | Testing Time (S) |
|---|---|---|---|
| Random Forest [ | 0.870 | 0.900 | 29.62 |
| CNN [ | - 1 | 0.851 | 2.29 |
| DeepConvLSTM [ | 0.895 | 0.915 | 9.82 |
| Attention-RNN (ours) | 0.898 | 0.911 | 3.75 |
1 “-” means there is no relevant data in the original paper.
Figure 4Confusion matrix of ML task.
Figure 5Confusion matrix of GR task.
Experiments on different model structures.
| Model | Structure | F1 (ML Task) | F1 (GR Task) |
|---|---|---|---|
| A | BN + 2BiLSTM + Dense | 0.894 | 0.903 |
| B | 2BiLSTM + Attention + Dense | 0.891 | 0.886 |
| C | BN + 1BiLSTM + Attention + Dense | 0.891 | 0.899 |
| D | 2BiLSTM + BN + Attention + Dense | 0.893 | 0.903 |
| E | 2BiLSTM + Attention + BN + Dense | 0.894 | 0.903 |
| F | BN + 3BiLSTM + Attention + Dense | 0.891 | 0.904 |
| G | BN + 4BiLSTM + Attention + Dense | 0.891 | 0.901 |
| H | BN + 5BiLSTM + Attention + Dense | 0.891 | 0.906 |
| I | BN + Attention + 2BiLSTM + Dense | 0.878 | 0.898 |
| J | BN + BiLSTM + Attention + BiLSTM + Dense | 0.892 | 0.891 |
| K | BN + BiLSTM + Attention + BiLSTM + Attention + Dense | 0.890 | 0.901 |
| L | BN + Attention + BiLSTM + Attention + BiLSTM + Dense | 0.881 | 0.899 |
| M | BN + Attention + BiLSTM + Attention + BiLSTM + Attention + Dense | 0.857 | 0.898 |
| Attention-RNN | BN + 2BiLSTM + Attention + Dense | 0.898 | 0.911 |
Information gain and ranking of each sensor.
| Sensor Name | Channels |
|
|
|---|---|---|---|
| RKN^ | 1–3 | 1.797 (8) | 0.558 (15) |
| HIP | 4–6 | 0.840 (18) | 0.471 (19) |
| LUA^ | 7–9 | 1.092 (13) | 0.615 (12) |
| RUA_ | 10–12 | 0.927 (16) | 0.600 (14) |
| LH | 13–15 | 1.617 (9) | 0.972 (9) |
| BACK (Acc) | 16–18 | 0.861 (17) | 0.618 (11) |
| RKN_ | 19–21 | 1.332 (10) | 0.603 (13) |
| RWR | 22–24 | 1.308 (11) | 1.464 (8) |
| RUA^ | 25–27 | 0.822 (19) | 0.474 (18) |
| LUA_ | 28–30 | 1.119 (12) | 0.510 (16) |
| LWR | 31–33 | 1.011 (14) | 0.492 (17) |
| RH | 34–36 | 0.963 (15) | 0.741 (10) |
| BACK (IMU) | 37–45 | 2.817 (3) | 2.088 (3) |
| RUA | 46–54 | 2.610 (6) | 1.890 (6) |
| RLA | 55–63 | 2.241 (7) | 1.971 (4) |
| LUA | 64–72 | 2.664 (5) | 1.818 (7) |
| LLA | 73–81 | 2.772 (4) | 1.899 (5) |
| L-SHOE | 82–97 | 4.832 (1) | 2.400 (2) |
| R-SHOE | 98–113 | 4.784 (2) | 2.448 (1) |
Figure 6F1 scores for ML task with different numbers of sensors.
Figure 7F1 scores for GR task with different numbers of sensors.
Figure 8Top 6 information gain sensors in GR task and top 12 information gain sensors in ML task.