| Literature DB >> 35161448 |
Aitor Almeida1, Unai Bermejo1, Aritz Bilbao1, Gorka Azkune2, Unai Aguilera1, Mikel Emaldi1, Fadi Dornaika2,3, Ignacio Arganda-Carreras3,4.
Abstract
Behavior modeling has multiple applications in the intelligent environment domain. It has been used in different tasks, such as the stratification of different pathologies, prediction of the user actions and activities, or modeling the energy usage. Specifically, behavior prediction can be used to forecast the future evolution of the users and to identify those behaviors that deviate from the expected conduct. In this paper, we propose the use of embeddings to represent the user actions, and study and compare several behavior prediction approaches. We test multiple model (LSTM, CNNs, GCNs, and transformers) architectures to ascertain the best approach to using embeddings for behavior modeling and also evaluate multiple embedding retrofitting approaches. To do so, we use the Kasteren dataset for intelligent environments, which is one of the most widely used datasets in the areas of activity recognition and behavior modeling.Entities:
Keywords: attention; behavior modeling; convolutional neural networks; embeddings; graph neural networks; intelligent environments; knowledge graphs; recurrent neural networks; transformers; user behavior prediction
Mesh:
Year: 2022 PMID: 35161448 PMCID: PMC8838738 DOI: 10.3390/s22030701
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Transformer architecture for behavior modeling.
Figure 2Evaluation steps.
Proposed architectures. ID is the architecture identifier. Description is a brief explanation of the underlying model. Attention describes the type of attention mechanism used in the model. Dense # is the number of fully connected layers with a ReLU activation (all the architectures have a final fully connected layer with a softmax activation). Dense size is the size of the fully connected layers with an ReLU activation. Dropout is the value of the dropout regularizations.
| ID | Description | Attention | Dense # | Dense Size | Dropout |
|---|---|---|---|---|---|
| A3 [ | Standard LSTM size 512 | None | 2 | 1024 | 0.8 |
| A10 | Standard LSTM size 512 | Embedding-level | 2 | 1024 | 0.8 |
| M1 [ | Multi-scale CNN N-grams: 2, 3, 4, 5 | None | 2 | 512 | 0.8 |
| M5 [ | Multi-scale CNN N-grams: 2, 3, 4, 5 | Embedding-level | 2 | 512 | 0.8 |
| G1 | GCN size 50 step num 1 and A3 | None | 2 | 1024 | 0.8 |
| G2 | GCN size 50 step num 1 and M1 | None | 2 | 512 | 0.8 |
| F1 | Transformer with flatten | Multi-head | 1 | 800 | 0.8 |
| F2 | Transformer with GAP | Multi-head | 1 | 800 | 0.8 |
Architecture experiments. Accuracy at k. Mean accuracy and standard deviation for 100 executions. Maximum accuracy at 1 is also reported. ID is the architecture identifier.
| ID | Mean | Mean | Mean | Mean | Mean | Max |
|---|---|---|---|---|---|---|
| A3 | 0.4388 ± 0.0077 | 0.6064 ± 0.0189 | 0.4573 | |||
| A10 | 0.4253 ± 0.0060 | 0.6038 ± 0.0098 | 0.6863 ± 0.0127 | 0.7802 ± 0.0123 | 0.8318 ± 0.0096 | 0.4487 |
| M1 | 0.4197 ± 0.0090 | 0.5938 ± 0.0121 | 0.6941 ± 0.0109 | 0.7583 ± 0.0089 | 0.8052 ± 0.0090 | 0.4530 |
| M5 | 0.4387 ± 0.0125 | 0.7074 ± 0.0163 | 0.7925 ± 0.0119 | 0.8321 ± 0.0107 | 0.4701 | |
| G1 | 0.2150 ± 0.0143 | 0.3164 ± 0.0199 | 0.4150 ± 0.0424 | 0.5397 ± 0.0440 | 0.6274 ± 0.0348 | 0.2479 |
| G2 | 0.2092 ± 0.0093 | 0.3214 ± 0.0066 | 0.4331 ± 0.0420 | 0.5627 ± 0.0303 | 0.6469 ± 0.0298 | 0.2479 |
| F1 | 0.6237 ± 0.0182 | 0.7190 ± 0.0153 | 0.7748 ± 0.0148 | 0.8294 ± 0.0126 |
| |
| F2 | 0.4575 ± 0.0091 | 0.6233 ± 0.0199 | 0.7092 ± 0.0112 | 0.7883 ± 0.0120 | 0.8405 ± 0.0139 | 0.4829 |
Retrofitting experiments with different knowledge graphs. Accuracy at k. Mean accuracy for 100 executions. Maximum accuracy at 1 is also reported. ID is the architecture identifier. Source is the information from which the knowledge graph was generated. EAM: expert activity model, Dataset: information from Kasteren et al.’s manuscript [40] and the labeled data. Graph indicates the entity that was used to generate edges. ACT: activities, LOC: locations, ACT-LOC: activities and locations.
| ID | Source | Graph | Mean | Mean | Mean | Mean | Mean | Max |
|---|---|---|---|---|---|---|---|---|
| A3 | None | None | 0.4388 | 0.6064 |
|
| 0.8513 | 0.4573 |
| EAMs | ACT | 0.4345 | 0.6080 | 0.7096 | 0.7872 | 0.8492 | 0.4573 | |
| LOC | 0.4342 | 0.6094 | 0.7139 | 0.7877 | 0.8495 | 0.4530 | ||
| ACT-LOC | 0.4361 | 0.6108 | 0.7154 | 0.7895 | 0.8495 | 0.4573 | ||
| Dataset | ACT | 0.4329 | 0.6178 | 0.7189 | 0.7902 |
| 0.4573 | |
| LOC | 0.4368 | 0.6112 | 0.7142 | 0.7880 | 0.8491 | 0.4615 | ||
| ACT-LOC |
|
| 0.7199 | 0.7888 | 0.8476 |
| ||
| A10 | None | None | 0.4253 | 0.6038 | 0.6863 | 0.7802 | 0.8318 | 0.4487 |
| EAMs | ACT | 0.4329 | 0.6195 | 0.6988 | 0.7938 | 0.8330 | 0.4573 | |
| LOC | 0.4325 | 0.6198 | 0.6993 |
|
| 0.4658 | ||
| ACT-LOC | 0.4317 | 0.6190 | 0.6978 | 0.7944 | 0.8335 | 0.4487 | ||
| Dataset | ACT | 0.4353 |
| 0.6995 | 0.7926 | 0.8304 |
| |
| LOC | 0.4338 | 0.6215 | 0.7002 | 0.7953 | 0.8329 | 0.4573 | ||
| ACT-LOC |
| 0.6195 |
| 0.7922 | 0.8259 | 0.4615 | ||
| M1 | None | None | 0.4197 |
|
| 0.7583 | 0.8052 |
|
| EAMs | ACT | 0.4163 | 0.5899 | 0.6790 | 0.7559 | 0.8133 | 0.4359 | |
| LOC | 0.4182 | 0.5895 | 0.6816 | 0.7570 | 0.8134 | 0.4402 | ||
| ACT-LOC | 0.4174 | 0.5892 | 0.6803 | 0.7553 | 0.8148 | 0.4359 | ||
| Dataset | ACT | 0.4148 | 0.5884 | 0.6839 | 0.7578 | 0.8140 | 0.4402 | |
| LOC |
| 0.5892 | 0.6822 | 0.7587 | 0.8150 | 0.4487 | ||
| ACT-LOC | 0.4186 | 0.5892 | 0.6862 |
|
| 0.4402 | ||
| M5 | None | None |
|
|
|
|
| 0.4701 |
| EAMs | ACT | 0.4059 | 0.5799 | 0.6717 | 0.7493 | 0.8053 | 0.4615 | |
| LOC | 0.4123 | 0.5863 | 0.6757 | 0.7534 | 0.8065 | 0.4701 | ||
| ACT-LOC | 0.4090 | 0.5815 | 0.6747 | 0.7550 | 0.8083 | 0.4615 | ||
| Dataset | ACT | 0.4022 | 0.5756 | 0.6696 | 0.7489 | 0.8051 | 0.4615 | |
| LOC | 0.4121 | 0.5808 | 0.6768 | 0.7552 | 0.8060 | 0.4658 | ||
| ACT-LOC | 0.4125 | 0.5876 | 0.6813 | 0.7589 | 0.8076 |
|