| Literature DB >> 36236581 |
Lívia Almada Cruz1, Ticiana Linhares Coelho da Silva1, Régis Pires Magalhães1, Wilken Charles Dantas Melo1, Matheus Cordeiro1, José Antonio Fernandes de Macedo1, Karine Zeitouni2.
Abstract
Representation learning seeks to extract useful and low-dimensional attributes from complex and high-dimensional data. Natural language processing (NLP) was used to investigate the representation learning models to extract words' feature vectors using their sequential order in the text via word embeddings and language models that maintain their semantic meaning. Inspired by NLP, in this paper, we tackle the representation learning problem for trajectories, using NLP methods to encode external sensors positioned in the road network and generate the features' space to predict the next vehicle movement. We evaluate the vector representations of on-road sensors and trajectories using extrinsic and intrinsic strategies. Our results have shown the potential of natural language models to describe the space of features on trajectory applications as the next location prediction.Entities:
Keywords: location prediction; representation learning; sensors trajectory; trajectory embedding; trajectory modeling; trajectory prediction
Mesh:
Year: 2022 PMID: 36236581 PMCID: PMC9573231 DOI: 10.3390/s22197475
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1Map of Fortaleza city with external sensors (surveillance cameras) positioned on the road network in 2019.
Figure 2Histograms of road distances to nearest sensors on the road network. (a) Distances to the 1st nearest sensor. (b) Distances to the 3rd nearest sensor.
Figure 3Statistics of sensor observations satisfy Zipf’s Law.
Figure 4Histogram of the number of observations by each sensor.
Figure 5LSTM EST prediction architecture.
Figure 6Sliding window approach.
Word2Vec and BERT MLM Parameters.
| Word2Vec | BERT MLM |
|---|---|
| embedding size = [16, 32, 64, 128] | embedding size = [16, 32, 64, 128] |
| n-grams = [4, 6, 8, 10] | sequence size = [32, 64] |
| intermediate sizes = [64] | |
| hidden dims = [64, 128, 256] | |
| num hidden layers, num attention heads = [8] |
Accuracy of prediction models.
| ACC@1 | ACC@2 | ACC@3 | |
|---|---|---|---|
| LSTM | 64.10 | 74.96 | 80.32 |
| LBERT | 66.84 | 79.70 | 85.05 |
| LBERT-FT | 73.71 | 85.35 | 90.01 |
| LW2V | 52.14 | 63.09 | 72.38 |
| LW2V-FT | 53.33 | 66.90 | 72.14 |
Mean and percentiles of the closeness error for predictions (in km).
| Mean | 60 | 70 | 80 | 90 | |
|---|---|---|---|---|---|
| LSTM | 1.0 | 0.0 | 0.85 | 1.99 | 3.49 |
| LBERT | 0.97 | 0.0 | 0.56 | 1.67 | 2.78 |
| LBERT-FT | 0.79 | 0.0 | 0.0 | 1.36 | 2.78 |
| LW2V | 1.44 | 0.73 | 1.4 | 2.44 | 4.09 |
| LW2V-FT | 1.36 | 0.65 | 1.31 | 2.46 | 4.09 |
Figure 7Example of nearest sensors according cosine distance between BERT MLM embedding vectors.
Figure 8Mean reciprocal rank of embeddings among sensors inside a neighborhood in space.
Figure 9Example of most similar trajectories according to cosine distance between their embedding vectors. (a) High spatial similarity. (b) Medium spatial similarity. (c) Low spatial similarity.
Figure 10Mean reciprocal rank of the embeddings among trajectories pairs under a maximal distance.
Results for the best LSTM model with different window sizes.
| Window Size | Embedding Size |
|
|
|
|---|---|---|---|---|
| 5 | 128 | 0.641 | 0.750 | 0.803 |
| 7 | 128 | 0.611 | 0.714 | 0.760 |
| 15 | 128 | 0.554 | 0.669 | 0.718 |
| 31 | 128 | 0.357 | 0.433 | 0.482 |
Results for the best LW2V model with different window sizes.
| Window Size | Embedding Size |
|
|
|
|---|---|---|---|---|
| 5 | 128 | 0.521 | 0.631 | 0.724 |
| 7 | 128 | 0.517 | 0.638 | 0.683 |
| 15 | 128 | 0.507 | 0.614 | 0.662 |
| 31 | 128 | 0.452 | 0.540 | 0.614 |
Results for the best LW2V-FT model with different window sizes.
| Window Size | Embedding Size |
|
|
|
|---|---|---|---|---|
| 5 | 128 | 0.533 | 0.669 | 0.721 |
| 7 | 64 | 0.526 | 0.650 | 0.695 |
| 15 | 32 | 0.514 | 0.624 | 0.664 |
| 31 | 128 | 0.460 | 0.550 | 0.614 |
Results for the best LBERT model with different window sizes.
| Window Size | Embedding Size |
|
|
|
|---|---|---|---|---|
| 5 | 128 | 0.668 | 0.797 | 0.851 |
| 7 | 128 | 0.587 | 0.748 | 0.814 |
| 15 | 128 | 0.405 | 0.579 | 0.654 |
| 31 | 128 | 0.480 | 0.659 | 0.783 |
Results for the best LBERT-FT model with different window sizes.
| Window Size | Embedding Size |
|
|
|
|---|---|---|---|---|
| 5 | 128 | 0.737 | 0.854 | 0.900 |
| 7 | 128 | 0.735 | 0.854 | 0.899 |
| 15 | 128 | 0.724 | 0.839 | 0.881 |
| 31 | 128 | 0.709 | 0.832 | 0.879 |