| Literature DB >> 33809743 |
Xue-Bo Jin1,2, Ruben Jonhson Robert Jeremiah3, Ting-Li Su1,2, Yu-Ting Bai1,2, Jian-Lei Kong1,2.
Abstract
State estimation is widely used in various automated systems, including IoT systems, unmanned systems, robots, etc. In traditional state estimation, measurement data are instantaneous and processed in real time. With modern systems' development, sensors can obtain more and more signals and store them. Therefore, how to use these measurement big data to improve the performance of state estimation has become a hot research issue in this field. This paper reviews the development of state estimation and future development trends. First, we review the model-based state estimation methods, including the Kalman filter, such as the extended Kalman filter (EKF), unscented Kalman filter (UKF), cubature Kalman filter (CKF), etc. Particle filters and Gaussian mixture filters that can handle mixed Gaussian noise are discussed, too. These methods have high requirements for models, while it is not easy to obtain accurate system models in practice. The emergence of robust filters, the interacting multiple model (IMM), and adaptive filters are also mentioned here. Secondly, the current research status of data-driven state estimation methods is introduced based on network learning. Finally, the main research results for hybrid filters obtained in recent years are summarized and discussed, which combine model-based methods and data-driven methods. This paper is based on state estimation research results and provides a more detailed overview of model-driven, data-driven, and hybrid-driven approaches. The main algorithm of each method is provided so that beginners can have a clearer understanding. Additionally, it discusses the future development trends for researchers in state estimation.Entities:
Keywords: Kalman filter; data-driven; deep learning; hybrid-driven; model-driven; state estimation
Year: 2021 PMID: 33809743 PMCID: PMC8002332 DOI: 10.3390/s21062085
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1A flowchart of the Gaussian mixture filter. (a) The system noise is a complicated one with multiple Gaussian components. (b) The complex Gaussian noise of the system is decomposed into several standard Gaussian noises , . (c) The state is estimated based on each Gaussian noise. (d) The state estimation result based on the complex mixed high noise is obtained by using the estimated result and the weight.
Figure 2The flow of the particle filter algorithm. The particles are generated firstly, and then, the weights need to be calculated according to the states’ accuracy. Next, the weights are normalized, the state is estimated, and the particles are resampled. The loop is continued until all the measurement data are used.
Comparison of each filter.
| Filter | Requirements for the System | Accuracy for a Practical System | Calculation Cost | Description |
|---|---|---|---|---|
| Kalman filter | Linear, with Gaussian white noise | Low | Low | The requirements for the system are very high, so it is difficult to achieve high accuracy in the actual application system. |
| EKF | Nonlinear, with Gaussian noise | Medium | Low | The performance of UKF and CKF is better |
| UKF | Nonlinear, with Gaussian noise | Medium | Medium | |
| CKF | Nonlinear, with Gaussian noise | Medium | Medium | |
| Gaussian mixture filters | Nonlinear, with non-Gaussian noise | Medium | Medium | These filters have low requirements for |
| Particle filters | Nonlinear, with non-Gaussian noise | High | High |
Figure 3The optimization loop of the system model and state.
Data-driven modeling methods with deep learning networks.
| References | Network Cell | Hyperparameter Optimization | Type of Network | Purpose |
|---|---|---|---|---|
| [ | Long short-term memory (LSTM) | Not mentioned | Classic deep learning network | Classify sequence |
| [ | Gated recurrent unit (GRU) | Not mentioned | Classic deep learning network | Forecasting time-series data |
| [ | Recurrent neural network (RNN) | Not mentioned | Classic deep learning network | Machine translation |
| [ | Attention-based LSTM | Not mentioned | Classic deep learning network | Machine translation |
| [ | Convolution network | Bayesian optimization | Classic deep learning network | Prediction |
| [ | GRU | Bayesian optimization | Classic deep learning network | Prediction |
| [ | Bidirectional RNN | Not mentioned | Classic deep learning network | Detection |
| [ | ConvLSTM | Not mentioned | Classic deep learning network | Prediction |
| [ | RNN | Not mentioned | Classic deep learning network | State estimation |
| [ | GRU | Manual search | Classic deep learning network | Prediction |
| [ | LSTM | Manual search | Bayesian deep learning network | Prediction |
| [ | GRU | Bayesian optimization | Classic deep learning network | Prediction |
| [ | Multi-layer forward neural network | Not mentioned | Bayesian deep learning network | State estimation |
Figure 4Data-driven model. (a) GRU cell; (b) LSTM cell; (c) deep learning network.
Figure 5Hyperparameter optimization process. (a) The training process; (b) The weights’ optimization process.
Figure 6The differences between the Bayesian deep network and non-Bayesian deep network are (a) the weight optimization process for the non-Bayesian deep network, where the weight is a certain value, and (b) the weight optimization process for the Bayesian deep network, where the weight’s distribution is obtained.