| Literature DB >> 35669518 |
Ahmad Alsharef1, Karan Aggarwal2, Manoj Kumar3, Ashutosh Mishra4.
Abstract
Time-series forecasting is a significant discipline of data modeling where past observations of the same variable are analyzed to predict the future values of the time series. Its prominence lies in different use cases where it is required, including economic, weather, stock price, business development, and other use cases. In this work, a review was conducted on the methods of analyzing time series starting from the traditional linear modeling techniques until the automated machine learning (AutoML) frameworks, including deep learning models. The objective of this review article is to support identifying the time-series forecasting challenge and the different techniques to meet the challenge. This work can be additionally an assist and a reference for researchers and industries demanding to use AutoML to solve the problem of forecasting. It identifies the gaps of the previous works and techniques used to solve the problem of forecasting time series.Entities:
Year: 2022 PMID: 35669518 PMCID: PMC9159649 DOI: 10.1007/s11831-022-09765-0
Source DB: PubMed Journal: Arch Comput Methods Eng ISSN: 1134-3060 Impact factor: 8.171
Fig. 1Linear regression example
Fig. 2Recurrent neural network architecture
Fig. 3GRU network architecture
Fig. 4LSTM network architecture
Fig. 5IndRNN architecture
AutoML frameworks comparison
| AutoML framework | Scope | Working method |
|---|---|---|
| H2O [ | Random forests, extremely randomized trees, generalized linear models, XGBoost, H2O gradient boosting machine, and deep neural networks | Trains a random grid of algorithms using a hyperparameter space. Individual models have been tuned using cross-validation. Two stacked ensembles are trained. One contains all the models optimized for model performance, and the other contains only the best-performing model from each algorithm. Then, it returns a sorted leaderboard of all models [ |
| Auto-Sklearn [ | Covers 15 models, 14 feature preprocessing methods, and 4 data preprocessing methods. The models KNN) gradient boosting, SGD, random forest, AdaBoost, and others | Starts by extracting dataset meta-feature to find the similarity to the knowledge base relying on meta-learning. Then, Bayesian optimization will try to find and select the out-performing ML pipelines. Finally, it builds the ensemble model based on the best ML workflow in the search space [ |
| AutoGluon [ | Covers many models including neural networks, LightGBM boosted trees, CatBoost, boosted trees, random forests, extremely randomized trees, and KNN | Trains a set of generated models under different configurations, optimizes hyperparameters, and selects the best of them. The search strategy for the best set of parameters is based on a random search, grid search, or Bayesian optimization [ |
| TPOT [ | Covers the following models: decision tree, random forest, eXtreme gradient boosting (XGBoost), logistic regression, and KNN | ML model design and optimization start by generating 100 tree-base pipelines evaluating them and selecting the top 20 on each generation. Each of these 20 pipelines is copied with crossovers or other mutations to produce 5 copies. This process is repeated 100 times until it finally outputs the top-performing models of the last generation [ |
| Auto-Weka [ | Bayes net, naive Bayes, linear regression, logistic regression, single-layer perceptron, stochastic gradient descent, SVM, KNN, K-star, decision tree, random forest, and others [ | Considers selecting a learning model and setting its hyperparameters simultaneously, going beyond other models that process the two tasks in isolation. Auto-WEKA does this fully automated approach using the Bayesian optimization approach |
| TSPO [ | XGBoost, random forest, quantile random forest, CatBoost, feed-forward neural networks | Utilizes a genetic algorithm to find an appropriate set of time series features, machine learning models, hyperparameters. It takes a raw time series, automatically decomposes it, extracts time series features for each decomposition and finds a model with a fitting hyper-parameter |
| AutoKeras [ | Focuses on deep learning rather than simple modelling | Utilizes Keras API and uses a process of searching through neural network architectures to best address a modeling task, referred to as Neural Architecture Search (NAS) |
| EvalML [ | Supports different supervised ML problem types: regression, binary classification, multiclass classification, time series regression, time series binary classification, and time series multiclass classification | Builds and optimizes ML pipelines using specific objective functions. It can automatically perform feature selection, model building, hyper-parameter tuning, cross-validation, etc |
| TransmogrifAI [ | Covers the following machine learning models: Decision Trees, Gradient Boosted Trees (GBT), Linear Regression, Linear Support Vector Machines, Logistic Regression, Naïve Bayes, and Random Forest (RF) [ | The user has to specify the dataset, the schema, and the target column and the framework automatically discards input features that do not present a predictive value and train a predefined set of algorithms with a predefined set of hyperparameters depending on the type of problem |
Related works comparison
| Author(s) and year | Methodologies | Findings | Identified research gaps |
|---|---|---|---|
| Li et al. (2010) [ | Surveyed the usage of AI and ANNs for time-series stock market prediction | Artificial neural networks are an effective forecast tool in financial economics | Surveyed only basic state-of-art models |
| Peng et al. (2015) [ | Text mining techniques with word embeddings and neural networks to extract information from financial News and detect stocks movements | The proposed methods improved the accuracy of prediction over the baseline system that uses time-series historical price information solely | Usage of a simple state-of-art method |
| McNally et al. (2018) [ | Compared deep learning with Linear models in predicting time-series data like cryptocurrency and stock market prices | Deep learning-based models gave better accuracy compared to the linear models | Usage of state-of-art prediction models |
| Ji et al. (2019) [ | Evaluated various state-of-art deep learning models to predict time-series Bitcoin prices | LSTM-based prediction models slightly outperformed the other models in regressing while deep neural networks (DNN) performed better in classifying price ups and downs | Didn't consider other recent deep learning models, such as transformer networks and few-shot/one-shot learning |
| Aniruddha Dutta et al. (2020) [ | Different machine learning models to predict time-series Bitcoin prices including LSTM and GRU | GRU yielded better accuracy than LSTM | Didn’t explore the potential of Convolutional Neural Network (CNN) and the work used a relatively small dataset |
| Pintelas et al. (2020) [ | Deep learning in predicting time-series cryptocurrency prices | Deep learning models were not able to solve this problem effectively | Didn't investigate the possibility of some minor information loss due to non-stationarity in the time series |
| Iqbal et al. (2021) [ | Linear models including ARIMAX, FBProphet, and XG Boosting in forecasting Bitcoin prices | ARIMAX was found as the best model for forecasting Bitcoin prices | Didn't Hypertune the parameters of the algorithms in order to improve accuracy performance |
| Hamayel. et al. (2021) [ | Three types of RNN models including GRU, LSTM, and Bi-LSTM to predict the prices of cryptocurrencies | Bi-LSTM performed the lowest compared to the other two models whereas the GRU model performed the best and yielded the lowest MAPE and RMSE | Didn't study the effect that social media in general and tweets in particular, can have on the price and trading volume of cryptocurrencies |
| Awoke et al. (2021) [ | LSTM ad GRU deep learning models to predict Bitcoin with high accuracy in handling its volatility | GRU-based models are more efficient in forecasting a highly volatile time series. However, LSTM was better at a sliding window of size 12 or smaller than 7 | Only compared basic deep learning-based models |
| Dahl (2020) [ | AutoML system and called it “TSPO” | It outperformed different statistical and ML benchmarks in 9 out of 12 randomly selected time-series datasets in different domains | The current implementation of the proposed system is computationally expensive |
| Xu et al. (2021) [ | Organized a challenge to predict time-series data that utilized 10 datasets from different domains | The first winner used the hyperparameters optimization techniques of AutoML demonstrating the feasibility and efficiency of automating time series regression | Some datasets have been too easy or too difficult to model |
| Paldino et al. (2021) [ | Compared four AutoML frameworks (AutoGluon, H2O, TPOT, Auto-sklearn) against a benchmark of traditional forecasting techniques (naïve, exponential smoothing, Holt-Winter’s) on different time-series forecasting challenges | AutoML techniques are not yet mature to address forecasting problems and this should encourage researchers to a more rigorous validation and work in the field | AutoML frameworks offer deep customization to improve their performance, which hadn't been considered |
| Javeri et al. (2021) [ | A data augmentation method to improve the performance of neural networks and unlock their power on intermediate length time-series data | Combining data augmentation with AutoML techniques such as Neural Architecture Search can help to find the best neural architecture for a given time-series dataset along with significant enhancement in the accuracy | Incorporated only a few deep learning architectures |