| Literature DB >> 35017781 |
Sule Birim1, Ipek Kazancoglu2, Sachin Kumar Mangla3, Aysun Kahraman4, Yigit Kazancoglu5.
Abstract
In recent years, machine learning models based on big data have been introduced into marketing in order to transform customer data into meaningful insights and to make strategic decisions by making more accurate predictions. Although there is a large amount of literature on demand forecasting, there is a lack of research about how marketing strategies such as advertising and other promotional activities affect demand. Therefore, an accurate demand-forecasting model can make significant academic and practical contributions for business sustainability. The purpose of this article is to evaluate machine learning methods to provide accuracy in forecasting demand based on advertising expenses. The study focuses on a prediction mechanism based on several Machine Learning techniques-Support Vector Regression (SVR), Random Forest Regression (RFR) and Decision Tree Regressor (DTR) and deep learning techniques-Artificial Neural Network (ANN), Long Short Term Memory (LSTM),-to deal with demand forecasting based on advertising expenses. Deep learning is a powerful technique that can solve marketing problems based on both classification and regression algorithms. Accordingly, a television manufacturer's real market dataset consisting of advertising expenditures, sales and demand forecasting via chosen machine learning methods was analyzed and compared in terms of the accuracy of demand forecasting. As a result, Long Short Term Memory has been found to be superior to other models in providing highly accurate prediction results for demand forecasting based on advertising expenses.Entities:
Keywords: Advertisement; Demand forecasting; Machine learning; Marketing intelligence
Year: 2022 PMID: 35017781 PMCID: PMC8736292 DOI: 10.1007/s10479-021-04429-x
Source DB: PubMed Journal: Ann Oper Res ISSN: 0254-5330 Impact factor: 4.854
Fig. 1Steps of the proposed methodology
Fig. 2Traditional Feed Forward ANN model
Fig. 3A perceptron structure
Fig. 4LSTM Cell—Source: (Kang, 2017)
Stationary analysis results
| Value | |
|---|---|
| ADF Statistic | − 3.433 |
| 0.034205 | |
| Critical Value 1% | − 2.863 |
| Critical Value 5% | − 2.567 |
| Critical Value 10% | − 2.567 |
The steps and the results of stepwise regression
| Steps | Included Variable | Excluded Variable | |
|---|---|---|---|
| 1 | Point of Sales(POS) Data | 0.0000 < 0.05 | |
| 2 | Consumer Price Index (CPI) | 0.0000 < 0.05 | |
| 3 | Sales ($) | 0.0000 < 0.05 | |
| 4 | Advertising Expenses(Internet) | 0.0000 < 0.05 | |
| 5 | Unit Price ($) | 0.0003 < 0.05 | |
| 6 | Advertising Expenses(Radio) | 0.0069 < 0.05 | |
| 7 | Advertising Expenses(TV) | 0.0045 < 0.05 | |
| 8 | Advertising Expenses (SMS) | 0.0000 < 0.05 | |
| Advertising Expenses(Radio) | 0.1589 > 0.05 | ||
| 9 | Consumer Confidence Index(CCI) | 0.0359 < 0.05 | |
| 10 | Loops have ended since there are no variables with p values < 0.05 | ||
| Selected Variables based on stepwise regression | ['POS/ Supply Data', 'Consumer Price Index (CPI)', 'SALES ($)', 'Advertising Expenses(Internet)', 'Unit Price ($)', 'Advertising Expenses(TV)', 'Advertising Expenses (SMS)', 'Consumer Confidence Index(CCI)'] |
Descriptive statistics for the variables in the models
| Demand | SALES ($) | Consumer Price Index (CPI) | POS Data | Advertising Expenses (Internet) | Unit Price ($) | Advertising Expenses (TV) | Advertising Expenses (SMS) | Consumer Confidence Index(CCI) | |
|---|---|---|---|---|---|---|---|---|---|
| Number of observations | 2613 | 2613 | 2613,00 | 2613 | 2613 | 2613 | 2613 | 2613 | 2613 |
| Mean | 5021 | 1,641,507 | 102,61 | 4523 | 3079 | 363 | 1.325 | 60 | 103 |
| Std. Deviation | 2681 | 941,667.3 | 1,38 | 2604 | 1521 | 26 | 124 | 14 | 3 |
| Minimum | 1610 | 462,709.6 | 101,30 | 1510 | 0 | 282 | 1067 | 38 | 96 |
| Maximum | 18,565 | 5,960,221.0 | 106,50 | 16,482 | 6355 | 400 | 1479 | 90 | 108 |
F-Test: Two-Sample for Variances
| Low Ad expenses Group (Q1) | High Ad expenses Group (Q3) | |
|---|---|---|
| Mean | 2425,400,795 | 6274,31,072 |
| Variance | 388,516,802 | 291,278,2142 |
| Number of Observations | 652 | 653 |
| F | 1,333,834,057 | |
| P value one-tail | 0,000,121,901 | |
| F Critical one-tail | 1,137,661,401 |
t-Test: Two-Sample Assuming Unequal Variances
| t Statistic | − 119,2,381,117 |
| P value (one-tail) | < 0.01 |
| t Critical one-tail | 1,646,048,676 |
| P value (two-tail) | < 0.01 |
| t Critical two-tail | 1,961,824,866 |
A part of the formed instances
| Time | Demand(t−1) | Demand(t) |
|---|---|---|
| 1 | 4384 | 4366 |
| 2 | 4366 | 4006 |
| 3 | 4006 | 4076 |
| 4 | 4076 | 4834 |
Candidate values for the hyperparameters
| Model | Hyperparameters | Values |
|---|---|---|
| SVR | Kernel function | Radial Basis Function (RBF), Linear, Polynomial, Sigmoid |
| C Parameter | 0.1, 1, 10, 100, 1000 | |
| DTR | Maximum depth | [10, 20, 30, 40, 50, 60, 70, 80] |
| The minimum number of samples in an internal node | [0.5, 2, 4, 6] | |
| The minimum number of observations at a terminal node | [1, 2, 4, 6] | |
| RFR | Number of trees | [400, 600, 800, 1000, 1200, 1400, 1600] |
| Maximum depth | [10, 20, 30, 40, 50, 60, 70, 80] | |
| The minimum number of observations at a terminal node | [1, 2, 4, 6] | |
| ANN, LSTM | Number of Hidden layers | [1,2,3] |
| Number of Neurons | [4, 8, 16, 32, 64] | |
| Epoch size | [50, 100, 250, 500, 750] | |
| Batch Size | [1, 5, 10, 25, 50, 75, 100] |
The best hyperparameters for the utilized models
| Model | Hyperparameters | Values | |
|---|---|---|---|
| Low Ad Expenses | High Ad expenses | ||
| SVR | Kernel function | RBF | RBF |
| C Parameter | 10 | 10 | |
| DTR | Maximum depth | 10 | 10 |
| The minimum number of samples in an internal node | 0.5 | 2 | |
| The minimum number of observations at a terminal node | 1 | 6 | |
| RFR | Number of trees | 1000 | 1600 |
| Maximum depth | 30 | 60 | |
| The minimum number of observations at a terminal node | 6 | 6 | |
| ANN | Number of Hidden layers | 1 | 1 |
| Number of Neurons | 4 | 32 | |
| Epoch size | 750 | 50 | |
| Batch Size | 5 | 1 | |
| LSTM | Number of Hidden layers | 1 | 1 |
| Number of Neurons | 8 | 8 | |
| Epoch size | 50 | 250 | |
| Batch Size | 25 | 25 |
RMSE and MAE values for the utilized models
| Method | Low Ads | High Ads | Change Between Low and High Ads | |||
|---|---|---|---|---|---|---|
| RMSE | MAE | RMSE | MAE | Improvement in RMSE (%) | Improvement in MAE (%) | |
| SVR | 1473.96 | 1166.37 | 1056.29 | 879.76 | 28 | 25 |
| DTR | 1581.91 | 1196.18 | 956.71 | 716.94 | 40 | 40 |
| RFR | 1678.37 | 1308.17 | 953.96 | 723.81 | 43 | 45 |
| ANN | 1365.38 | 1035.64 | 918.61 | 686.54 | 33 | 34 |
| LSTM | 19 | 12 | ||||
*The lowest RMSE and MAE values mean the highest accuracy
Fig. 5a LSTM results for Low Ad Group. b. LSTM results for High Ad Group
Fig. 6a ANN results for Low Ad Group. b. ANN results for High Ad Group