| Literature DB >> 35953059 |
Kyung-Tae Lee1, Juhyeong Han1, Kwang-Hyung Kim1.
Abstract
To predict rice blast, many machine learning methods have been proposed. As the quality and quantity of input data are essential for machine learning techniques, this study develops three artificial neural network (ANN)-based rice blast prediction models by combining two ANN models, the feed-forward neural network (FFNN) and long short-term memory (LSTM), with diverse input datasets, and compares their performance. The Blast_Weather_FFNN model had the highest recall score (66.3%) for rice blast prediction. This model requires two types of input data: blast occurrence data for the last 3 years and weather data (daily maximum temperature, relative humidity, and precipitation) between January and July of the prediction year. This study showed that the performance of an ANN-based disease prediction model was improved by applying suitable machine learning techniques together with the optimization of hyperparameter tuning involving input data. Moreover, we highlight the importance of the systematic collection of long-term disease data.Entities:
Keywords: artificial intelligence; artificial neural network; machine learning; rice blast
Year: 2022 PMID: 35953059 PMCID: PMC9372109 DOI: 10.5423/PPJ.NT.04.2022.0062
Source DB: PubMed Journal: Plant Pathol J ISSN: 1598-2254 Impact factor: 2.321
Fig. 1Distribution map of National Crop Pest Management System (NCPMS) data and weather observation data used in the study. Blue dots indicate the NCPMS data of rice blast occurrence and red dots indicate the weather station sites for weather observation data.
Features of the hyperparameters for each model compared in this study
| Model | Hyperparameters | Range |
|---|---|---|
| Blast_FFNN[ | Year_size[ | 1–9 |
| Units[ | 22–25 | |
| Activation function | Relu, sigmoid, tanh | |
| Blast_Weather_FFNN | Units | 23–27 |
| Activation function | Relu, sigmoid, tanh | |
| Months[ | Jan–Jul to May–Jul | |
| Period[ | 2–30 | |
| Weather_variables | Combinations of tmax, tmin, wspd, prec, rhum | |
| Blast_Weather_LSTM | Units | 22–26 |
| Activation function | Relu, sigmoid, tanh | |
| Months | Jan–Jul to May–Jul | |
| Period | 2–30 | |
| Weather_variables | Combinations of tmax, tmin, wspd, prec, rhum |
FFNN, feed-forward neural network; LSTM, long short-term memory.
As indicated by the name, Blast_FFNN uses only historical rice blast occurrence data, while the Blast_Weather_FFNN and Blast_Weather_LSTM models use both blast occurrence and weather data as inputs.
The number of consecutive years of rice blast occurrence observed in the past.
The number of nodes in each hidden layer.
Selected months during which weather variables are used as input.
The number of timesteps during the selected months.
Fig. 2A flowchart of the development of Blast_Weather_FFNN and Blast_Weather_LSTM models used in this study. Blast_FFNN does not include weather data indicated as the light green box in the figure. FFNN, feed-forward neural network; LSTM, long short-term memory; NCPMS, National Crop Pest Management System.
Performances of rice blast prediction models after optimizing each hyperparameter
| Model | Hyperparameters | Validation | Optimal values[ | |
|---|---|---|---|---|
|
| ||||
| Accuracy (%) | Recall (%) | |||
| Blast_FFNN | Year_size | 67.93–75.26 | 40.06–55.77 | 3 |
| Units | 72.24–73.09 | 52.78–55.99 | 16, 16[ | |
| Activation function | 72.62–73.07 | 55.19–55.99 | Tanh, sigmoid | |
| Blast_Weather_FFNN | Units | 69.32–70.68 | 56.35–65.35 | 8 |
| Activation function | 70.48–71.31 | 64.52–65.92 | Relu | |
| Months | 70.62–70.87 | 63.23–65.08 | Jan–Jul | |
| Period | 70.39–71.41 | 62.03–66.41 | 20 | |
| Weather_variables | 69.86–71.59 | 63.47–66.33 | tmax, prec, rhum | |
| Blast_Weather_LSTM | Units | 70.99–71.53 | 61.85–64.32 | 16 |
| Activation function | 71.18–71.43 | 57.10–64.77 | Relu | |
| Months | 70.25–71.51 | 63.12–64.03 | Mar–Jul | |
| Period | 69.56–70.65 | 61.40–64.34 | 24 | |
| Weather_variables | 69.80–71.30 | 58.66–64.50 | tmax, prec, rhum | |
FFNN, feed-forward neural network; LSTM, long short-term memory.
Optimal values were selected to maximize the recall indicator.
The Blast_FFNN model had two optimal values for each of the two hidden layers, while the Blast_Weather_FFNN and Blast_Weather_LSTM models had one hidden layer.