Literature DB >> 34045493

Simultaneous corn and soybean yield prediction from remote sensing data using deep transfer learning.

Saeed Khaki1, Hieu Pham2, Lizhi Wang3.   

Abstract

Large-scale crop yield estimation is, in part, made possible due to the availability of remote sensing data allowing for the continuous monitoring of crops throughout their growth cycle. Having this information allows stakeholders the ability to make real-time decisions to maximize yield potential. Although various models exist that predict yield from remote sensing data, there currently does not exist an approach that can estimate yield for multiple crops simultaneously, and thus leads to more accurate predictions. A model that predicts the yield of multiple crops and concurrently considers the interaction between multiple crop yields. We propose a new convolutional neural network model called YieldNet which utilizes a novel deep learning framework that uses transfer learning between corn and soybean yield predictions by sharing the weights of the backbone feature extractor. Additionally, to consider the multi-target response variable, we propose a new loss function. We conduct our experiment using data from 1132 counties for corn and 1076 counties for soybean across the United States. Numerical results demonstrate that our proposed method accurately predicts corn and soybean yield from one to four months before the harvest with an MAE being 8.74% and 8.70% of the average yield, respectively, and is competitive to other state-of-the-art approaches.

Entities:  

Year:  2021        PMID: 34045493      PMCID: PMC8159996          DOI: 10.1038/s41598-021-89779-z

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

The use of satellites and other remote sensing mechanisms has proven to be vital in monitoring the growth of various crops both spatially and temporally[1]. Common uses of remote sensing are to extract various descriptive indices such as normalized difference vegetation index (NDVI), temperature condition index, enhanced vegetation index, and leaf area index[2]. This information can be used for drought monitoring, detecting excessive soil wetness, quantifying weather impacts on vegetation, the evaluation of vegetation health and productivity, and crop yield forecasting[3-5]. Moreover, a large-scale collection of environmental descriptors such as surface temperature and precipitation can also be identified from satellite images[6,7]. Accurate and non-destructive identification of crop yield throughout a crop’s growth stage enables farmers, commercial breeding organizations, and government agencies to make decisions to maximize yield output, ultimately for a nation’s economic benefit. Early estimations of yield at the field/farm level play a vital role in crop management in terms of site-specific decisions to maximize a crop’s potential. Numerous approaches have been applied for crop yield prediction, such as manual surveys, agro-meteorological models, and remote sensing based methods[8]. However, given the size of farming operations, manual field surveys are neither efficient nor practical. Moreover, modeling approaches have limitations due to the difficulty in determining model parameters, the scalability to large areas, and the computational resources necessary[9]. Recently, satellite data has become widely available in various spatial, temporal, and spectral resolutions and can be used to predict crop yield over different geographical locations and scales[10]. The availability of Earth observation (EO) data has created new ways for efficient, large-scale agricultural mapping[11]. EO data enables a unique mechanism to capture crop information over large areas with regular updates to create maps of crop production and yield. Nevertheless, due to the high spatial resolution needed for accurate yield predictions, Unmanned Aerial Vehicles (UAV) have been promoted for data acquisition[12]. Although UAV platforms have demonstrated superior image capturing abilities, without the assistance of a large workforce, accurately measuring the yield for large regions is not feasible. Indeed, advances in technology have made yield prediction more accessible and accurate both through machine learning and statistical approaches. Historically, common methods to predict crop yield include random forests, linear regression, and ensemble approaches[13-15]. However, recently crop yield predictions have been dominated by deep learning approaches. Recent works have applied multi-layer perceptrons to predict yield in wheat, corn, and strawberry yield by combining observed phenotypic data and environmental data[16-19]. Moreover, there is an increasing amount of literature combining convolutional neural networks and yield prediction from UAV imagery[20-22]. It is evident from recent works that no matter the data acquisition mechanism, there is an apparent shift in utilizing deep neural networks for crop yield predictions. In regards to the scope of our paper, remote sensing capabilities have been extensively used for estimating crop yield around the world of various scales (field, county, state). Various studies have been performed using different vegetation indices to estimate yield in maize, wheat, grapes, rice, corn, and soybeans using random forests, neural networks, multiple linear regression, partial least squares regression, and crop models[23-31]. Similar to tabular yield data and UAV imagery, remote sensing yield predictions were historically dominated by traditionally machine learning and statistical approaches. However, there is a recent trend towards combining convolutional neural networks with satellite imagery for crop yield predictions[32-34]. These research papers showcase the potential, power, and accuracy deep learning and remote sensing have on estimating yield at a large scale. For this paper, we consider two main crops in the Midwest United States (US), corn (Zea mays L.) and soybeans (Glycine max). According to the United States Department of Agriculture, in 2019, 89.7 million and 76.1 million acres of corn and soybean were planted, respectively. The combination of these two crops makes up approximately 20% of active US farmland[35]. Given the sheer size of these farming operations combined with the growing population, actions must be taken to maximize yield. With the help of remote sensing, government officials, as well as farmers, can monitor crop growth (at various scales) to ensure proper crop management decisions are made to maximize the yield potential. Although approaches exist to estimate yield with various means, current models are limited in that they only estimate yield for a single crop. That is, there does not exist any model making use of remote sensing data to predict the yield of multiple crops simultaneously. More specifically, the scientific question we answer in this paper is determining how to extract transferable information about the yield prediction of one crop to be used for the yield prediction of another crop. Moreover, because individual models need to be constructed for each crop, long computational times prevent machine learning methods from being adopted over large areas[36,37]. One way to alleviate these issues is by considering simultaneous (multi-target) regression of yield. By predicting multiple response variables simultaneously, interaction effects are considered between response variables leading to improved accuracy, and computational performance is reduced due to having a single model predict multiple outputs as opposed to multiple models predicting multiple outputs. Multi-target regression using support vector regression has been successful in estimating different biophysical parameters from remote sensing images outperforming single-target regression techniques both in terms of computational performance and accuracy[38]. Additionally, a multi-target Gaussian regressor created to identify canopy biomass in rice paddies was shown to obtain superior accuracy when compared to a single-target Gaussian regression[39]. Indeed, across a variety of domains and applications, simultaneous regression of multiple responses has shown promise with improved accuracy and computational performance[40-42]. Given the promising attributes of a multi-target regression model, we apply this approach to our remote sensing use case. Having such a model would, hopefully, result in more accurate predictions by considering the interactions between crops and estimating yields accordingly. Additionally, because predictions are simultaneous, the results are less computationally intensive than computing a model for each crop. Therefore, we propose a new deep learning framework called YieldNet utilizing a novel deep neural network architecture. This model makes use of a transfer learning methodology that acts as a bridge to share model weights for the backbone feature extractor. Moreover, because of the uniqueness associated with simultaneous yield prediction for two crops, we propose a new loss function that can handle multiple response variables. Specifically, the novelties of our approach include: A new model that extracts transferable information from both corn and soybean to improve the yield prediction of both crops simultaneously. The proposed method utilizes transfer learning through sharing weights of the backbone feature extractor among multiple crops, therefore, improving computational efficiency. To the best of our knowledge, this is the first deep learning approach combining multi-target regression, convolutional neural networks, and remote sensing data for crop yield prediction. A new loss function is proposed to consider the multi-target response variable. The weight sharing property of the proposed method substantially decreases the number of model parameters and subsequently helps the training process despite having limited labeled data. The effectiveness of our proposed method is demonstrated on large-scale geospatial data from 1132 counties for corn and 1076 counties for soybean covering 13 states across the United States.

Methodology

The goal of this paper is to simultaneously predict the average yield per unit area of two different crops (corn and soybean) both grown in regions of interest (US counties) based on a sequence of images taken by satellite before harvest. Let and denote, respectively, the remotely sensed multi-spectral image taken at time during the growing season and the ground truth average yield of crop type planted in location l at year k, where H and W are the image’s height and width and d is the number of bands (channels). Thus, the dataset can be represented aswhere L and K are the numbers of locations and years, respectively. For our approach, we do not use end-to-end training due to the main following reasons: (1) the number of labeled training data is limited, and (2) inability to use transfer learning from popular benchmark datasets such as Imagenet[43] due to domain difference and multi-spectral nature of satellite images. Therefore, we reduce the dimensionality of the raw remote sensing images under the permutation invariance assumption which states the average yield mostly depends on the number of different pixel types rather than the position of the pixels in images due to the infeasibility of end-to-end training. Similar approaches have been used in other studies[34,44]. As a result, we separately discretize the pixel values for each band of a multi-spectral image into b bins and obtain a histogram representation . If we obtain histogram representations of the sequence of multi-spectral images denoted as and concatenate them through time dimension, we can produce a compact histogram representations of the sequence of multi-spectral images taken during growing season. As such, the dataset D can be re-written as the following and we will use this notation throughout the rest of the paper:Given the above dataset, this paper proposes a deep learning based method, named YieldNet, that learns the desired mapping to predict the yields of two different crops simultaneously.

Network architecture

Crop yield is a highly complex trait that is affected by many factors such as environmental conditions and crop genotype which requires a complex model to reveal the functional relationship between these interactive factors and crop yield. We propose a novel convolutional neural network architecture which is a highly non-linear and complex model. Convolutional neural networks belong to the class of representation learning methods which automatically extract necessary features from raw data without the need for any handcrafted features[45,46]. Figure 1 outlines the architecture of the proposed method. Table 1 shows the detailed architecture of the proposed model. We use a 2-D convolution operation which is performed over the ‘time’ and ‘bin’ dimensions while considering bands as channels. As such, convolution operation over the time dimension can help capture the temporal effect of satellite images collected over time intervals.
Figure 1

Outline of the YieldNet architecture. The input of the YieldNet is 3-D histograms and the 2-D convolution operation is performed over the ‘time’ and ‘bin’ dimensions while considering bands as channels. The number under feature maps indicates the number of channels. Convolutional layers denoted with yellow color work as a backbone feature extractor and share weights for both corn and soybean yield predictions.

Table 1

YieldNet architecture.

Type/stridePaddingFilter sizeNumber of Filters
Conv/s2Valid\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$7\times 7$$\end{document}7×748
Conv/s2Valid\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$5\times 5$$\end{document}5×564
Conv/s2Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$5\times 5$$\end{document}5×596
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3128
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3128
Corn head
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3148
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3148
FC-100
FC-50
Soybean head
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3148
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3$$\end{document}3×3148
FC-100
FC-50

The first five convolutional layers work as a backbone feature extractor and share weights for both corn and soybean yield predictions.

Outline of the YieldNet architecture. The input of the YieldNet is 3-D histograms and the 2-D convolution operation is performed over the ‘time’ and ‘bin’ dimensions while considering bands as channels. The number under feature maps indicates the number of channels. Convolutional layers denoted with yellow color work as a backbone feature extractor and share weights for both corn and soybean yield predictions. YieldNet architecture. The first five convolutional layers work as a backbone feature extractor and share weights for both corn and soybean yield predictions. In our proposed network, the first five convolutional layers share weights for corn and soybean yield predictions which are denoted with yellow color in Fig. 1. These layers extract relevant features from input data for both corn and soybean yield predictions. The intuition behind using one common feature extractor is that it significantly decreases the number of parameters of the network, which helps training the model more efficiently given the scarcity of the labeled data. In addition, many low-level features captured by the comment feature extractor reflect general environmental conditions that are transferable between corn and soybean yields. All convolutional layers are followed by batch normalization[47] and ReLU nonlinearities in our proposed network. Batch normalization is used to accelerate the training process by reducing internal covariate shifts and regularizing the network. We use two convolutional layers in both corn and soybean heads which are followed by two fully connected layers.

Network loss

To jointly minimize the prediction errors of corn and soybean yield forecasting, we propose the following loss function:where , , , , , , , and denote ith average ground truth corn yield, ith predicted corn yield, ith average ground truth soybean yield, ith predicted soybean yield, number of corn samples, number of soybean samples, average corn yield, and average soybean yield, respectively. Our proposed loss function is a normalized Euclidean loss which makes the corn and soybean losses to have the same scale. We use the maximum function in our proposed loss function to make the training process more stable and ensure that both corn and soybean losses are optimized.

Experiments and results

In this section, we present the dataset used in our study and then report the results of our proposed method along with other competing methods in corn and soybean yield predictions. We conducted all experiments in Tensorflow[48] on an NVIDIA Tesla V100 GPU.

Data

The data analyzed in this study included three sets: yield performance, satellite images, and cropland data layers. The yield performance dataset includes the observed county-level average yield for corn and soybean between 2004 and 2018 across 1132 counties for corn and 1076 for soybean within 13 states of the US Corn Belt: Indiana, Illinois, Iowa, Minnesota, Missouri, Nebraska, Kansas, North Dakota, South Dakota, Ohio, Kentucky, Michigan, and Wisconsin, where corn and soybean are considered the dominant crops[49]. Figures 2 and 4 depict the US Corn Belt and histogram of yields, respectively. The summary statistics of the corn and soybean yields are shown in Table 2.
Figure 2

The map of US Corn Belt.

Figure 4

The histograms of the corn and soybean yields. The unit of yield is bushels per acre.

Table 2

The summary statistics of yield data across all years.

Summary statisticsCornSoybean
Year range2004–20182004–2018
Average yield146.6845.02
Standard deviation of yield36.0310.08
Minimum yield18.39.3
Maximum yield246.782.3
Number of locations11321076
Number of observations13,99212,502

The unit of yield is bushels per acre.

Satellite data contains MODIS products including MOD09A1 and MYD11A2. MOD09A1 product provides an estimate of the surface spectral reflectance of Terra MODIS bands 1-7 at 500m resolution and corrected for atmospheric conditions[50]. The MYD11A2 product provides an average land surface temperature which has the day and night-time surface temperature bands at 1km resolution[51]. It is worth noting that MOD09Q1 data exists at 250m resolution, however, it encompasses only 2 reflectance bands. Thus, although higher resolution, the data contains less information[52]. Figure 3 depicts an example multispectral image of land surface temperature and surface spectral reflectance for Adams county in Illinois. These satellite images were captured at 8-days intervals and we only use satellite images captured during growing seasons (March–October). As such, satellite images are collected 30 times a year in our study. We discretize all multispectral images using 32 bins to generate the 3-D histograms , where , , and .
Figure 3

The multispectral images of surface spectral reflectance and land surface temperature for Adams county in Illinois captured on May 8th, 2011. Corn and soybean farmland images indicate cropland areas within the Adams county in Illinois in year 2011.

USDA-NASS cropland data layers (CDL) is crop-specific land cover data that are produced annually for different crops based on moderate resolution satellite imagery and extensive agricultural ground truth[53]. In this paper, cropland data layers are used for both corn and soybean to focus on only croplands within each county and exclude non-croplands such as buildings and streets from satellite images. The map of US Corn Belt. The summary statistics of yield data across all years. The unit of yield is bushels per acre. The multispectral images of surface spectral reflectance and land surface temperature for Adams county in Illinois captured on May 8th, 2011. Corn and soybean farmland images indicate cropland areas within the Adams county in Illinois in year 2011. The histograms of the corn and soybean yields. The unit of yield is bushels per acre.

Design of experiments

We compare our proposed method with the following models to evaluate the efficiency of our method.

Random forest (RF)[54]

RF is a non-parametric ensemble learning method that is robust against overfitting. We set the number and the maximum depth of trees in RF to be 150 and 20, respectively. We tried different numbers and the maximum depth of trees and found that these hyperparameters resulted in the most accurate predictions.

Deep feed forward neural network (DFNN)

DFNN is a highly nonlinear model which stacks multiple fully connected layers to learn the underlying functional relationship between inputs and outputs. The DFNN model has nine hidden layers, each having 50 neurons. We used batch normalization for all layers. The ReLU activation function was used in all layers except the output layer. The model was trained for 120,000 iterations. Adam optimizer was used to minimize the Euclidean loss function.

3-dimensional convolutional neural network (3D-CNN)

3D-CNN is a highly non-linear model which employs 3-dimensional kernels in its convolution operations to capture spatio-temporal features in the data[55]. The 3D-CNN captures the temporal effects of remote sensing data as well as spatial and intra-band features extracted from the individual images[56]. In this paper, we used a homogeneous network architecture for the 3D-CNN which is found effective in other studies[55,57]. Table 3 shows the network architecture of the 3D-CNN.
Table 3

3D-CNN network architecture.

Type/stridePaddingFilter sizeNumber of filters
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×164
Max poolSame\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$4\times 4\times 2$$\end{document}4×4×2
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×164
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×164
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×164
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×164
Conv/s1Same\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3\times 3\times 1$$\end{document}3×3×1128
Max poolSame\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$2\times 2\times 1$$\end{document}2×2×1
FC-256
FC-128

The input of the 3D-CNN is 3-D histograms and the 3-D convolution operation is performed over the ‘time’, ‘bin’, and ‘band’ dimensions.

3D-CNN network architecture. The input of the 3D-CNN is 3-D histograms and the 3-D convolution operation is performed over the ‘time’, ‘bin’, and ‘band’ dimensions.

Regression tree (RT)[58]

RT is a nonlinear model which does not make any assumption regarding the mapping function. We set the maximum depth of the regression tree to be 12 to decrease the overfitting.

Lasso[59]

Lasso adopts norm to reduce the effect of unimportant features. A Lasso model also can be used as a baseline model to compare the linear and nonlinear effects of input features. We set the coefficient of our Lasso model to be 0.05.

Ridge[60]

Ridge regression is similar to the Lasso model except it uses norm as a penalty to reduce the effect of unimportant features. We set the coefficient of our Ridge model to be 0.05.

Training details

YieldNet network was trained in an end-to-end manner. We initialized the network parameters with Xavier initialization[61]. To minimize the loss function defined in Eq. (1), we used Adam optimizer[62] with a learning rate of 0.05% and a mini-batch size of 32. The network was trained 4000 iterations to convergence. We do not use dropout[63] because batch normalization also has a regularization effect on the network.

Final results

After having trained all models, we evaluated the performances of our proposed method along with other competing models to predict corn and soybean yields. To completely evaluate all models, we took 3 years 2016, 2017, and 2018 as test years and predicted corn and soybean yields 4 times a year during the growing season on the 23rd day of July, August, September, and October. From a practical perspective, monitoring crop yield throughout the growing season is vital for optimal farm management practices (when to add fertilizer, apply pesticides, irrigate crops, etc). In addition, crop yield monitoring affects crop commodity market, which determines the future prices of crops. Therefore, our results emulate the situation where we are progressively predicting yield at the end of the growing season months earlier as a means of tracking expected crop growth. Tables 4 and 5 present the corn and soybean yield prediction results, respectively, and compare the performances of models with respect to the root–mean–square error (RMSE) evaluation metric which is defined as follows:where, N, , and denote the number of observations, the predicted yield for ith observation, and the ground truth yield for ith observation, respectively.
Table 4

The RMSE of corn yield prediction performance of models.

Test dateModels
YearMonthRidgeLassoRFDFNNRT3D-CNNYieldNet
2016July23.1221.0322.4822.1629.4118.8418.73
August23.1619.6820.9520.4829.1615.2515.76
September24.5320.621.2321.0429.3116.5515.96
October24.9321.0521.1520.7427.9616.6515.85
2017July30.5527.5326.6126.4033.6422.5020.88
August25.1622.2722.2520.8528.0216.6017.74
September24.1521.521.9919.2126.815.7115.53
October25.7320.9422.1418.9026.7815.6915.40
2018July27.5121.2122.3822.8527.6920.6422.08
August24.519.4621.5221.1429.3418.8118.25
September25.118.6921.720.5728.9117.5816.89
October32.519.222.2821.6328.917.7216.75
Average25.9121.1022.2221.3328.8317.7117.49

The average ± standard deviation for corn yield in years 2016, 2017, and 2018 are, respectively, , , and . The numbers of test samples in years 2016, 2017, 2018 for corn yield prediction are 885, 882, and 784, respectively. The unit of RMSE is bushels per acre.

Best results are given in bold.

Table 5

The RMSE of soybean yield prediction performance of models.

Test dateModels
YearMonthRidgeLassoRFDFNNRT3D-CNNYieldNet
2016July9.879.129.388.2710.827.085.43
August8.28.429.156.8910.655.364.59
September8.38.279.127.1910.374.84.45
October8.498.859.027.6210.74.94.24
2017July8.817.016.626.309.635.665.83
August6.776.115.655.427.854.615.11
September6.826.445.625.577.884.424.55
October6.746.635.675.407.794.384.35
2018July8.779.888.138.629.746.586.36
August7.758.087.787.149.585.85.59
September8.258.047.787.539.645.675.16
October8.018.147.97.419.45.415.19
Average8.077.927.656.959.505.385.07

The average ± standard deviation for corn yield in years 2016, 2017, and 2018 are, respectively, , , and . The numbers of test samples in years 2016, 2017, and 2018 for soybean yield prediction are 773, 776, and 663, respectively. The unit of RMSE is bushels per acre.

Best results are given in bold.

The RMSE of corn yield prediction performance of models. The average ± standard deviation for corn yield in years 2016, 2017, and 2018 are, respectively, , , and . The numbers of test samples in years 2016, 2017, 2018 for corn yield prediction are 885, 882, and 784, respectively. The unit of RMSE is bushels per acre. Best results are given in bold. The RMSE of soybean yield prediction performance of models. The average ± standard deviation for corn yield in years 2016, 2017, and 2018 are, respectively, , , and . The numbers of test samples in years 2016, 2017, and 2018 for soybean yield prediction are 773, 776, and 663, respectively. The unit of RMSE is bushels per acre. Best results are given in bold. As shown in Tables 4 and 5, our proposed method outperforms other methods to varying extents. The Ridge and Lasso had comparable performances for soybean yield prediction, but, Lasso performed considerably better compared to the Ridge for corn yield prediction. DFNN showed a better performance than RF, RT, and Ridge for both corn and soybean yield predictions. DFNN had similar performance with Lasso for corn yield prediction while having higher prediction accuracy for soybean yield prediction. Despite the linear modeling structure, Lasso performed better than RF and RT for corn yield prediction, which indicates that RF and RT cannot successfully capture the nonlinearity of remote sensing data, resulting in poor performance compared to Lasso. RT had a weak performance compared to other methods due to being prone to overfitting. RF performs considerably better than RT because of using ensemble learning, which makes it robust against overfitting. 3D-CNN outperformed other models except for our proposed method due to capturing the temporal effects and spatial and intra-band features of remote sensing data. 3D-CNN performed slightly better than YieldNet in three cases for soybean and corn yield predictions. But, YieldNet outperformed the 3D-CNN on average for both corn and soybean yield predictions while having a smaller number of parameters and computation time. Our proposed method outperformed the other methods due to multiple factors: (1) the convolution operation in the YieldNet model captures both the temporal effect of remote sensing data collected over growing season and the spatial information of bins in histograms, (2) the YieldNet network uses transfer learning between corn and soybean yield predictions by sharing the weights of the backbone feature extractor, and (3) using a shared backbone feature extractor in the YieldNet model substantially decreases the number of model’s parameters and subsequently helps training process despite having the limited labeled data. The prediction accuracy decreases as we try to make predictions earlier during the growing season (e.g. July and August) due to loss of information. All models except our proposed model do not show a clear decreasing pattern in performance accuracy as we go from October to July, which indicates they cannot fully learn the functional mapping from satellite images to the yield. Table 6 compares the number of parameters and training time of the competing methods for crop yield prediction. As shown in Table 6, Lasso and Ridge have the lowest number of parameters compared to neural network based models such as DFNN, 3D-CNN, and YieldNet due to their linear modeling structure. Compared to the YieldNet model, other models should be trained separately for corn and soybean, which results in having twice the total number of parameters. 3D-CNN has the highest number of parameters among the neural network based models. From a computation time perspective, linear models and RT had the lowest training time. RF had the second longest training time after 3D-CNN compared to other models due to using ensemble learning. Among neural network based models, YieldNet had the shortest training time and 3D-CNN had the longest training time. The models were trained on an Intel i7-4790 CPU 3.60 GHz. The inference times of all methods are less than a second. However, the inference time of our model is less than that of the 3D-CNN model.
Table 6

The number of parameters and training time of the competing methods for crop yield prediction.

MeasuresModels
RidgeLassoRFDFNNRT3D-CNNYieldNet
Number of parameters86418641Non-parametric923,882Non-parametric3,167,4901,436,050
Training time (min)0.421.94155.1431.461.44759.42.89

The training time indicates the total time needed for training models for both corn and soybean yield prediction. The models were trained on an Intel i7-4790 CPU 3.60 GHz.

The number of parameters and training time of the competing methods for crop yield prediction. The training time indicates the total time needed for training models for both corn and soybean yield prediction. The models were trained on an Intel i7-4790 CPU 3.60 GHz. We also report the yield prediction performance of our proposed model with respect to another evaluation metric, mean absolute error (MAE), in Table 7. As shown in Table 7, our proposed method accurately predicted corn yield 1 month, 2 months, 3 months, and 4 months before the harvest with MAE being 9.92%, 8.88%, 8.36%, and 7.8% of the average corn yield, respectively. The proposed model also accurately predicted soybean yield one month, two months, three months, and four months before harvest with MAE being 10.05%, 9.06%, 8.01%, and 7.67% of the average soybean yield, respectively. The proposed model is slightly more accurate in soybean yield forecasting than corn yield forecasting, which is due to the higher variation in the corn yield compared to the soybean yield.
Table 7

The MAE of the corn and soybean yield prediction performances of our proposed model.

Crop201620172018
Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.
Corn14.9214.3214.3614.4818.2415.9212.7113.4016.1114.5913.3513.23
Soybean6.054.984.103.665.05.133.734.115.184.903.914.19

The unit of MAE is bushels per acre.

The MAE of the corn and soybean yield prediction performances of our proposed model. The unit of MAE is bushels per acre. We visualized the error percentage maps for the corn and soybean yield predictions for the year 2018. As shown in Figs. 5 and 6, the error percentage is below 5% for most counties, which indicates that our proposed model provides a robust and accurate yield prediction across US Corn Belt.
Figure 5

The error percentage maps for the 2018 corn yield prediction which is done during growing season in July, August, September, and October. The counties with white color indicate the ground truth yield were not available for those counties in 2018.

Figure 6

The error percentage maps for the 2018 soybean yield prediction which is done during the growing season in July, August, September, and October. The counties with white color indicate the ground truth yield were not available for those counties in 2018.

The error percentage maps for the 2018 corn yield prediction which is done during growing season in July, August, September, and October. The counties with white color indicate the ground truth yield were not available for those counties in 2018. The error percentage maps for the 2018 soybean yield prediction which is done during the growing season in July, August, September, and October. The counties with white color indicate the ground truth yield were not available for those counties in 2018. To further evaluate the prediction results of our proposed model, we created the scatter plots of ground truth yield against the predicted yield for the year 2018. Figure 7 depicts the scatter plots for the corn yield prediction during the growing season in the months July, August, September, and October. The corn scatter plots indicate that the YieldNet model can successfully forecast yield months prior to harvesting.
Figure 7

The scatter plots for the 2018 corn yield prediction during the growing season in the months July, August, September, and October. MAE and r stand for mean absolute error and correlation coefficient, respectively. The unit of yield is bushels per acre.

The scatter plots for the 2018 corn yield prediction during the growing season in the months July, August, September, and October. MAE and r stand for mean absolute error and correlation coefficient, respectively. The unit of yield is bushels per acre. Figure 8 depicts the scatter plots for the soybean yield prediction during the growing season in the months July, August, September, and October. The corn scatter plots indicate that the YieldNet model provides reliable and accurate yield months prior to the harvest.
Figure 8

The scatter plots for the 2018 soybean yield prediction during the growing season in months July, August, September, and October. MAE and r stand for mean absolute error and correlation coefficient, respectively. The unit of yield is bushels per acre.

The scatter plots for the 2018 soybean yield prediction during the growing season in months July, August, September, and October. MAE and r stand for mean absolute error and correlation coefficient, respectively. The unit of yield is bushels per acre.

Ablation study

In order to examine the usefulness of using a single deep learning model for simultaneously predicting the yield of two crops, we perform the following analysis. We train two separate models one for corn yield prediction and another for soybean yield prediction which are as follows: This model has exactly the same network architecture as the YieldNet model except we removed the soybean head from the original YieldNet network. As a result, can only predict the corn yield. This model has exactly the same network architecture as the YieldNet model except we removed the corn head from the original YieldNet network. As a result, can only predict the soybean yield. Tables 8 and 9 compare the yield prediction performances of the above-mentioned models with the original YieldNet model.
Table 8

The RMSE of the corn yield prediction performances of the YieldNet and models.

Model201620172018
Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{YieldNet}^{\textit{corn}}$$\end{document}YieldNetcorn21.7418.8817.5516.9023.8018.8416.8316.8323.7118.6017.8417.62
YieldNet18.7315.7615.9615.8520.8817.7415.5315.4022.0818.2516.8916.75

The unit of RMSE is bushels per acre.

Table 9

The RMSE of the soybean yield prediction performances of the YieldNet and models.

Model201620172018
Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.Jul.Aug.Sep.Oct.
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\textit{YieldNet}^{\textit{soy}}$$\end{document}YieldNetsoy6.495.495.154.576.965.824.864.867.806.585.465.57
YieldNet5.434.594.454.245.835.114.554.356.365.595.165.19

The unit of yield is bushels per acre.

The RMSE of the corn yield prediction performances of the YieldNet and models. The unit of RMSE is bushels per acre. The RMSE of the soybean yield prediction performances of the YieldNet and models. The unit of yield is bushels per acre. As shown in Tables 8 and 9, the YieldNet model which simultaneously predicts corn and soybean yields outperforms individual and models. The YieldNet provides more robust and accurate yield predictions compared to the other two individual models, which indicates that transfer learning between corn and soybean yield prediction improves the yield prediction accuracy for both crops.

Discussion and conclusion

Our numerical results illustrate that our approach to simultaneously predicting yield for both corn and soybeans is possible and can achieve higher accuracy than individual models. By utilizing transfer learning between corn and soybean yield to share the weights of the backbone feature extractor, YieldNet was able to substantially decrease the number of learning parameters. Our transfer learning approach enabled us to save on computation resources while also maximizing our prediction accuracy. Moreover, the accuracy achieved using a 4-month look-ahead has a lot of important implications for crop management decisions. With accurate yield predictions at various time points, decision-makers now have the ability to change crop management practices to ensure yield is being maximized throughout its growth stage. Although our approach highlighted corn and soybean in the US market, this approach is applicable to any number of crops in any region. Due to the strength of our deep learning framework in combination with a generalized loss function, our approach is ready for scale. To improve the accuracy of our methodology, more data can be gathered and this can be left as a future extension to this work, alongside more crops and more regions. It is the hope of this paper that our approach and results showcase the power deep learning for simultaneous yield prediction can have on the remote sensing community and the larger agricultural community as a whole.
  2 in total

Review 1.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

2.  A CNN-RNN Framework for Crop Yield Prediction.

Authors:  Saeed Khaki; Lizhi Wang; Sotirios V Archontoulis
Journal:  Front Plant Sci       Date:  2020-01-24       Impact factor: 5.753

  2 in total
  4 in total

1.  Construction of Remote Sensing Model of Fresh Corn Biomass Based on Neural Network.

Authors:  Jianjian Chen; Hui Zhang; Yunlong Bian; Xiangnan Li; Guihua Lv
Journal:  Comput Intell Neurosci       Date:  2022-05-31

2.  Winter wheat yield prediction using convolutional neural networks from environmental and phenological data.

Authors:  Nima Safaei; Saeed Khaki; Amit Kumar Srivastava; Gina Lopez; Wenzhi Zeng; Frank Ewert; Thomas Gaiser; Jaber Rahimi
Journal:  Sci Rep       Date:  2022-02-25       Impact factor: 4.379

3.  Research on Rice Yield Prediction Model Based on Deep Learning.

Authors:  Xiao Han; Fangbiao Liu; Xiaoliang He; Fenglou Ling
Journal:  Comput Intell Neurosci       Date:  2022-04-26

4.  Sugarcane Nitrogen Concentration and Irrigation Level Prediction Based on UAV Multispectral Imagery.

Authors:  Xiuhua Li; Yuxuan Ba; Muqing Zhang; Mengling Nong; Ce Yang; Shimin Zhang
Journal:  Sensors (Basel)       Date:  2022-04-01       Impact factor: 3.576

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.