Next Article in Journal
Post-Cryogenic Viability of Peach (Persica vulgaris Mill.) Dormant Buds from the VIR Genetic Collection
Next Article in Special Issue
Monitoring of Wheat Fusarium Head Blight on Spectral and Textural Analysis of UAV Multispectral Imagery
Previous Article in Journal
Soil and Irrigation Water Salinity, and Its Consequences for Agriculture in Ethiopia: A Systematic Review
Previous Article in Special Issue
Improving Land Use/Cover Classification Accuracy from Random Forest Feature Importance Selection Based on Synergistic Use of Sentinel Data and Digital Elevation Model in Agriculturally Dominated Landscape
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Monitoring of Soybean Maturity Using UAV Remote Sensing and Deep Learning

1
College of Information and Management Science, Henan Agricultural University, Zhengzhou 450002, China
2
Key Laboratory of Quantitative Remote Sensing in Agriculture of Ministry of Agriculture, Beijing Research Center for Information Technology in Agriculture, Beijing 100097, China
3
Key Lab of Smart Agriculture System, Ministry of Education, China Agricultural University, Beijing 100083, China
4
College of Agriculture, Nanjing Agricultural University, Nanjing 210095, China
5
Institute of Agricultural Equipment, Zhejiang Academy of Agricultural Sciences (ZAAS), Hangzhou 310000, China
*
Author to whom correspondence should be addressed.
Agriculture 2023, 13(1), 110; https://doi.org/10.3390/agriculture13010110
Submission received: 22 November 2022 / Revised: 27 December 2022 / Accepted: 28 December 2022 / Published: 30 December 2022

Abstract

:
Soybean breeders must develop early-maturing, standard, and late-maturing varieties for planting at different latitudes to ensure that soybean plants fully utilize solar radiation. Therefore, timely monitoring of soybean breeding line maturity is crucial for soybean harvesting management and yield measurement. Currently, the widely used deep learning models focus more on extracting deep image features, whereas shallow image feature information is ignored. In this study, we designed a new convolutional neural network (CNN) architecture, called DS-SoybeanNet, to improve the performance of unmanned aerial vehicle (UAV)-based soybean maturity information monitoring. DS-SoybeanNet can extract and utilize both shallow and deep image features. We used a high-definition digital camera on board a UAV to collect high-definition soybean canopy digital images. A total of 2662 soybean canopy digital images were obtained from two soybean breeding fields (fields F1 and F2). We compared the soybean maturity classification accuracies of (i) conventional machine learning methods (support vector machine (SVM) and random forest (RF)), (ii) current deep learning methods (InceptionResNetV2, MobileNetV2, and ResNet50), and (iii) our proposed DS-SoybeanNet method. Our results show the following: (1) The conventional machine learning methods (SVM and RF) had faster calculation times than the deep learning methods (InceptionResNetV2, MobileNetV2, and ResNet50) and our proposed DS-SoybeanNet method. For example, the computation speed of RF was 0.03 s per 1000 images. However, the conventional machine learning methods had lower overall accuracies (field F2: 63.37–65.38%) than the proposed DS-SoybeanNet (Field F2: 86.26%). (2) The performances of the current deep learning and conventional machine learning methods notably decreased when tested on a new dataset. For example, the overall accuracies of MobileNetV2 for fields F1 and F2 were 97.52% and 52.75%, respectively. (3) The proposed DS-SoybeanNet model can provide high-performance soybean maturity classification results. It showed a computation speed of 11.770 s per 1000 images and overall accuracies for fields F1 and F2 of 99.19% and 86.26%, respectively.

1. Introduction

Soybeans are a high-quality source of plant protein and raw materials for the production of hundreds of chemical products [1,2]. China’s soybean-growing areas include the Northeast China Plain [3] and the North China Plain [4] (ranging from the north latitude of 30° to 48°). Soybean breeders must develop early-maturing, standard, and late-maturing varieties for planting at different latitudes to ensure that soybean plants fully utilize solar radiation. Therefore, timely and accurate monitoring of soybean breeding line maturity can facilitate soybean breeding decision-making and agricultural management [5,6,7,8].
Traditional methods for measuring field breeding line maturity are time-consuming and labor-intensive [7]. Meanwhile, the expertise and bias of the investigators can affect the accuracy of field surveys. Breeding fields have thousands of breeding lines with different maturation times. Manual surveys cannot quickly provide high-frequency breeding line maturity information to meet harvesting and yield measurement scheduling requirements. Unmanned aerial vehicle (UAV) remote sensing technology can be used to collect high-resolution crop canopy images and has thus been widely used in precision agricultural crop trait monitoring [9,10,11,12]. Compared with satellite and airborne remote sensing technologies, UAV remote sensing technology is relatively inexpensive and flexible in its operation, and it requires less space for landing and takeoff [13]. More importantly, the digital images obtained by low-altitude UAVs have a high ground spatial resolution (centimeter-scale or higher); thus, they contain rich crop-canopy surface information for crop phenotypic research [14,15]. In recent years, UAV remote sensing technology has been widely used to collect crop trait information [9,10,11,12,16,17]. UAVs equipped with high-definition digital cameras can acquire soybean canopy ultrahigh ground spatial resolution digital images over a field scale [14,15]. Many UAV-based methods have been proposed for monitoring various types of crop trait information, including the leaf area index (LAI) [18], leaf chlorophyll content [18,19,20,21], biomass [15,22], and crop height [23].
Machine learning has been successfully applied in several areas, such as image classification, target recognition, and language translation [24,25,26]. In recent years, machine learning techniques have been widely used to recognize various crop traits based on remote sensing images [27]. Gniewko et al. [28] used an artificial neural network (ANN), growing degree days, and total precipitation to estimate soybean yields. Letícia et al. [29] conducted a study to identify nematode damage to soybeans through the use of UAV remote sensing and a random forest (RF) model. The results obtained by Eugenio et al. [30] and Paulo et al. [31] indicated that machine learning techniques are efficient and flexible for remote sensing monitoring of soybean yields. Abdelbaki et al. [32] conducted a study to predict the soybean LAI and fractional vegetation cover (FVC) based on the RF model and UAV remote sensing. Compared with traditional machine learning methods (e.g., SVM and RF), deep learning methods such as long short-term memory (LSTM) [33,34], deep convolutional neural networks (CNNs) [26,35], and transformers [14] have been applied to image recognition, medical image analysis, climate change, and Weiqi game analysis, where they can provide results with similar or even higher precision than human experts. Deep learning uses multiple layers to extract higher-level features from the raw input. In recent years, deep learning techniques have been widely used to recognize various crop traits in remote sensing images, e.g., in leaf disease identification, weed identification, and crop trait recognition [1,26,33,34,35,36,37]. Wang et al. [34] developed an LSTM model by integrating MODIS LAI data to predict crop yields in China. Khan et al. [37] used a YOLOv4 model to identify apple leaf diseases in digital images captured by mobile phones. Zhang et al. [26] used a YOLOv4 model to identify weeds in digital photos of a peanut field. Khalied et al. [38] proposed a model based on MobileNetV2 for fruit identification and classification. Yonis et al. [39] proposed a CNN model adopting the VGG16 architecture for seed identification and classification. Notably, most of these widely used networks (e.g., YOLOv4 [40], ResNet50 [41], MobileNet [42], VGG16 [39], and InceptionResNetV2 [43]) did not take full advantage of shallow features. Shallow features derived from the shallow layers of CNNs are rich in image details, which are generally used in areas such as fine texture detection or small target detection [44,45]. Fusing the deep and shallow features of CNNs may improve performance in soybean maturity classification [44,45,46].
The objective of this work was to monitor soybean maturity using UAV remote sensing and deep learning. We designed a new convolutional neural network architecture (DS-SoybeanNet) to extract and utilize both shallow and deep image features to improve the performance of UAV-based soybean maturity information monitoring. We used a high-definition digital camera on board a UAV to collect high-definition soybean canopy digital images from two soybean breeding fields. We compared the UAV-based soybean maturity information monitoring performances of conventional machine learning methods (support vector machine (SVM) and random forest (RF)), current deep learning methods (InceptionResNetV2, MobileNetV2, and ResNet50), and our proposed DS-SoybeanNet method. Our results indicate that the proposed DS-SoybeanNet method can extract both shallow and deep image feature information and can realize high-performance soybean maturity classification.

2. Materials

2.1. Study Area

The study area was located at the Shengfeng Experimental Station (E: 116°22′10″–116°22′20″, N: 35°25′50″–35°26′20″, Figure 1) of the National Center for Soybean Improvement, Jiaxiang County, Jining City, Shandong Province, China. Jiaxiang County is situated on the North China Plain, with a warm continental monsoon climate, concentrated precipitation, and an average annual sunshine duration of 2405.2 h. The average annual temperature is 13.9 °C.

2.2. UAV Flights and Soybean Canopy Image Collection

We used a high-definition digital camera on board an eight-rotor electric UAV to collect high-resolution soybean canopy remote sensing images (Table 1). In the soybean breeding experimental field, the size of each planting area was approximately 2.5 m × 5 m. As shown in Figure 1, we selected two independent soybean planting fields (fields F1 and F2) in the study area to obtain soybean canopy digital images and maturity information.
For field F1, we conducted five UAV flights (29 July, 13 August, 31 August, 17 September, and 28 September 2015). A total of 2116 soybean canopy digital images and their maturity information were obtained, which were used to calibrate the SoybeanNet model. For field F2, we made only one observation on 30 September, 2015. There were immature, near-mature, mature, and harvested soybean breeding lines in field F2 on 30 September. A total of 546 planting areas were set up in field F2 for the mapping and independent evaluation of the DS-SoybeanNet model.
The soybean image collection and image stitching process mainly included the following three steps:
(1)
Before the UAV took off, we set the flight route information according to the field size; the heading and lateral overlap were set to 80%. Table 1 shows the digital camera exposure parameters.
(2)
During the UAV flight, the soybean canopy images and corresponding position and orientation system (POS) information were collected using the digital camera, inertial measurement unit, and global positioning system device on board the UAV.
(3)
After the UAV flight, we imported the digital images and POS information into PhotoScan software to stitch together the high-definition digital images collected by the UAV. After the image stitching process, five soybean canopy digital orthophoto maps (ground spatial resolution (GSD): 0.016 m) for field F1 and one soybean canopy digital orthophoto map (GSD: 0.016 m) for field F2 were acquired.

2.3. Soybean Canopy Image Labeling

In this study, soybean maturity information was manually labeled. The labeling method was based on the standards of soybean harvesting. The labeling method is described in Table 2. For workers to customize schedules for harvesting soybean planting plots, four categories were used: immature (L0), near-mature (L1), mature (L2), and harvested (L3). L2 plots have the highest harvesting priority and need to be harvested as soon as possible, L1 plots have a high priority because the soybean will mature in less than a week, L0 and L3 plots have a lower priority because L0 plots generally take longer to grow, and no outdoor work is required for L3 plots.
Since different soybean breeding lines have different maturation times, the numbers of images corresponding to the four labels varied between the two fields. Sixty percent of the images of each type in the dataset were randomly chosen to train the model, and the remaining 40% were used to evaluate the model’s accuracy. Table 3 shows the numbers of samples used to train and validate the DS-SoybeanNet model. Figure 2 shows the soybean images used for model calibration and validation.

2.4. Data Enhancement

In this study, we produced a DOM for the entire area by mosaicking together the digital images collected during each UAV flight. Since an orthoimage has a uniform scale, the ground spatial resolutions and solar angles were the main differences between the five DOMs. We used image rotation (four rotation angles: 0° (i.e., the original image), 90°, 180°, and 270°) and scaling (four scaling factors: 1.0 (i.e., the original image), 1.2, 1.5, 1.8, and 2.0) to enhance the soybean canopy image dataset collected from field F1. Image rotation and magnification helped us to obtain soybean canopy images with different resolutions and angles; in addition, they helped prevent overfitting of the model due to the small number of samples collected in the field.
After data enhancement, the number of original soybean canopy images obtained from field F1 was increased by 20 times. The number of independent validation datasets obtained from field F2 was not increased. In this study we used the Python open-cv and NumPy libraries to extract, rotate, and magnify the soybean canopy images.

3. Methods

3.1. Proposed DS-SoybeanNet

CNNs were originally proposed based on the receptive field mechanism in biology and they are a widely used deep learning technology [47]. CNNs are designed to process images with a lattice-like structure. The multilayer convolution, weight sharing, and rotational-shift invariance of CNNs make them effective in image classification and feature recognition. The deep and complex features extracted by CNNs are often used to effectively describe differences between different image categories and can be used to quickly and accurately complete classification tasks. Currently, widely used networks (e.g., ResNet50 and MobileNetV2) ignore shallow image feature information. We designed a network structure (Figure 3) that considers both shallow and deep image features to enhance the model’s generalization ability. The advantage of DS-SoybeanNet is that the shallow and deep features are linked together by means of a concatenation module. Consequently, DS-SoybeanNet can extract and utilize both shallow and deep image features to improve the accuracy of soybean maturity information classifications. Figure 3 shows the architecture of DS-SoybeanNet. DS-SoybeanNet contains five convolutional layers, five flattening modules, one concatenation module, and four fully connected layers. The layers are described as follows:
(1) Input layer
The input data were collected via UAV remote sensing technology in the form of soybean canopy orthophotos and were then manually labeled and cropped to produce sample data. The sample size was 108 × 108 × 3, and the sample data were divided into four types: immature, near-mature, mature, and harvested.
(2) Convolutional and pooling layers
The purpose of the convolution operation was to extract the different features of the input images. DS-SoybeanNet was designed with five convolutional layers; each convolutional layer was combined with the ReLU activation function to achieve delinearization. The pooling layers can reduce the dimensions of the feature maps by summarizing the presence of features in patches of the feature map.
(3) Flattening and concatenation layers
A flattening layer can reshape the feature maps into the dimensions required for the subsequent layers. A concatenation layer concatenates inputs along a specified dimension.
(4) Fully connected layers and output layer
Four fully connected layers were designed, and dropout layers were attached to the first three layers to prevent overfitting and improve model generalization. The output of the model was soybean maturity information derived from the input images.

3.2. Transfer Learning Based on InceptionResNetV2, MobileNetV2, and ResNet50

Transfer learning is a strategy for solving similar or related tasks using existing methods and data. Many deep learning networks show effective performance in image classification and target recognition from natural images (e.g., InceptionResNetV2 [43], ResNet50 [41], and MobileNetV2 [42]). Using a pretrained model to extract the features of remote sensing images can solve, to a certain extent, the problems involved with training a network for remote sensing image scene classification when there is a lack of training data. In this study, we used InceptionResNetV2, MobileNetV2, and ResNet50 as the pretrained deep learning models for transfer learning and performance comparisons with the proposed DS-SoybeanNet model.
(1)
ResNet50: The ResNet50 network contains 49 convolutional layers and a fully connected layer. The core CNN components are the convolutional filter and the pooling layer. ResNet50 is a CNN derivative with a core component skip-connection to circumvent the gradient disappearance problem. The ResNet structure can accelerate training and improve performance (preventing gradient dispersion).
(2)
InceptionResNetV2: The Inception module can obtain sparse or nonsparse features in the same layer. InceptionResNetV2 performs very well, but compared with ResNet, InceptionResNetV2 has a more complex network structure.
(3)
MobileNetV2: MobileNetV2 is a lightweight CNN model proposed by Google for embedded devices, such as mobile phones, with a focus on optimizing latency while considering the model’s size. MobileNetV2 can effectively balance latency and accuracy.
Transfer learning requires a low learning rate for retraining because the feature extraction module of the model already has some ability to extract image feature information after pretraining. An ideal learning rate can promote model convergence, whereas an unsuitable rate can cause training oscillations or even directly lead to the “explosion” of the loss value of the objective function. In addition to transfer learning methods based on InceptionResNetV2, MobileNetV2, and ResNet50, we also tested the performance of the AlexNet [48] and VGG16 [38] models to monitor soybean maturity.

3.3. SVM and RF

We also compared the soybean maturity information classification accuracy of our proposed DS-SoybeanNet with those of conventional machine learning models (SVM and RF). SVM is a generalized linear classifier that performs binary data classification in supervised learning [49]. Its decision boundary is the maximum marginal hyperplane solved for the learned samples, which reduces the classification problem to a convex quadratic programming problem. SVM has a low composition risk, its training is challenging to implement on large samples, and it is not ideal for solving multiclassification problems. RF is based on an integrated learning strategy, which combines multiple decision trees [50]. These decision trees are independent and unrelated to each other. Random forest uses the bagging strategy and repeated sampling to generate multiple trees. Under the bagging and bootstrap aggregation strategy, a subset of the samples are randomly selected from the dataset for training, and voting is conducted to obtain the average value as the resulting output. This strategy significantly avoids incorrect sample data, and thus shows improved accuracy.

3.4. Accuracy Evaluation

Figure 4 shows the experimental methodology used in this work. The canopy images of field F1 were used to calibrate and validate the models, whereas all canopy images of field F2 were used to validate the models.
The confusion matrix is a widely used tool for model accuracy evaluations. Table 4 shows the confusion matrix for the binary classification problem. Accuracy and recall can be obtained based on the confusion matrix. Generally, a higher accuracy and recall indicate a higher classification accuracy.
Accuracy = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
TP, TN, FP, and FN represent the true-positive, true-negative, false-positive, and false-negative categories, respectively, in the confusion matrix (Table 4). Confusion matrices are not limited to binary classification but can also be used for multiclass classification. In this study, we used the confusion matrix, accuracy, and recall to evaluate the soybean maturity classification accuracy of the proposed DS-SoybeanNet model.

4. Results and Discussion

4.1. Model Calibration and Validation Based on Field F1

We used the calibration dataset of field F1 to train the proposed DS-SoybeanNet, AlexNet, VGG16, InceptionResNetV2, MobileNetV2, ResNet50, SVM, and RF models. Each model was trained and validated three times, and the model with the highest performance was saved. The learning rates were set to 0.0005, 0.0001, and 0.00001 for the transfer learning models (InceptionResNetV2, MobileNetV2, and ResNet50), and the number of epochs was set to 100. For DS-SoybeanNet, we analyzed the model accuracy with different convolution window sizes.

4.1.1. Validation of AlexNet, VGG16, SVM, and RF

We tested the SVM and RF models for monitoring soybean breeding line maturity (Table 5) based on the validation dataset from field F1. The L0, L1, and L3 classification recall values were higher than 99% for the traditional machine learning models (SVM and RF). The classification accuracies of SVM and RF were 92.31% and 94.23%, respectively. We also tested the performance of the AlexNet and VGG16 models (Table 5). The performances of AlexNet (99.44%) and VGG16 (97.99%) were higher than those of SVM (92.31%) and RF (94.23%).

4.1.2. Validation of Transfer Learning Based on InceptionResNetV2, MobileNetV2, and ResNet50

We also tested the performance of the three deep learning models using three learning rates. Table 6 shows the accuracies of the models using different learning rates. The performances of the three deep learning models (InceptionResNetV2, MobileNetV2, and ResNet50) were similar when using different learning rates. Our results indicate that the soybean maturity classification accuracy of traditional machine learning models (RF: 94.23%; SVM: 92.31%) was lower than that of InceptionResNetV2, MobileNetV2, and ResNet50.
There were notable differences in recall among the four labels. For example, the L2 classification recall of InceptionResNetV2 was much lower than those of L0, L1, and L3 when the learning rate was 0.0005. The same was observed for MobileNetV2 and ResNet50, which had L2 classification recalls of 69.23% and 88.46%, respectively.

4.1.3. Validation of the Proposed DS-SoybeanNet Model

We tested the proposed DS-SoybeanNet model in the monitoring of soybean breeding line maturity. Table 7 shows the classification results of the DS-SoybeanNet model with the convolution kernel size set to 3 × 3, 5 × 5, 7 × 7, 9 × 9, 11 × 11, 16 × 16, and 21 × 21. The results indicate that there was little difference in performance among the seven convolution kernel sizes (with classification accuracies ranging from 97.52% to 99.19%). The results suggest that the model had the best soybean maturity classification accuracy when the convolution kernel size was set to 5 × 5 (99.17%) or 7 × 7 (99.19%). Figure 5 shows the training accuracy and loss curves of the DS-SoybeanNet with kernel sizes of 5 × 5 and 7 × 7. These results indicate that the model reached convergence at about 40 epochs. Training the DS-SoybeanNet (5 × 5) for about 100 epochs could take about 40 min and 5 s. Table A1 and Table A2 show the model architecture and parameter information of DS-SoybeanNet with 5 × 5 and 7 × 7 kernels.

4.2. Performance Comparison Based on Field F2

We used the 546 images from field F2 to test the performance of MobileNetV2, InceptionResNetV2, ResNet50, SVM, RF, and the proposed DS-SoybeanNet model in monitoring soybean maturity. Table 8 shows the confusion matrices of the soybean maturity classifications of the eight models. Table 9 shows the classification results of the eight models using the data from field F2. Our results (Table 8 and Table 9) indicated that the proposed DS-SoybeanNet model exhibited a higher classification accuracy than the other machine learning models.
The conventional machine learning models (SVM and RF) exhibited the highest classification recall (100%) in the classification of immature soybeans (L0) (see Table 9). AlexNet (96.95%) showed the highest classification recall for mature soybeans (L2). As shown in Table 8 and Table 9, the conventional machine learning models (SVM and RF) and deep learning models (MobileNetV2, InceptionResNetV2, and ResNet50) showed lower recalls for near-mature soybeans (L1), which led to lower overall classification accuracies for these models. DS-SoybeanNet (84.47%) had the highest classification recall for near-mature soybeans (L1) (see Table 9).
As shown in Table 9, the ResNet50 model exhibited a high classification accuracy of 72.16%. The RF (65.38%) and SVM (63.37%) models had similar classification accuracies. The soybean classification accuracies of InceptionResNetV2 (55.86%) and MobileNetV2 (52.75%) were lower than those of the other five models. The accuracies of DS-SoybeanNet based on 5 × 5 and 7 × 7 convolution kernels, namely, 86.26% and 84.25%, respectively, were notably higher than those of the other models.
Note that the eight models’ performance decreased when using the field F2 dataset to test the models (Table 5, Table 6, Table 7 and Table 9). As shown in Table 9, the top 3 models were DS-SoybeanNet, AlexNet, and VGG16 when monitoring soybean maturity using the field F2 dataset. Recently, the AlexNet [48] and VGG16 [39] models have been used to detect crop maturity by many researchers. Our results show that the new DS-SoybeanNet model performed better than the AlexNet and VGG16 models in the classification of immature (L0) and near-mature soybeans (L1). For the field F1 dataset, the recall of L0 for DS-SoybeanNet was 100%, which is higher than that of AlexNet (99.69%) and VGG16 (98.74%). For the field F2 dataset, the recall of L0 and L1 for DS-SoybeanNet was 92.19% and 84.47%, which was notably higher than that of the AlexNet (L0: 79.37%, L1: 43.89%) model.
To further evaluate the fusion of deep and shallow CNN features and to explore the efficiency of the proposed DS-SoybeanNet model, we set up three ablation experiments for DS-SoybeanNet, as described below. Figure 6 shows the architectures of the CNNs used for experiments 2 and 3. Each model was trained and validated three times, and the model with the highest performance was saved.
  • Experiment 1. DS-SoybeanNet (Figure 3);
  • Experiment 2. DS-SoybeanNet with only shallow image features (Figure 6a); and
  • Experiment 3. DS-SoybeanNet with only deep image features (Figure 6b);
Our results (Table 10) indicate that the soybean maturity classification accuracy in experiment 2 (only shallow image features) and experiment 3 (only deep image features) was lower than that in experiment 1. This further proved that fusing deep and shallow CNN features [44,45,46] may improve the performance of the model in image classification tasks.

4.3. Soybean Maturity Mapping

For soybean maturity mapping, the following three steps were carried out:
(a)
A soybean canopy DOM of field F2 was obtained after the UAV flight and the image stitching process. Then, all soybean breeding line plots (26 rows and 21 columns) were manually labeled, and the soybean plot image coordinates (plot center) were recorded.
(b)
The soybean canopy images (108 × 108 × 3) were extracted automatically using the image coordinates and soybean canopy DOM using a Python script. Then, we used DS-SoybeanNet to classify these soybean canopy images.
(c)
We then mapped the soybean maturity based on the soybean maturity information and soybean plot image coordinates.
Figure 7 shows a true-color RGB image and the maturity maps calculated for field F2 using DS-SoybeanNet with 5 × 5 and 7 × 7 convolution kernels. Our results indicate that the estimated soybean maturity information for field F2 had a high accuracy. The soybean maturity information obtained from the DS-SoybeanNet model with 5 × 5 and 7 × 7 convolution kernels was similar.

4.4. Advantages and Disadvantages of UAV + DS-SoybeanNet

As soybeans mature, the leaf chlorophyll level gradually decreases, contributing to a slow change in the leaves’ color from green to yellow [51,52]. Crop leaf chlorophyll variation is asynchronous among layers of leaves [52]. For example, leaves in the top layer of a soybean canopy tend to have a younger leaf age and thus turn yellow later than the leaves in the bottom layer. Consequently, green and yellow leaves appear in the soybean canopy when the soybeans are nearly mature (Figure 2). Breeding fields commonly have thousands of breeding lines with different maturation times. Thus, timely monitoring of soybean breeding line maturity is crucial for soybean harvesting management and yield measurements [5,6,7,8]. UAV remote sensing technology can be utilized to collect high-resolution crop canopy images and has been widely used in precision agricultural crop trait monitoring [14,15]. Many studies have evaluated the crop parameter monitoring performance of digital cameras and multispectral sensors on board lightweight UAVs [17,18,19]. In our study, we attempted to evaluate the potential of using UAV remote sensing to monitor soybean breeding line maturity. We developed DS-SoybeanNet, which can extract and utilize both shallow and deep image features, and which thus helps to provide soybean breeding line maturity monitoring that is more robust than that offered by conventional machine learning methods. DS-SoybeanNet achieved the best accuracy of 86.26% (Table A1), which was notably higher than those of the conventional machine learning models (SVM and RF). However, DS-SoybeanNet has various disadvantages compared with conventional machine learning methods, such as its long elapsed time and large size (Table 11). In machine learning, CNNs have a more complex network structure and higher computational complexity than conventional machine learning models with larger model sizes.
Table 11 shows the time required to process 1000 samples using each model and the models’ sizes. The computation times of the CNN models (ranging from 6.607 s to 67.080 s) were notably higher than those of the conventional machine learning models, SVM and RF (0.003 s and 0.007 s). In addition, a high-performance device is required to calibrate CNN models. As shown in Table 11, the model sizes of DS-SoybeanNet, ResNet50, and InceptionResNetV2 were more than 300 MB. The proposed DS-SoybeanNet model had the largest size (2616 MB) compared to the other models. The DS-SoybeanNet model’s large size may mean that it requires large storage when deployed on lightweight platforms (e.g., Raspberry Pi) for stationary observations. Nevertheless, DS-SoybeanNet (5 × 5) had approximately the same calculation speed as MobileNetV2 and a much higher monitoring accuracy than the other deep learning models. Therefore, we consider DS-SoybeanNet a fast and high-performance deep-learning tool for monitoring soybean maturity.
Many previous studies have used AlexNet, VGG16, Inception-V3, and VGG19 in crop maturity classifications. Faisal et al. [53] compared the performance of pre-trained VGG-19 (99.4%), Inception-V3 (99.4%), and NASNet (99.7%) in detecting fruit maturity. Atif et al. [54] used AlexNet and VGG16 to classify the maturity levels of jujube fruits (best: VGG16 = 99.17%). Sahil et al. [55] developed a method that used YOLOv3 to pinpoint the locations of tomatoes (94.67%) and used an AlexNet-like CNN model to classify their maturity levels (90.67%). In this work, we compared the results of conventional machine learning models (SVM (92.31%) and RF (94.23%)) and six CNN machine learning models (DS-SoybeanNet (99.19%), VGG16 (97.99%), AlexNet (99.44%), ResNet50 (98.97%), InceptionResNetV2 (99.49%), and MobileNetV2 (97.52%)) in soybean maturity information monitoring based on UAV remote sensing. The accuracy results reported in this study were close to those of previous studies based on AlexNet, VGG16, Inception-V3, and VGG16. Thus, our results further proved that deep learning is a good tool for crop maturity information monitoring [48,53,54,55,56]. The combination of UAV remote sensing and deep learning can be used for high-performance soybean maturity information monitoring. However, our results indicate that selected machine learning models’ performance decreased when using the field F2 dataset to test the models (Table 5, Table 6, Table 7 and Table 9). We suspect that changes in the UAV’s working environment—for example, varying sunlight intensity over time—led to a direct decline in the models’ performance. This is perhaps not surprising because the farmland environment is affected by varying cropland conditions (e.g., irrigation, wind). Thus, future research should be focused on the factors influencing cropland images.
In this study, the performance obtained when using soybean canopy images captured by the UAV’s remote sensing digital camera may have been limited by the varying sunlight intensity over time. Since DS-SoybeanNet did not normalize the image differences due to sunlight, a normalization module may improve its performance in soybean maturity classification. Therefore, future studies need to develop a normalization module to weaken the effect of the sun. Thus, more experiments with different varieties and regions of soybeans are needed to improve the generalizability of the DS-SoybeanNet model. In this study, the proposed DS-SoybeanNet was validated using only two breeding fields from a single site; thus, further validation is required from additional fields and study sites.

5. Conclusions

In this study, we designed a network, namely, DS-SoybeanNet, to extract and utilize both shallow and deep image features to improve the performance of UAV-based soybean maturity information monitoring. We compared conventional machine learning methods (SVM and RF), current deep learning methods (AlexNet, VGG16, InceptionResNetV2, MobileNetV2, and ResNet50), and our proposed DS-SoybeanNet model in terms of their soybean maturity classification accuracy. The results were as follows.
(1)
The conventional machine learning methods (SVM and RF) had lower calculation times than the deep learning methods (AlexNet, VGG16, InceptionResNetV2, MobileNetV2, and ResNet50) and our proposed DS-SoybeanNet model. For example, the computation speed of RF was 0.03 s per 1000 images. However, the overall accuracies of the conventional machine learning methods were notably lower than those of the deep learning methods and the proposed DS-SoybeanNet model.
(2)
The current deep learning methods were outperformed in terms of universality by the DS-SoybeanNet model in the monitoring of soybean maturity. The overall accuracies of MobileNetV2 for fields F1 and F2 were 97.52% and 52.75%, respectively.
(3)
The proposed DS-SoybeanNet model was able to provide high-performance soybean maturity classification results. Its computation speed was 11.770 s per 1000 images and its overall accuracies for fields F1 and F2 were 99.19% and 86.26%, respectively.
(4)
Furthermore, future studies are needed in order to develop a normalization module to weaken the effect of the sun. Moreover, further validation is required using additional fields and study sites.

Author Contributions

J.Y., H.F. (Haikuan Feng), S.Z. and H.F. (Hao Feng) designed the experiments. J.Y., H.F. (Haikuan Feng), Z.S., H.X. and C.Z. collected the soybean images. J.Y. and S.Z. analyzed the data and wrote the manuscript. Y.L. and S.H. made comments and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NO. 42101362).

Data Availability Statement

The data that support the findings of this study are available from the corresponding author, J.Y., upon reasonable request.

Acknowledgments

We thank Bo Xu and Guozheng Lu for field management and data collection.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1 shows the attention regions of different models in the soybean canopy images. Regarding interpretability, the top three models performed differently when their attention regions were visualized by means of the Grad-CAM technique (Figure A1). VGG16 models focused only on luxuriant leaves for all four categories (Figure A1). The AlexNet model showed acceptable attention regions when dealing with L0 and L1 soybean images, whereas it focused only on branches and leaves when analyzing L2 and L3 soybean images (Figure A1). Compared with AlexNet and VGG16 models, DS-SoybeanNet showed acceptable attention regions for the four categories (Figure A1). In most cases, DS-SoybeanNet was able to differentiate among the soybean images accurately based on the leaves, branches, and soil pixels, similarly to farm workers. Table A1 and Table A2 show the model architecture and parameter information of DS-SoybeanNet with 5 × 5 and 7 × 7 kernels.
Figure A1. The attention regions of the top 3 (Table 9) models in soybean canopy images.
Figure A1. The attention regions of the top 3 (Table 9) models in soybean canopy images.
Agriculture 13 00110 g0a1
Table A1. Details of the proposed DS-SoybeanNet with 5 × 5 kernels.
Table A1. Details of the proposed DS-SoybeanNet with 5 × 5 kernels.
Layer (Type)Output ShapeParamConnected to
input_1 (Input Layer)[(None,108,108,3)0
conv2d (Conv2D)(None,108,108,32)2432input_1 [0][0]
conv2d_1 (Conv2D)(None,108,108,16)12816conv2d [0][0]
max_pooling2d_1 (MaxPooling2D)(None,27,27,16)0conv2d_1 [0][0]
conv2d_2 (Conv2D)(None,27,27,32)12832max_pooling2d_1 [0][0]
conv2d_3 (Conv2D)(None,27,27,16)12816conv2d_2 [0][0]
max_pooling2d (MaxPooling2D)(None,27,27,32)0conv2d [0][0]
max_pooling2d_2 (MaxPooling2D)(None,13,13,32)0conv2d_2 [0][0]
max_pooling2d_3 (MaxPooling2D)(None, 13,13,16)0conv2d_3 [0][0]
conv2d_4 (Conv2D)(None,27,27,16)6416conv2d_3 [0][0]
flatten (Flatten)(None,23328)0max_pooling2d [0][0]
flatten_1 (Flatten)(None,11664)0max_pooling2d_1 [0][0]
flatten_2 (Flatten)(None,5408)0max_pooling2d_2 [0][0]
flatten_3 (Flatten)(None,2704) 0max_pooling2d_3 [0][0]
flatten_4 (Flatten)(None,11664)0conv2d_4 [0][0]
concatenate (Concatenate)(None,54768)0flatten [0][0]
flatten_1 [0][0]
flatten_2 [0][0]
flatten _3 [0][0]
flatten _4 [0][0]
dropout (Dropout)(None,54768)0concatenate [0][0]
dense (Dense)(None,4096)224333824dropout [0][0]
dropout_1 (Dropout)(None,4096)0dense [0][0]
dense_1 (Dense)(None,512)4195328dropout_1 [0][0]
dropout_2 (Dropout)(None,512)0dense_1 [0][0]
dense_2 (Dense)(None,4)4100dropout_2 [0][0]
Total params: 228,580,564
Trainable params: 228,580,564
Non-trainable params: 0
Table A2. Details of the proposed DS-SoybeanNet with 7 × 7 kernels.
Table A2. Details of the proposed DS-SoybeanNet with 7 × 7 kernels.
Layer (Type)Output ShapeParamConnected to
input_1 (Input Layer)[(None,108,108,3)0
conv2d (Conv2D)(None,108,108,32)4736input_1 [0][0]
conv2d_1 (Conv2D)(None,108,108,16)25104conv2d [0][0]
max_pooling2d_1 (MaxPooling2D)(None,27,27,16)0conv2d_1 [0][0]
conv2d_2 (Conv2D)(None,27,27,32)25120max_pooling2d_1 [0][0]
conv2d_3 (Conv2D)(None,27,27,16)25104conv2d_2 [0][0]
max_pooling2d (MaxPooling2D)(None,27,27,32)0conv2d [0][0]
max_pooling2d_2 (MaxPooling2D)(None,13,13,32)0conv2d_2 [0][0]
max_pooling2d_3 (MaxPooling2D)(None, 13,13,16)0conv2d_3 [0][0]
conv2d_4 (Conv2D)(None,27,27,16)12560conv2d_3 [0][0]
flatten (Flatten)(None,23328)0max_pooling2d [0][0]
flatten_1 (Flatten)(None,11664)0max_pooling2d_1 [0][0]
flatten_2 (Flatten)(None,5408)0max_pooling2d_2 [0][0]
flatten_3 (Flatten)(None,2704) 0max_pooling2d_3 [0][0]
flatten_4 (Flatten)(None,11664)0conv2d_4 [0][0]
concatenate (Concatenate)(None,54768)0flatten [0][0]
flatten_1 [0][0]
flatten_2 [0][0]
flatten _3 [0][0]
flatten _4 [0][0]
dropout (Dropout)(None,54768)0concatenate [0][0]
dense (Dense)(None,4096)224333824dropout [0][0]
dropout_1 (Dropout)(None,4096)0dense [0][0]
dense_1 (Dense)(None,512)4195328dropout_1 [0][0]
dropout_2 (Dropout)(None,512)0dense_1 [0][0]
dense_2 (Dense)(None,4)4100dropout_2 [0][0]
Total params: 228,625,876
Trainable params: 228,625,876
Non-trainable params: 0

References

  1. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Hartling, S.; Esposito, F.; Fritschi, F.B. Soybean Yield Prediction from UAV Using Multimodal Data Fusion and Deep Learning. Remote Sens. Environ. 2020, 237, 111599. [Google Scholar] [CrossRef]
  2. Qin, P.; Wang, T.; Luo, Y. A Review on Plant-Based Proteins from Soybean: Health Benefits and Soy Product Development. J. Agric. Food Res. 2022, 7, 100265. [Google Scholar] [CrossRef]
  3. Liu, X.; Jin, J.; Wang, G.; Herbert, S.J. Soybean Yield Physiology and Development of High-Yielding Practices in Northeast China. Field Crop. Res. 2008, 105, 157–171. [Google Scholar] [CrossRef]
  4. Zhang, Y.M.; Li, Y.; Chen, W.F.; Wang, E.T.; Tian, C.F.; Li, Q.Q.; Zhang, Y.Z.; Sui, X.H.; Chen, W.X. Biodiversity and Biogeography of Rhizobia Associated with Soybean Plants Grown in the North China Plain. Appl. Environ. Microbiol. 2011, 77, 6331–6342. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Vogel, J.T.; Liu, W.; Olhoft, P.; Crafts-Brandner, S.J.; Pennycooke, J.C.; Christiansen, N. Soybean Yield Formation Physiology—A Foundation for Precision Breeding Based Improvement. Front. Plant Sci. 2021, 12, 719706. [Google Scholar] [CrossRef]
  6. Maranna, S.; Nataraj, V.; Kumawat, G.; Chandra, S.; Rajesh, V.; Ramteke, R.; Patel, R.M.; Ratnaparkhe, M.B.; Husain, S.M.; Gupta, S.; et al. Breeding for Higher Yield, Early Maturity, Wider Adaptability and Waterlogging Tolerance in Soybean (Glycine max L.): A Case Study. Sci. Rep. 2021, 11, 22853. [Google Scholar] [CrossRef]
  7. Volpato, L.; Dobbels, A.; Borem, A.; Lorenz, A.J. Optimization of Temporal UAS-Based Imagery Analysis to Estimate Plant Maturity Date for Soybean Breeding. Plant Phenome J. 2021, 4, e20018. [Google Scholar] [CrossRef]
  8. Moeinizade, S.; Pham, H.; Han, Y.; Dobbels, A.; Hu, G. An Applied Deep Learning Approach for Estimating Soybean Relative Maturity from UAV Imagery to Aid Plant Breeding Decisions. Mach. Learn. Appl. 2022, 7, 100233. [Google Scholar] [CrossRef]
  9. Zhou, J.; Mou, H.; Zhou, J.; Ali, M.L.; Ye, H.; Chen, P.; Nguyen, H.T. Qualification of Soybean Responses to Flooding Stress Using UAV-Based Imagery and Deep Learning. Plant Phenomics 2021, 2021. [Google Scholar] [CrossRef]
  10. Habibi, L.N.; Watanabe, T.; Matsui, T.; Tanaka, T.S.T. Machine Learning Techniques to Predict Soybean Plant Density Using UAV and Satellite-Based Remote Sensing. Remote Sens. 2021, 13, 2548. [Google Scholar] [CrossRef]
  11. Luo, S.; Liu, W.; Zhang, Y.; Wang, C.; Xi, X.; Nie, S.; Ma, D.; Lin, Y.; Zhou, G. Maize and Soybean Heights Estimation from Unmanned Aerial Vehicle (UAV) LiDAR Data. Comput. Electron. Agric. 2021, 182, 106005. [Google Scholar] [CrossRef]
  12. Fukano, Y.; Guo, W.; Aoki, N.; Ootsuka, S.; Noshita, K.; Uchida, K.; Kato, Y.; Sasaki, K.; Kamikawa, S.; Kubota, H. GIS-Based Analysis for UAV-Supported Field Experiments Reveals Soybean Traits Associated with Rotational Benefit. Front. Plant Sci. 2021, 12, 637694. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, G.; Li, C.; Wang, Y.; Yuan, H.; Feng, H.; Xu, B.; Yang, X. The DOM Generation and Precise Radiometric Calibration of a UAV-Mounted Miniature Snapshot Hyperspectral Imager. Remote Sens. 2017, 9, 642. [Google Scholar] [CrossRef] [Green Version]
  14. Zhou, C.; Ye, H.; Sun, D.; Yue, J.; Yang, G.; Hu, J. An Automated, High-Performance Approach for Detecting and Characterizing Broccoli Based on UAV Remote-Sensing and Transformers: A Case Study from Haining, China. Int. J. Appl. Earth Obs. Geoinf. 2022, 114, 103055. [Google Scholar] [CrossRef]
  15. Yue, J.; Yang, G.; Tian, Q.; Feng, H.; Xu, K.; Zhou, C. Estimate of Winter-Wheat above-Ground Biomass Based on UAV Ultrahigh-Ground-Resolution Image Textures and Vegetation Indices. ISPRS J. Photogramm. Remote Sens. 2019, 150, 226–244. [Google Scholar] [CrossRef]
  16. Haghighattalab, A.; González Pérez, L.; Mondal, S.; Singh, D.; Schinstock, D.; Rutkoski, J.; Ortiz-Monasterio, I.; Singh, R.P.; Goodin, D.; Poland, J. Application of Unmanned Aerial Systems for High Throughput Phenotyping of Large Wheat Breeding Nurseries. Plant Methods 2016, 12, 35. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Singhal, G.; Bansod, B.; Mathew, L.; Goswami, J.; Choudhury, B.U.; Raju, P.L.N. Chlorophyll Estimation Using Multi-Spectral Unmanned Aerial System Based on Machine Learning Techniques. Remote Sens. Appl. Soc. Environ. 2019, 15, 100235. [Google Scholar] [CrossRef]
  18. Roosjen, P.P.J.; Brede, B.; Suomalainen, J.M.; Bartholomeus, H.M.; Kooistra, L.; Clevers, J.G.P.W. Improved Estimation of Leaf Area Index and Leaf Chlorophyll Content of a Potato Crop Using Multi-Angle Spectral Data—Potential of Unmanned Aerial Vehicle Imagery. Int. J. Appl. Earth Obs. Geoinf. 2018, 66, 14–26. [Google Scholar] [CrossRef]
  19. Yue, J.; Feng, H.; Tian, Q.; Zhou, C. A Robust Spectral Angle Index for Remotely Assessing Soybean Canopy Chlorophyll Content in Different Growing Stages. Plant Methods 2020, 16, 104. [Google Scholar] [CrossRef]
  20. Wang, W.; Gao, X.; Cheng, Y.; Ren, Y.; Zhang, Z.; Wang, R.; Cao, J.; Geng, H. QTL Mapping of Leaf Area Index and Chlorophyll Content Based on UAV Remote Sensing in Wheat. Agriculture 2022, 12, 595. [Google Scholar] [CrossRef]
  21. Wójcik-Gront, E.; Gozdowski, D.; Stępień, W. UAV-Derived Spectral Indices for the Evaluation of the Condition of Rye in Long-Term Field Experiments. Agriculture 2022, 12, 1671. [Google Scholar] [CrossRef]
  22. Yue, J.; Feng, H.; Li, Z.; Zhou, C.; Xu, K. Mapping Winter-Wheat Biomass and Grain Yield Based on a Crop Model and UAV Remote Sensing. Int. J. Remote Sens. 2021, 42, 1577–1601. [Google Scholar] [CrossRef]
  23. Han, L.; Yang, G.; Yang, H.; Xu, B.; Li, Z.; Yang, X. Clustering Field-Based Maize Phenotyping of Plant-Height Growth and Canopy Spectral Dynamics Using a UAV Remote-Sensing Approach. Front. Plant Sci. 2018, 9, 1638. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Ofer, D.; Brandes, N.; Linial, M. The Language of Proteins: NLP, Machine Learning & Protein Sequences. Comput. Struct. Biotechnol. J. 2021, 19, 1750–1758. [Google Scholar] [CrossRef] [PubMed]
  25. Janiesch, C.; Zschech, P.; Heinrich, K. Machine Learning and Deep Learning. Electron. Mark. 2021, 31, 685–695. [Google Scholar] [CrossRef]
  26. Zhang, H.; Wang, Z.; Guo, Y.; Ma, Y.; Cao, W.; Chen, D.; Yang, S.; Gao, R. Weed Detection in Peanut Fields Based on Machine Vision. Agriculture 2022, 12, 1541. [Google Scholar] [CrossRef]
  27. Yue, J.; Feng, H.; Yang, G.; Li, Z. A Comparison of Regression Techniques for Estimation of Above-Ground Winter Wheat Biomass Using Near-Surface Spectroscopy. Remote Sens. 2018, 10, 66. [Google Scholar] [CrossRef] [Green Version]
  28. Niedbała, G.; Kurasiak-Popowska, D.; Piekutowska, M.; Wojciechowski, T.; Kwiatek, M.; Nawracała, J. Application of Artificial Neural Network Sensitivity Analysis to Identify Key Determinants of Harvesting Date and Yield of Soybean (Glycine max [L.] Merrill) Cultivar Augusta. Agriculture 2022, 12, 754. [Google Scholar] [CrossRef]
  29. Santos, L.B.; Bastos, L.M.; de Oliveira, M.F.; Soares, P.L.M.; Ciampitti, I.A.; da Silva, R.P. Identifying Nematode Damage on Soybean through Remote Sensing and Machine Learning Techniques. Agronomy 2022, 12, 2404. [Google Scholar] [CrossRef]
  30. Eugenio, F.C.; Grohs, M.; Venancio, L.P.; Schuh, M.; Bottega, E.L.; Ruoso, R.; Schons, C.; Mallmann, C.L.; Badin, T.L.; Fernandes, P. Estimation of Soybean Yield from Machine Learning Techniques and Multispectral RPAS Imagery. Remote Sens. Appl. Soc. Environ. 2020, 20, 100397. [Google Scholar] [CrossRef]
  31. Teodoro, P.E.; Teodoro, L.P.R.; Baio, F.H.R.; da Silva Junior, C.A.; Dos Santos, R.G.; Ramos, A.P.M.; Pinheiro, M.M.F.; Osco, L.P.; Gonçalves, W.N.; Carneiro, A.M.; et al. Predicting Days to Maturity, Plant Height, and Grain Yield in Soybean: A Machine and Deep Learning Approach Using Multispectral Data. Remote Sens. 2021, 13, 4632. [Google Scholar] [CrossRef]
  32. Abdelbaki, A.; Schlerf, M.; Retzlaff, R.; Machwitz, M.; Verrelst, J.; Udelhoven, T. Comparison of Crop Trait Retrieval Strategies Using UAV-Based VNIR Hyperspectral Imaging. Remote Sens. 2021, 13, 1748. [Google Scholar] [CrossRef] [PubMed]
  33. Sun, J.; Di, L.; Sun, Z.; Shen, Y.; Lai, Z. County-Level Soybean Yield Prediction Using Deep CNN-LSTM Model. Sensors 2019, 19, 4363. [Google Scholar] [CrossRef] [Green Version]
  34. Wang, J.; Si, H.; Gao, Z.; Shi, L. Winter Wheat Yield Prediction Using an LSTM Model from MODIS LAI Products. Agriculture 2022, 12, 1707. [Google Scholar] [CrossRef]
  35. Tian, H.; Wang, P.; Tansey, K.; Han, D.; Zhang, J.; Zhang, S.; Li, H. A Deep Learning Framework under Attention Mechanism for Wheat Yield Estimation Using Remotely Sensed Indices in the Guanzhong Plain, PR China. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102375. [Google Scholar] [CrossRef]
  36. Khaki, S.; Wang, L. Crop Yield Prediction Using Deep Neural Networks. Front. Plant Sci. 2019, 10, 621. [Google Scholar] [CrossRef] [Green Version]
  37. Khan, A.I.; Quadri, S.M.K.; Banday, S.; Latief Shah, J. Deep Diagnosis: A Real-Time Apple Leaf Disease Detection System Based on Deep Learning. Comput. Electron. Agric. 2022, 198, 107093. [Google Scholar] [CrossRef]
  38. Albarrak, K.; Gulzar, Y.; Hamid, Y.; Mehmood, A.; Soomro, A.B. A Deep Learning-Based Model for Date Fruit Classification. Sustainability 2022, 14, 6339. [Google Scholar] [CrossRef]
  39. Gulzar, Y.; Hamid, Y.; Soomro, A.B.; Alwan, A.A.; Journaux, L. A Convolution Neural Network-Based Seed Classification System. Symmetry 2020, 12, 2018. [Google Scholar] [CrossRef]
  40. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  41. Sangeetha, V.; Prasad, K.J.R. Syntheses of Novel Derivatives of 2-Acetylfuro[2,3-a]Carbazoles, Benzo[1,2-b]-1,4-Thiazepino[2,3-a]Carbazoles and 1-Acetyloxycarbazole-2- Carbaldehydes. Indian J. Chem. Sect. B Org. Med. Chem. 2006, 45, 1951–1954. [Google Scholar] [CrossRef]
  42. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  43. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence (AAAI-17), San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar] [CrossRef]
  44. Miao, Y.; Lin, Z.; Ding, G.; Han, J. Shallow Feature Based Dense Attention Network for Crowd Counting. In Proceedings of the 34th AAAI Conference on Artificial Intelligence (AAAI-20), New York, NY, USA, 7–12 February 2020; pp. 11765–11772. [Google Scholar] [CrossRef]
  45. Wei, J.; Wang, Q.; Li, Z.; Wang, S.; Zhou, S.K.; Cui, S. Shallow Feature Matters for Weakly Supervised Object Localization. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2021, 1, 5989–5997. [Google Scholar] [CrossRef]
  46. Bougourzi, F.; Dornaika, F.; Mokrani, K.; Taleb-Ahmed, A.; Ruichek, Y. Fusing Transformed Deep and Shallow Features (FTDS) for Image-Based Facial Expression Recognition. Expert Syst. Appl. 2020, 156, 113459. [Google Scholar] [CrossRef]
  47. Lecun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  48. Behera, S.K.; Rath, A.K.; Sethy, P.K. Maturity Status Classification of Papaya Fruits Based on Machine Learning and Transfer Learning Approach. Inf. Process. Agric. 2021, 8, 244–250. [Google Scholar] [CrossRef]
  49. Hosseini, M.; McNairn, H.; Mitchell, S.; Robertson, L.D.; Davidson, A.; Ahmadian, N.; Bhattacharya, A.; Borg, E.; Conrad, C.; Dabrowska-Zielinska, K.; et al. A Comparison between Support Vector Machine and Water Cloud Model for Estimating Crop Leaf Area Index. Remote Sens. 2021, 13, 1348. [Google Scholar] [CrossRef]
  50. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  51. Huang, W.; Wang, Z.; Huang, L.; Lamb, D.W.; Ma, Z.; Zhang, J.; Wang, J.; Zhao, C. Estimation of Vertical Distribution of Chlorophyll Concentration by Bi-Directional Canopy Reflectance Spectra in Winter Wheat. Precis. Agric. 2011, 12, 165–178. [Google Scholar] [CrossRef]
  52. Wang, J.; Zhao, C.; Huang, W. Fundamental and Application of Quantitative Remote Sensing in Agriculture; Science China Press: Beijing, China, 2008. [Google Scholar]
  53. Faisal, M.; Alsulaiman, M.; Arafah, M.; Mekhtiche, M.A. IHDS: Intelligent Harvesting Decision System for Date Fruit Based on Maturity Stage Using Deep Learning and Computer Vision. IEEE Access 2020, 8, 167985–167997. [Google Scholar] [CrossRef]
  54. Mahmood, A.; Singh, S.K.; Tiwari, A.K. Pre-Trained Deep Learning-Based Classification of Jujube Fruits According to Their Maturity Level. Neural Comput. Appl. 2022, 34, 13925–13935. [Google Scholar] [CrossRef]
  55. Mutha, S.A.; Shah, A.M.; Ahmed, M.Z. Maturity Detection of Tomatoes Using Deep Learning. SN Comput. Sci. 2021, 2, 441. [Google Scholar] [CrossRef]
  56. Zhou, X.; Lee, W.S.; Ampatzidis, Y.; Chen, Y.; Peres, N.; Fraisse, C. Strawberry Maturity Classification from UAV and Near-Ground Imaging Using Deep Learning. Smart Agric. Technol. 2021, 1, 100001. [Google Scholar] [CrossRef]
Figure 1. Study area (a) and experimental soybean field (b).
Figure 1. Study area (a) and experimental soybean field (b).
Agriculture 13 00110 g001
Figure 2. Examples of the four labels.
Figure 2. Examples of the four labels.
Agriculture 13 00110 g002
Figure 3. Architecture of DS-SoybeanNet.
Figure 3. Architecture of DS-SoybeanNet.
Agriculture 13 00110 g003
Figure 4. Flowchart of the experimental methodology.
Figure 4. Flowchart of the experimental methodology.
Agriculture 13 00110 g004
Figure 5. Training accuracy (a) and loss (b) of the DS-SoybeanNet with kernel sizes of 5 × 5 and 7 × 7.
Figure 5. Training accuracy (a) and loss (b) of the DS-SoybeanNet with kernel sizes of 5 × 5 and 7 × 7.
Agriculture 13 00110 g005
Figure 6. Architecture of CNNs used for experiments 2 (a) and 3 (b).
Figure 6. Architecture of CNNs used for experiments 2 (a) and 3 (b).
Agriculture 13 00110 g006aAgriculture 13 00110 g006b
Figure 7. Maturity maps. (a) RGB true-color image; (b) DS-SoybeanNet (5 × 5); and (c) DS-SoybeanNet (7 × 7). Note: The red rectangle indicates the soybean plot region.
Figure 7. Maturity maps. (a) RGB true-color image; (b) DS-SoybeanNet (5 × 5); and (c) DS-SoybeanNet (7 × 7). Note: The red rectangle indicates the soybean plot region.
Agriculture 13 00110 g007aAgriculture 13 00110 g007b
Table 1. Parameters of the UAV and digital camera used in this study.
Table 1. Parameters of the UAV and digital camera used in this study.
UAVParameterCameraParameter
UAV nameDJI S1000Camera nameSONY DSC-QX100
Flight heightApproximately 50 mImage size5472 × 3648
Flight speedApproximately 8 m/sImage dpi350
Flight time>20 minAperturef/11
Exposure1/1250 s
ISOISO-1600
Focal length10 mm
ChannelsRed, green, blue
Ground spatial resolution0.016 m
Table 2. Standards used for labeling the soybean plots.
Table 2. Standards used for labeling the soybean plots.
LabelPriorityDescription
L0ImmatureLowAll upper canopy leaves are green or there are a few yellow leaves.
L1Near-matureHighApproximately half of the upper canopy leaves are yellow.
L2MatureHighestThe upper leaves of the canopy are yellow but have yet to be harvested.
L3HarvestedLowThe soybean planting area has been harvested.
Table 3. Numbers of soybean images for model calibration and validation.
Table 3. Numbers of soybean images for model calibration and validation.
LabelTraining Dataset
(Field F1)
Validation Dataset
(Field F1)
Independent Validation Dataset
(Field F2)
L054231864
L1257163219
L27052198
L340031465
Total1269847546
Enhancement25,38016,940-
Table 4. Confusion matrix.
Table 4. Confusion matrix.
TypePredicted condition
Actual
condition
LabelPositive (P)Negative (N)
Positive (T)True Positive (TP)False Negative (FN)
Negative (N)False Positive (FP)True Negative (TN)
Table 5. Classification results of AlexNet, VGG16, SVM, and RF.
Table 5. Classification results of AlexNet, VGG16, SVM, and RF.
LabelSVMRFAlexNetVGG16
L099.69%99.06%99.69%98.74%
L1100%100%99.39%100%
L290.38%90.38%98.08%84.62%
L399.04%99.36%99.36%98.41%
Accuracy92.31%94.23%99.44% *97.99%
Note: * indicates the highest accuracy.
Table 6. Classification results of transfer learning based on InceptionResNetV2, MobileNetV2, and ResNet50.
Table 6. Classification results of transfer learning based on InceptionResNetV2, MobileNetV2, and ResNet50.
LabelInceptionResNetV2MobileNetV2ResNet50
Rate 1Rate 2Rate 3Rate 1Rate 2Rate 3Rate 1Rate 2Rate 3
L098.09%100%99.69%100%100%99.69%99.69%100%99.69%
L196.93%100%98.16%95.09%96.32%92.02%100%96.93%98.16%
L282.69%98.08%98.08%69.23%84.62%82.69%88.46%96.15%94.23%
L399.36%98.73%99.04%99.36%97.77%97.77%99.04%98.73%99.36%
Accuracy97.41%99.49%99.09%96.93%97.52%96.46%98.93%98.77%98.97%
Note: Rate 1 = 0.0005; Rate 2 = 0.0001; Rate 3 = 0.00001.
Table 7. Classification results of the proposed DS-SoybeanNet.
Table 7. Classification results of the proposed DS-SoybeanNet.
LabelDS-SoybeanNet
3 × 35 × 57 × 79 × 911 × 1116 × 1621 × 21
RecallL0100%100%100%100%100%100%100%
L196.93%100%100%100%99.39%99.39%99.39%
L292.31%90.38%90.38%78.85%88.46%80.77%75.00%
L399.36%99.36%99.68%99.36%99.68%96.50%97.77%
Accuracy98.70%99.17% *99.19% *98.47%99.06%97.40%97.52%
Note: * indicates the highest accuracy.
Table 8. Confusion matrices of MobileNetV2 (a), InceptionResNetV2 (b), ResNet50 (c), SVM (d), RF (e), DS-SoybeanNet with kernel sizes of 5 × 5 (f) and 7 × 7 (g), AlexNet (h), and VGG16 (i).
Table 8. Confusion matrices of MobileNetV2 (a), InceptionResNetV2 (b), ResNet50 (c), SVM (d), RF (e), DS-SoybeanNet with kernel sizes of 5 × 5 (f) and 7 × 7 (g), AlexNet (h), and VGG16 (i).
(a)Predicted Condition(b)Predicted Condition(c)Predicted Condition
Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3
L052048L0390178L0461611
L120016831L151818412L19971094
L2101725L2011934L2031914
L300164L3001055L300560
(d)Predicted Condition(e)Predicted Condition(f)Predicted Condition
Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3
L064000L064000L059500
L19811911L18994333L116185180
L22851029L214013720L20271710
L320261L300362L300956
(g)Predicted Condition (5 × 5)(h)Predicted Condition(i)Predicted Condition
Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3Actual conditionLabelL0L1L2L3
L059500L0511102L059500
L113179252L15971152L125169196
L20221697L2031923L20281637
L3001253L300560L300758
Table 9. Classification results of eight models from field F2.
Table 9. Classification results of eight models from field F2.
ModelRankPrecisionAccuracy
L0L1L2L3
DS-SoybeanNet (5 × 5)192.19%84.47%86.36%86.15%86.26%
DS-SoybeanNet (7 × 7)92.19%81.74%85.35%81.54%84.25%
VGG16292.19%77.17%82.32%89.23%82.23%
AlexNet379.37%43.89%96.95%92.31%72.89%
ResNet50471.87%44.29%96.46%92.31%72.16%
RF5100%42.92%69.19%95.38%65.38%
SVM6100%54.34%51.52%93.85%63.37%
InceptionResNetV2760.93%8.22%97.47%84.62%55.86%
MobileNetV2881.25%0%39.53%98.46%52.75%
Table 10. Classification results of three experiments with 5 × 5 and 7 × 7 kernels.
Table 10. Classification results of three experiments with 5 × 5 and 7 × 7 kernels.
LabelExperiment 1Experiment 2Experiment 3
Validation
Dataset
(Field F1)
Independent
Validation
Dataset
(Field F2)
Validation
Dataset
(Field F1)
Independent
Validation
Dataset
(Field F2)
Validation
Dataset
(Field F1)
Independent
Validation
Dataset
(Field F2)
RecallL0100%92.19%100%98.44%100%96.88%
L1100%84.47%100%74.89%99.39%83.11%
L290.38%86.36%84.62%87.37%69.23%71.21%
L399.36%86.15%98.09%75.38%98.09%87.69%
Accuracy99.17% *86.26% *98.35%82.23%97.28%80.95%
RecallL0100%92.19%100%75.00%100%89.06%
L1100%81.74%99.39%78.08%99.39%81.74%
L290.38%85.35%86.54%82.83%78.85%75.25%
L399.68%81.54%98.73%92.31%98.41%81.54%
Accuracy99.19% *84.25% *98.58%81.14%97.99%80.22%
Note: Bold and * indicate the highest accuracy.
Table 11. Models’ elapsed times and sizes.
Table 11. Models’ elapsed times and sizes.
ModelTime (s)/1000 SamplesSize
RF0.00324.1 KB
SVM0.0077.70 KB
MobileNetV26.60753.3 MB
DS-SoybeanNet (5 × 5)11.7702616 MB
AlexNet19.011151 MB
DS-SoybeanNet (7 × 7)22.9552616 MB
ResNet5036.099306 MB
InceptionResNetV244.328653 MB
VGG1667.080623 MB
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, S.; Feng, H.; Han, S.; Shi, Z.; Xu, H.; Liu, Y.; Feng, H.; Zhou, C.; Yue, J. Monitoring of Soybean Maturity Using UAV Remote Sensing and Deep Learning. Agriculture 2023, 13, 110. https://doi.org/10.3390/agriculture13010110

AMA Style

Zhang S, Feng H, Han S, Shi Z, Xu H, Liu Y, Feng H, Zhou C, Yue J. Monitoring of Soybean Maturity Using UAV Remote Sensing and Deep Learning. Agriculture. 2023; 13(1):110. https://doi.org/10.3390/agriculture13010110

Chicago/Turabian Style

Zhang, Shanxin, Hao Feng, Shaoyu Han, Zhengkai Shi, Haoran Xu, Yang Liu, Haikuan Feng, Chengquan Zhou, and Jibo Yue. 2023. "Monitoring of Soybean Maturity Using UAV Remote Sensing and Deep Learning" Agriculture 13, no. 1: 110. https://doi.org/10.3390/agriculture13010110

APA Style

Zhang, S., Feng, H., Han, S., Shi, Z., Xu, H., Liu, Y., Feng, H., Zhou, C., & Yue, J. (2023). Monitoring of Soybean Maturity Using UAV Remote Sensing and Deep Learning. Agriculture, 13(1), 110. https://doi.org/10.3390/agriculture13010110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop