Next Article in Journal
Sustainable Evolution of the Geographic System in the Regional Park “Carrascoy y El Valle” in the Region of Murcia (Southeast Spain)
Next Article in Special Issue
Sample Expansion and Classification Model of Maize Leaf Diseases Based on the Self-Attention CycleGAN
Previous Article in Journal
Characterization of a New Powdered, Milk-Based Medicinal Plant (Alcea rosea) Drink Product
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis and Research on Rice Disease Identification Method Based on Deep Learning

1
College of Information Technology, Jilin Agricultural University, Changchun 130118, China
2
Jilin Precision Agriculture and Big Data Engineering Research Center, Changchun 130118, China
*
Author to whom correspondence should be addressed.
Sustainability 2023, 15(12), 9321; https://doi.org/10.3390/su15129321
Submission received: 30 March 2023 / Revised: 28 May 2023 / Accepted: 31 May 2023 / Published: 9 June 2023
(This article belongs to the Special Issue Sustainable Development of Intelligent Agriculture)

Abstract

:
Rice is one of the most important food crops in China and around the world. However, with the continuous transformation of human activities, the quality of climate, soil, and water sources has also changed, and disease affecting rice has become increasingly serious. Traditional artificial pest identification methods have been unable to adapt to the occurrence of a large number of diseases, and artificial naked eye identification also increases the uncertainty of the identification results, and cannot “suit the remedy to the case”, which will not cure the disease, or even achieve half the result with half the effort. In the incidence range of rice diseases, rice blast, rice false smut, and bacterial blight have the highest incidence rate, the greatest harm, and are the most representative. Therefore, this paper mainly focuses on the above three categories. In this paper, the identification of rice diseases is further studied. First, sample pictures of rice blast, rice false smut, and bacterial leaf blight diseases are collected. Due to the differences in the distance and light of the sample photos, their size and angle is biased. Therefore, some means are needed to unify the specifications of these images, so as to improve the efficiency of network model recognition. Neural network recognition needs to absorb many sample images to classify and learn features. The main research objects of this paper are rice blast, rice false smut, and bacterial wilt. Therefore, this paper also expands the data set for this kind of disease, and unifies the specifications through size cutting, angle change, and vertical symmetrical mirror image processing. Then, we built a new network model based on deep learning to realize the parameter initialization design. The accuracy of the rice disease identification model built at the beginning does not satisfy the practical requirements. In order to upgrade the model in depth, this experiment increases the entry point of analysis and research, and integrates four parameters: iteration times, batch size, learning rate, and optimization algorithm in order to strive for the optimization of the experimental results. In this study, the confusion matrix is selected as the evaluation standard, and experimental results with more objectivity and reference value are obtained through the horizontal comparison of visual graphics generator (VGG) and residual network (ResNet), two highly referential network models. The results show that the recognition accuracy of the optimized model is 98.64%, which achieves the goal of accurately identifying diseases.

1. Introduction

Rice is one of the most important rations of the people in China. In the past millennium, rice has changed from an indispensable food source to a source of life in promoting national economic development and ensuring national survival and development. China has a huge demand for rice, and there is a need to minimize the risk of disease in the supply chain. The occurrence of rice diseases and pests has always been a key factor in reducing output [1,2]. According to statistics, the occurrence of rice bacterial blight will reduce the yield by 20–30%. In serious cases, it may lead to a no-grain harvest. Allowing rice diseases to flourish will have a bad impact on China’s food security, and even national health and the national economy. In rice planting, accurate identification of diseases, successful disease control, and early treatment of diseases and pests play decisive roles in the final yield of grain. The most commonly used identification methods of rice diseases are mainly divided into two types: one is to obtain the results through the identification of disease samples by experts, and the other is to diagnose by the growers themselves through plant disease diagnosis or consulting relevant books. However, there are certain difficulties in rice disease identification. Human judgment requires high theoretical knowledge and practical experience. This method is not only time-consuming and labor-consuming, but also prone to misjudgment. Therefore, we should focus on the research of efficient and high accuracy automatic identification methods. Deep learning technology can simplify the steps of identification, improve the accuracy and efficiency of disease identification, and provide great help for the precise treatment and subsequent prevention and control of diseases.
In the field of crop disease image recognition based on deep learning, A Fuentes and other authors used tomato disease and insect pest data as experimental samples. The analysis showed that the recognition accuracy of faster region-based convolutional neural network (R-CNN) combined with VGG-16 was the highest [3]. BARANWAL S and other authors designed a six-layer apple leaf disease identification structure based on the LeNet model, selected the DenseNet and perception modules that were pre-trained on ImageNet, and used them in the network. Compared with other latest methods, this method has higher performance. In the public data set, its average prediction accuracy is not less than 98.54% [4]. Priyadharshini R A and other authors proposed a disease classification based on LeNet architecture, including a Bayesian classifier and support vector machine classifier. In their research, 68.1% accuracy was obtained through the combination of 10 different training and test data sets. The LeNet optimization model increases the probability of successful identification of diseases by 30%, greatly improving the effectiveness of inspection. There are also many scholars in China who pay attention to the use of deep learning methods in agriculture. Most of the existing studies focus on the image recognition of magnified images on crop leaves [5]. Li Jing and others established a deep learning model for identifying 14 crop varieties and 26 crop diseases, explored the impact of resolution and iteration times, and achieved 94.1% accuracy in the test set of tobacco diseases [6]. Sun Jun and others propose an optimized recognition model, which also uses the convolutional neural network structure, but is upgraded and transformed one-by-one in terms of model parameters and training process. Multiple convolution and rectified linear unit (ReLU) layers are used to process images to locate features. Different pooling layers with various filters are used to identify specific parts of the image, and flatten and feed the combined element map to the fully connected layer to obtain the final output. The recognition accuracy of many common diseases is close to 100% [7]. Xu Dong and others used a convolution neural network to verify images of soybeans. Compared with traditional models and other classifiers (such as support vector machines), the strategy of evaluating sparse data does not require high technical efficiency, nor does it require the model to have larger data or deeper network saturation, so as to obtain better results. The recognition rate of soybean disease has reached 96.7% [8]. Yang Jindan and others identified and sorted out the model framework of rice leaf disease according to the plant virus leaf powdery mildew, mainly including three kinds of rice leaf diseases, namely, bacterial leaf blight, brown spot, and leaf smut. In order to extract features accurately, the AlexNet model of deep convolution neural network is used, and the accuracy rate is 98.61% [9]. The main research objectives of Lin Zhongqi and others are wheat leaf diseases. The research method magnifies the differences of different leaves through image recognition, and integrates the difference features into local support vector machines, while conducting research in combination with a CNN network [10]. Guo Xiaoqing and others proposed a multi-scale AlexNet structure that optimizes the image diagnosis model [11].
To sum up, there are many technical methods for plant disease identification, but there are still some problems in some aspects: (1) complex characteristic parameters and poor generalization of the model. As plant disease recognition involves many characteristic parameters, such as disease shape, color, disease spot size, and so on, the calculation process is very complex. For a variety of feature types, we need to choose a variety of suitable extraction methods, which makes the experimental process more difficult. (2) The requirement of a neural network model is high, and the experimental effect is poor. With the continuous development of neural network models, the number of network layers is increasing, and the number of convolution cores is increasing, so we need to use a large number of data samples for practice. However, some experiments have a small number of samples due to environmental reasons, resulting in a poor recognition effect in a deep convolution network structure.
In this study, a convolutional neural network is used to recognize rice disease images, and a database containing 2000 images of three common diseases of rice (rice blast, rice false smut, and white leaf blight) is established. The problem of the multi-sample demand of complex models is solved, and a new network model based on deep learning is built to realize parameter initialization design. The accuracy of the rice disease identification model built at the beginning does not meet the practical requirements. In order to carry out in-depth upgrading and transformation of the model, this experiment increases the entry point of analysis and research, integrating the four parameters of iteration times, batch size, learning rate, and optimization algorithm so as to strive to optimize the test results. In this study, the confusion matrix is selected as the evaluation standard, and experimental results with more objectivity and reference value are obtained through the horizontal comparison of VGG and ResNet, two highly referential network models. The results show that the recognition accuracy of the optimized model is 98.64%, which achieves the goal of accurately identifying diseases. This study will improve the accuracy and efficiency of rice disease identification through the improvement of a convolutional neural network, and provide assistance for the precise treatment and subsequent prevention and control of rice diseases.

2. Materials and Methods

2.1. Dataset

Based on the complexity of the current convolutional neural network, too little sample data cause the model to over-fit and the training ends, so a large number of learning and training pictures are required. The experimental data are image samples obtained from the website of the national plant pathology Association of America (URL: https://www.kaggle.com/datasets), open-source image samples obtained on the website, and image samples taken in actual rice fields. It mainly includes rice blast, rice false smut, and rice bacterial blight. The tools for obtaining data in actual rice fields are HD cameras and smartphones. The original data obtained include 555 images of rice blast, 646 of rice false smut, and 345 of bacterial leaf blight, as shown in Figure 1. Through the data enhancement methods of clipping, multi-angle rotation, and vertical image, the sample library of disease images was expanded, and, finally, 2000 images of the three diseases were obtained.

2.1.1. Image Normalization

In the process of deep learning, image preprocessing is very important. Some of the data in the experiment were taken from the internet, and many factors led to the different states. This paper uses bicubic interpolation to adjust the size of the image to match the model of 224 × 224× input size of 3. Assuming the size of cropped image A is m × n. The size of image B with equal scaling is m × n. Based on the size ratio of the image, the corresponding coordinates of B (x, y) on A can be obtained, and the 16 points closest to the pixel point (x, y) are used as parameters to calculate the pixel value B (x, y) of the target image. The calculation of the entire process is shown in Formula (1).
B X , Y = i = 0 3 j = 0 3 a i j × W ( i ) × W ( j )

2.1.2. Image Enhancement

The number of sample sets affects the recognition performance of the convolution training model. When there are few samples, the training model has an overfitting problem [12,13]. To improve the generalization ability of the model, the amount of data and the amount of data changes should be increased as much as possible. Common data enhancement methods include cropping, rotation, mirroring, adding noise, color jitter, and other methods [14]. The data augmentation methods used in this experiment include cropping, multi angle rotation, and vertical mirroring. The results are shown in Figure 2. By expanding the disease image sample database, 2000 images of each of these three diseases were ultimately obtained. The expanded data are divided into training sets, validation sets, and testing sets at 6:3:1, namely, 3600 training sets, 1800 validation sets, and 600 testing sets.

2.2. Models

In recent years, people’s ability to utilize computation and process problems has gradually improved, and the complexity of problems that can be solved has also increased. The number of times that neural network models are used to process problems has also gradually increased [15,16]. Different network models have been established to solve different problems, such as the VGG model, RNN model, and ResNet model. However, due to the limitations of RNN, such as difficulty in training [17], it is not consistent with the model construction in this article, Therefore, the concept of model design in this article refers to the VGG model and ResNet model
The VGG network architecture is mainly composed of four parts: convolution, ReLU activation function, maximum pooling layer, and full connection layer. The input images are processed through 13 convolution layers, 5 pooling layers, and the full connection layer, and finally the images are classified.
In the ResNet model, the input picture is not activated by the activation function immediately after passing through any convolution layer, but first passes through the standardization layer and then arrives at the activation function, which is composed of a convolution layer, standardization layer, and an activation layer.
By analyzing the above two models, a neural network model can be constructed to identify rice diseases. However, due to the existence of gradient problems, the design of the network model structure needs to be optimized. Therefore, we designed our own network model, and its structure diagram is shown in Figure 3.
The improved network model in this article is based on VGG and ResNet, consisting of 5 convolutional layers, 5 pooling layers, and an activation layer. The pooling layers and convolutional layers are interleaved.

2.2.1. Convolution Layer Design

The convolutional layer design in this article mainly refers to the VGG network. The convolution kernel defines the size range of convolution, which represents the size of the receptive field in the network. The most common two-dimensional convolution kernel is 3 × the convolutional kernel of 3. In general, the larger the convolution kernel, the larger the receptive field, the more picture information you see, and the better the global features you can obtain. However, large convolutional kernels can lead to a significant increase in computational complexity, a decrease in computational performance, and the inability to accurately capture disease information. Therefore, this article has chosen a relatively suitable convolutional kernel size of 3 × 3.In this experiment, five convolutional layers were designed, namely, C1, C2, C3, C4, and C5. Different sizes of convolutional kernels were used to extract lesions, and a zero-padding method was also needed to ensure consistency in image size. The details are as follows:
First, we used 3 for all layers from C1 to C5. The size 3 convolution checked the input image for convolution. The sliding step size of all convolution layers was set to 1, and the filling method was also set to the same. After 5 convolutions, the number of feature maps obtained after each convolution is 64128256512 and 512, respectively, from C1 to C5.

2.2.2. Pool Layer Design

In model design, a pooling layer is usually set after the convolution operation for dimensionality reduction. The pooling method selected in this article is maximum pooling.
This experiment designs 5 pooling layers, namely, P1, P2, P3, P4, and P5. The filter sampling size was set to 2 × 2. We selected 2 for the sliding step size without filling, and the result is as follows:
The P1 layer mainly down-samples the features processed by the C2 layer. Due to the zero filling of the convolutional layers, the input feature map size of the P1 layer is 224 × 224, P1 layer output size is (224 − 2)/2 + 1 = 112;
The output sizes from the P2 layer to the P5 layer are (112 − 2)/2 + 1 = 56, (56 − 2)/2 + 1 = 28, (28 − 2)/2 + 1 = 14, and (14 − 2)/2 + 1 = 7, respectively.

2.2.3. Full Connection Layer Design

The main function of the fully connected layer is to achieve classification [18]. There are relatively many parameters in the fully connected layer. Therefore, this paper also draws on the idea of three connection layers.
The number of nodes in the fully connected layer is usually set to 2n (n=1, 2, 3 … N). Therefore, for the first two fully connected layers, the values of 4096 and 1024 are selected. The third layer is different from the first two layers, and the number of nodes is 3.

2.3. Model Training

2.3.1. Solution of Over-Fitting Problem

Over-fitting refers to over-modeling of training data [19]. When a model levels (ML) model relies heavily on training data, the over-fitting phenomenon usually occurs. In order to effectively avoid the occurrence of over-fitting, this experiment optimizes the full connection layer by adding regularization.
In mathematics, statistics, finance, and computer science, regularization is used to solve the ill-conditioned problem of information processing or prevent over-fitting. The additional term controls the function of excessive fluctuation so that the coefficient does not take the extreme value. L1 regularization technology (lasso regression) and L2 (ridge regression) regularization and their mathematical expressions are shown in (2) and (3), respectively.
L 1 = i = 1 m y i f ( x i )
L 2 = i = 1 m ( y i f ( x i ) ) 2
where yi represents the target value, f (xi) is the estimated value, the L1 parameter penalty term calculates the sum of the absolute difference between the target value yi and the estimated value f (xi), and the L2 parameter penalty term is the sum of the squares of the difference between the target value and the estimated value, resulting in different regularization and different penalty effects. L2 is different from L1. One difference is that there is only one best prediction line, and the other may have multiple optimal solutions. If the regular term is equal to 0, the minimum value is obtained. However, if the regular term is very large, it increases to too much weight, which leads to insufficient fitting. L2 regularization can avoid the over-fitting problem.

2.3.2. Cross-Entropy Loss Calculation

After the model is built, it is necessary to check whether its performance can guarantee the accuracy of the calculation results. Generally, the performance of the model is expressed by the error, that is, the difference between the predicted value and the true value. The difference is the result of the calculation using the loss function. The greater the error value, the worse the performance of the model. The model needs to be improved, to better the performance of the model. The loss function is usually softmax cross-entropy loss function. Softmax output probability distribution softmax function, also known as the exponential normalization function, is a normalized form of logistic function. In general, when encountering problems such as the uncertainty of the output value range, which makes it difficult to intuitively determine the meaning of these values, and the difficulty in measuring the error between these discrete values and the output value of the uncertain range, due to the fact that the real label is a discrete value, softmax normalization processing is used to solve the problem. The softmax function can compress a k-dimensional real vector into a real vector in the range [0–1], as shown in Formula (4):
S j = S o f t max ( Z j ) = e Z j k = 1 K e Z k ( j = 1,2 , , K )
Among them, S j represents the output probability, K represents the total number of categories, and Z j represents the output of the previous output unit.
Formula (4) shows that the output of the original neural network is used as confidence to generate new outputs, and the new outputs meet all the requirements of the probability distribution, transforming the output of the neural network into a probability distribution.
After the probability distribution is obtained from the softmax function output, the distance between the predicted probability distribution and the probability distribution of the real answer can be calculated through cross-entropy, which is the error. The calculation expression of cross entropy is shown in Formula (5).
H ( y , t ) = H t ( y ) = i t i log y i
wherein H t ( y ) represents the cross-entropy between the predicted value and the real value, y represents the output predicted value, and t represents the output real value.

2.4. Development Environment

This experiment uses pytorch’s GPU version as the platform for convolutional neural network model construction and training. The hardware environment adopts Intel Xeon e5-2680 V4 CPU, a Samsung SSD 860 512 G hard disk, Kingston DDR4 64GB memory, and GPU adopts a NVIDIA Titan XP graphics card, 12 GB video memory. The operating system adopts Windows 10, and python3.6 language is used to write programs in the integrated development environment PyCharm.

2.5. Evaluation Indices

In this study, accuracy, precision, recall, and F1 score are selected to evaluate the model. The formulae are shown in (6)–(9).
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Among them, true positive (TP): the predicted value matches the actual value; the actual value is positive, and the model prediction is positive. True negative (TN): the predicted value matches the actual value; the actual value is negative, and the model prediction is negative. False positive (FP)—type 1 error: the predicted value is incorrectly predicted; the actual value is negative, but the model prediction is positive, also known as type 1 error. False negative (FN)—type 2 error: the predicted value is incorrectly predicted; the actual value is positive, but the model prediction is negative, also known as type 2 error.

3. Results

3.1. Parameter Optimization Results

(1)
Iteration number experiment
The data support of this model is the expanded data set of 2000 samples of rice blast, false smut, and bacterial blight. The training set, verification set, and test set are divided according to the ratio of 6:3:1, i.e., 3600 training sets, 1800 verification sets, and 600 test sets.
Firstly, we initialized the data and set the size to 224 × 224, and set the learning rate to 0.01. Four groups of schemes are set up in this experiment, and 400, 600, 800, and 1000 are selected as the iteration times. The relationship between the iteration times and the accuracy of the model is observed, and the parameters that are most suitable for the experimental model are selected. Figure 4 shows the results of the identification model of the change in the number of iterations. As shown in the figure, when the number of iterations is less than 800, the accuracy of the verification set increases with the increase in the number of iterations. When the number of iterations exceeds 800, the accuracy of the training set changes slightly and gently. When the number of iterations is 796, the accuracy of the training set is the highest, which is 90.32%. Too many iterations increases the running time of the model and reduces the computational efficiency, so 800 iterations were selected in all subsequent experiments.
(2)
Batch size experiment
Batch size refers to the number of training examples in a forward or backward pass. Batch size affects the speed of model learning and the stability of the learning process. It is an important super parameter. A small batch usually works well. Therefore, we choose 16, 32, 64, and 128 as the batch sizes of the experiment to analyze the training results, as shown in Table 1.
Table 1 shows the impact of different batch sizes on the training results. Through analysis, we find that a large batch size has a good effect, but the disadvantage is that the convergence speed is slow, and too large a batch size may cause the system operation to jam or enter downtime. Therefore, we choose 64 as the parameter in a compromise, and Figure 5 shows the accuracy and loss results under the selection of batch size 64. By comparing with Figure 4, we can draw the following conclusion: when the value of the batch size increases, the gradient descent direction tends to be accurate.
(3)
Optimization algorithm and learning rate experiment
The goal of the optimization algorithm is to achieve the best point of maximum value optimization, and achieve a balance between excessive risk and excessive caution. The most commonly used optimization methods are the gradient descent method, moment optimization method, and adaptive learning rate optimization algorithm. Different algorithms have different effects on the results. In this study, three learning rates of 0.1, 0.01, and 0.001 are selected. The final results of the nine groups of models are shown in Table 2.
Through analysis, we find that the recognition results are the lowest when the learning rate of the three algorithms is 0.1, which does not meet the experimental requirements. RMSprop (92.31%), Adam (98.64%), and SGD (84.79%) are very similar algorithms. Obviously, Adam is slightly better. Therefore, this study finally selects the Adam optimization algorithm and sets the learning rate to 0.001.

3.2. Comparison Results with Other Models

In order to reduce the calculation error, this verification also refers to other methods [20,21,22]. The recognition accuracy of these three diseases under the VGG and ResNet models is selected, and the analysis results are shown in Figure 6. From Figure 6a,b, we can see that the highest accuracy for VGG and ResNet in the training set is 0.9258 and 0.886, respectively, and the highest accuracy of the improved model in this study in the training set is 0.9864, which is 0.0606 and 0.1004 higher than the aforementioned two models, respectively. The highest accuracy for VGG and ResNet in the verification set is 0.9186 and 0.9, respectively, and the highest accuracy of the improved model in this study in the verification set is 0.9849, which is 0.0663 and 0.0849 higher than the aforementioned two models, respectively. From Figure 6c,d, we can see that the lowest loss values for VGG and ResNet in the training set are 0.1108 and 0.125, respectively, and the lowest loss value for the improved model in this study in the training set is 0.0705, which is 0.0403 and 0.0545 lower than the aforementioned two models, respectively. The lowest loss values for VGG and ResNet in the validation set are 0.1035 and 0.132, respectively, and the lowest loss value for the improved model in this study in the validation set is 0.0704, which is 0.0331 and 0.0616 lower than the aforementioned two models, respectively. To sum up, the model adopted in this paper has advantages over VGG and ResNet in the recognition accuracy of the three kinds of diseases studied in this experiment. On the other hand, this model has fewer parameters and faster calculation.

3.3. Confusion Matrix Results

The matrix compares the actual target value with the target value predicted by the machine learning model [23]. Figure 7 shows the confusion matrix of the improved model. It can be seen from the figure that the classification accuracy of bacterial blight is up to 100%, and the classification effect of rice false smut is also up to 98%, which shows very high accuracy. There are 3 and 10 misclassified images of rice false smut and rice blast, respectively, and the misclassified images are basically distributed near the main diagonal of the confusion matrix, indicating that they are predicted to be close categories. We tested 600 pictures and calculated the accuracy, precision, recall, and F1 score of the improved model according to the performance test parameters of the network model. As shown in Table 3, the average values of these four indicators are 0.9766, 0.9786, 0.9783, and 0.9760, respectively, and are all above 0.97, indicating that the model has high accuracy in rice disease identification.

4. Conclusions

This study designs a rice disease recognition model based on a convolutional neural network. In order to reduce the risk of over-fitting from the parameters of the model, L2 regularization is used in this experiment, and the loss of the model is calculated by the cross-entropy loss function. In order to reduce the loss, the Adam algorithm is used. The first is to analyze the different effects of the number of iterations on the results. The optimal number of iterations is 800, and the accuracy of the training set of the initial model is 90.32%. Secondly, by comparing the batch size, optimization algorithm, and learning speed, further optimization experiments are carried out. The final results show that when the learning rate of identifying rice diseases is set to 0.001, the number of batches is 64, the maximum accuracy of the model is 98.64%.
In this paper, the images of three diseases are analyzed, and good results are achieved by improving the parameter model, so that the recognition accuracy of the three diseases reaches 98.64%. Even though the recognition rate has made great progress, there are still many aspects that need further research, such as expanding the rice disease data set, preprocessing image methods, and optimizing the neural network model.

Author Contributions

Conceptualization, H.L. and H.Y.; methodology, H.L.; software, Y.C. and J.W.; validation, Y.C.; investigation, J.W.; resources, H.Y.; visualization, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was financially supported by the Science and Technology Development Program of Jilin Province, grant number 20220508133RC.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The main source to obtain image samples was the website of the national plant pathology Association of the United States and we obtained open-source images on the website and also used image samples captured in actual rice fields.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Tang, L.; Zhou, J.; Yu, X. Empirical analysis of reducing rice yield loss by conservation tillage—Based on the survey data of 1080 rice farmers in 4 provinces. J. Nat. Resour. 2017, 32, 1016–1028. [Google Scholar]
  2. Sun, B.; Liu, T. Investigation on Obstacle Factors Affecting High and Stable Yield of Rice. Friends Farmers Wealth 2011, 33. [Google Scholar]
  3. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A Robust Deep-Learning-Based Detector for Real-Time Tomato Plant Diseases and Pests Recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Baranwal, S.; Khandelwal, S.; Arora, A. Deep learning convolutional neural network for apple leaves disease detection. In Proceedings of the International Conference on Sustainable Computing in Science, Technology and Management (SUSCOM), Amity University Rajasthan, Jaipur, India, 26–28 February 2019. [Google Scholar] [CrossRef]
  5. Priyadharshini, R.A.; Arivazhagan, S.; Arun, M.; Mirnalini, A. Maize leaf disease classification using deep convolutional neural networks. Neural Comput. Appl. 2019, 31, 8887–8895. [Google Scholar]
  6. Li, J. Research on tobacco disease automatic recognition based on convolution neural network. Ph.D. Thesis, Shandong Agricultural University, Taian, China, 2016. [Google Scholar]
  7. Sun, J.; Tan, W.; Mao, H.; Wu, X.; Chen, Y.; Wang, L. Recognition of multiple plant leaf diseases based on improved convolutional neural network. Trans. Chin. Soc. Agric. Eng. 2017, 33, 209–215. [Google Scholar]
  8. Xu, D. Research on soybean disease identification based on convolution neural network. Ph.D. Thesis, Anhui University, Hefei, China, 2018. [Google Scholar]
  9. Yang, J.; Yang, T.; Miao, T.; Chao, Z.; Qiucai, S.; Yufei, P.; Pozhang, M.; Yuqing, D. Identification of strawberry leaf powdery mildew disease based on convolution neural network. Jiangsu Agric. J. 2018, 34, 527–532. [Google Scholar]
  10. Lin, Z. Research on wheat leaf disease image recognition based on convolution neural network. Ph.D. Thesis, Shandong Agricultural University, Taian, China, 2018. [Google Scholar]
  11. Guo, X.; Fan, T.; Shu, X. Tomato leaf disease image recognition based on improved Multi-Scale AlexNet. Trans. Chin. Soc. Agric. Eng. 2019, 35, 162–169. [Google Scholar]
  12. Duan, H. Gesture recognition and Implementation Based on convolutional neural network. Ph.D. Thesis, Jinan University, Guangzhou, China, 2018. [Google Scholar]
  13. Gao, H.; Chen, Z.; Li, C. Sandwich Convolutional Neural Network for Hyperspectral Image Classification Using Spectral Feature Enhancement. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 3006–3015. [Google Scholar] [CrossRef]
  14. Guo, Y.; Lei, B.; Nie, H.; Li, N. Intelligent high-precision image recognition algorithm based on deep learning. Mod. Electron. Technol. 2021, 44, 173–176. [Google Scholar]
  15. Tan, Y.; Ouyang, C.; Li, L.; Liao, T.; Tang, P. Research on rice disease image recognition based on deep convolution neural network. J. Jinggangshan Univ. (Nat. Sci. Ed.) 2019, 2, 31–38. [Google Scholar]
  16. Su, B. Identification technology of rice diseases and pests based on image processing. Inf. Technol. Informatiz. 2018, 218, 96–98. [Google Scholar]
  17. Gong, Q. Text classification based on cyclic neural network model. Ph.D. Thesis, Huazhong University of Science and Technology, Wuhan, China, 2016. [Google Scholar]
  18. Chen, Y. Research on image classification of crop diseases and pests based on convolutional neural network. Ph.D. Thesis, Jiangxi Agricultural University, Nanchang, China, 2019. [Google Scholar]
  19. Ma, C.; Yuan, T.; Yao, X.; Ji, Y.; Li, L. Research on image recognition method of field rice disease based on hog + SVM. Acta Agric. Shanghai 2019, 35, 131–136. [Google Scholar]
  20. Huang, Y.; Pu, Y.; Zhang, H. Research on fruit image recognition algorithm based on multi-scale feature fusion. J. Chang. Univ. Technol. Nat. Sci. Ed. 2021, 44, 87–94. [Google Scholar]
  21. Tan, Y.; Long, L.; Chun, O.; Huan, L. Image recognition of camellia oleifera disease based on convolutional neural network and transfer learning. J. Agric. Eng. 2018, 34, 194–201. [Google Scholar]
  22. Zhong, L. Research on Image Recognition of Wheat Leaf Disease Based on Convolutional Neural Networks. Ph.D. Thesis, Shandong Agricultural University, Shandong, China, 2018. [Google Scholar]
  23. Yi, R.; Lu, L. Research on overfitting problem of convolutional neural network. Inf. Syst. Eng. 2019, 140–142. [Google Scholar]
Figure 1. Partial display of three disease samples.
Figure 1. Partial display of three disease samples.
Sustainability 15 09321 g001
Figure 2. Display of data enhancement results.
Figure 2. Display of data enhancement results.
Sustainability 15 09321 g002
Figure 3. Improved network model architecture.
Figure 3. Improved network model architecture.
Sustainability 15 09321 g003
Figure 4. Result of accuracy (a) and loss (b) under different iteration times.
Figure 4. Result of accuracy (a) and loss (b) under different iteration times.
Sustainability 15 09321 g004
Figure 5. Result of accuracy (a) and loss (b) (batch size = 64).
Figure 5. Result of accuracy (a) and loss (b) (batch size = 64).
Sustainability 15 09321 g005
Figure 6. The results of the improved model compared with VGG and ResNet.
Figure 6. The results of the improved model compared with VGG and ResNet.
Sustainability 15 09321 g006
Figure 7. Distribution of three rice diseases: confusion matrix (a); normalized confusion matrix (b).
Figure 7. Distribution of three rice diseases: confusion matrix (a); normalized confusion matrix (b).
Sustainability 15 09321 g007
Table 1. Effect of different batch size on training results.
Table 1. Effect of different batch size on training results.
Batch Size163264128
Training set accuracy0.80120.85640.92580.9245
Verification set accuracy0.78430.84210.92170.9134
Training time/s392402415439
Table 2. Training result display.
Table 2. Training result display.
Learning Rate/Optimization Algorithm0.10.010.001
Adam0.70560.88320.9864
SGD0.34480.78460.8479
RMSProp0.65210.80910.9231
Table 3. Four indicators report.
Table 3. Four indicators report.
Disease TypeAccuracyPreciseRecall RateF1 Score
Rice false smut0.98000.95630.98500.9640
Rice blast0.95000.98950.95000.9693
Bacterial blight1.00000.99001.00000.9949
Average value0.97660.97860.97830.9760
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, H.; Cui, Y.; Wang, J.; Yu, H. Analysis and Research on Rice Disease Identification Method Based on Deep Learning. Sustainability 2023, 15, 9321. https://doi.org/10.3390/su15129321

AMA Style

Liu H, Cui Y, Wang J, Yu H. Analysis and Research on Rice Disease Identification Method Based on Deep Learning. Sustainability. 2023; 15(12):9321. https://doi.org/10.3390/su15129321

Chicago/Turabian Style

Liu, He, Yuduo Cui, Jiamu Wang, and Helong Yu. 2023. "Analysis and Research on Rice Disease Identification Method Based on Deep Learning" Sustainability 15, no. 12: 9321. https://doi.org/10.3390/su15129321

APA Style

Liu, H., Cui, Y., Wang, J., & Yu, H. (2023). Analysis and Research on Rice Disease Identification Method Based on Deep Learning. Sustainability, 15(12), 9321. https://doi.org/10.3390/su15129321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop