Next Article in Journal
Experimental Evaluation of the Bending Behavior of a Drilled Shaft with Partial Casing under Lateral Loads
Next Article in Special Issue
Analysis of Water Consumption in Fruit and Vegetable Processing Plants with the Use of Artificial Intelligence
Previous Article in Journal
SESAM Mode-Locked Yb:Ca3Gd2(BO3)4 Femtosecond Laser
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Plant Diseases Identification through a Discount Momentum Optimizer in Deep Learning

1
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
2
School of Computer Science and Technology, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
3
Institute of Network Security and Trusted Computing, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(20), 9468; https://doi.org/10.3390/app11209468
Submission received: 3 July 2021 / Revised: 20 August 2021 / Accepted: 25 August 2021 / Published: 12 October 2021
(This article belongs to the Special Issue Applications of Computer Science in Agricultural Engineering)

Abstract

:
Deep learning proves its promising results in various domains. The automatic identification of plant diseases with deep convolutional neural networks attracts a lot of attention at present. This article extends stochastic gradient descent momentum optimizer and presents a discount momentum (DM) deep learning optimizer for plant diseases identification. To examine the recognition and generalization capability of the DM optimizer, we discuss the hyper-parameter tuning and convolutional neural networks models across the plantvillage dataset. We further conduct comparison experiments on popular non-adaptive learning rate methods. The proposed approach achieves an average validation accuracy of no less than 97% for plant diseases prediction on several state-of-the-art deep learning models and holds a low sensitivity to hyper-parameter settings. Experimental results demonstrate that the DM method can bring a higher identification performance, while still maintaining a competitive performance over other non-adaptive learning rate methods in terms of both training speed and generalization.

1. Introduction

The outbreak of plant diseases is a threat to food production and security at the global scale. It can cause disastrous consequences for smallholder farmers representing 85% of the world’s farms whose livelihoods depend on healthy crops [1]. In order to manage the detection and spread of plant diseases, several diagnostic protocols are developed in literature. However, challenges exist that prevent this kind of technology from being adopted in practice [2].
In previous research, a variety of generic machine learning (ML) methods are popularity in plant diseases identification including K-nearest neighbor (KNN), support vector machines (SVM), artificial neural networks (ANN), amongst others [3]. These methods are relatively successful under limited and constrained setups. However, these traditional machine learning methods have the problems of incomplete feature selection and fussy manual feature selection [4]. Deep Learning (DL) in particular offers very novel approaches to classify images because it extends classical ML by adding more “depth” (complexity) into the model [5]. These complex models can increase classification accuracy or reduce generalization error. The agricultural field, and especially the image-based plant diseases identification task, has not been an exception to this [6]. Indeed, since 2015, research on plant diseases detection has strongly veered towards using deep learning. AlexNet, GoogLeNet, VGG, ResNet, and DenseNet deep learning models are commonly used [7]. Solemane [8] identifies a mildew disease in crop millet and takes VGG16 as a pre-trained model with ImageNet as a source dataset. The method shows the effectiveness of transfer learning for disease classification with small data. The performance of the method gives 95.00% accuracy. It is not suitable for the identification of other plant diseases. Within the plantvillage data set, Mohanty trains plant diseases identification model and achieves an accuracy of 99.35% [9]. However, when tested on images taken under conditions different from the images used for training, the model’s accuracy is reduced substantially. In [10], the authors fine-tune pretrained deep convolutional neural networks of AlexNet, GoogLeNet, and VGGNet using the LifeCLEF 2015 plant task dataset.They have improved the overall validation accuracy of the top system by 15% points while outperforming the top three competition participants in all categories. In [11], fine-tuning and evaluation of state-of-the-art deep convolutional neural network for image-based plant disease classification are performed. DenseNets obtains a test accuracy score of 99.75% for the 30th epoch, beating the rest of the architectures. This research needs to be done to improve on the computational time and training process. As reported in [12], the authors focused on techniques to achieve an accuracy score of over 93% with class weight, SMOTE (Synthetic Minority Over-sampling Technique), and focal loss with deep convolutional neural networks from scratch. The goal was to counter high-class imbalance so that the model can accurately predict underrepresented classes. Their dataset biased towards Cassava Mosaic Disease and Cassava Brown Streak Virus Disease classes. They need further research though for multiple diseases on the same plant and multiple diseases on different plants.
Training CNNs (Convolutional Neural Networks) to achieve high multiple plant diseases identification accuracy was very challenging due to two reasons: (1) Deep learning is highly dependent on the dataset. However, there are few public data sets in the field of plant diseases identification. (2) The deep learning model has more network layers and parameters resulting in more time and cost for training and validation. (3) Existing methods only focus on single-target and few-target plant diseases image with simple backgrounds. In real life, crop diseases have the characteristics of various types, large quantity, and complex backgrounds. To solve the above problems, this paper studies the optimization method in deep learning towards plant identification tasks. The non-adaptive learning rate optimization method has been widely applied in deep learning, with the virtues of global optimization and rapid convergence. Based on the non-adaptive learning rate method, a new optimization algorithm is presented to increase the accuracy of identification. As a whole, the contribution of this article is as follows:
  • Applying the discount weighted moving average to the momentum buffer m t , a relative result reveals the higher recognition ability and faster convergence.
  • Another key contribution of this work is show that DM does provide performance gains over other non-adaptive learning rate methods on plant diseases classification task.
  • It is proved that discount momentum optimizer is insensitive to deep learning architectures and hyper-parameters.
  • The DM method is capable of recovering popular non-adaptive learning rate methods in an efficient and accessible manner.
The rest of this paper is organized as follows: Section 2 introduces a state-of-the-art of deep learning optimization technology. Section 3 describes the details of non-adaptive learning rate methods as well as the proposed DM optimizer. Section 4 presents the implementation, empirical results, and analysis. The major work is discussed and wrapped up in Section 5 and Section 6.

2. Related Work

Deep learning optimization methods are currently used to deal with the overfitting and performance deterioration problems. There are several common optimization methods. Here, we introduce the transfer learning method at first. As reported in [9,10,11,13], transfer learning techniques fine-tune transmitted sub-networks to adapt to new data and then mining depth features, which can effectively solve the small data sets problem. Another optimization method called data augmentation. This method enlarges the dataset to reduce the chance of over-fitting. Data enhancement methods include segmented symptom images, geometrical transformations, and intensity transformations [10,14]. With the exception of the above-mentioned two optimization methods, optimizing network parameters of the deep learning models is also commonly used. These optimization methods improve the overall performance of models from convergence, over-fitting, running time, and generalization. A stochastic gradient descent (SGD) optimizer is one of the heavily used optimization methods in deep learning. Stochastic gradient descent (SGD) serves as a popular optimizer in deep learning. It is a non-adaptive learning rate method. That is, the learning rate needs to be manually determined. In [8,11,13,14,15,16,17,18], they have improved the validation accuracy performance on plant diseases identification tasks by employing the SGD optimizer. However, beyond that, k-fold cross-validation, batch normalization, and dropout also have a positive impact on the performance in deep learning model training. The k-fold cross-validation methods solve the over-fitting problems [12], and the batch normalization method potentially helps in two ways: faster learning and higher overall accuracy [11,12], and the dropout operation [18] prevents over-fitting and improves the generalization ability. Performance gains achieved by different methods highlight the important role of the optimization algorithm in deep learning.

3. Non-Adaptive Learning Rate Methods

This paper contributes to the plant diseases identification by investigating non-adaptive learning rate techniques. A typical deep learning optimization task consists of minimizing the objective function f ( ω ) and fixing the best set of parameters. Non-adaptive learning rate methods heavily utilized in optimization problems to update the weights are the workhorse in literature. Inspired by classical and successful gradient descent methods, we focus on the non-adaptive learning rate methods. Therefore, this paper provides an expansion and improvement of SGDM for a more general and robust CNN model. A generic framework of non-adaptive learning rate methods is shown in Algorithm 1. This enables us to understand the rules of non-adaptive learning rate methods.
Algorithm 1 Generic framework of non-adaptive optimization methods
Require: x t R d , initial step size (learning rate) α , sequence of functions ϕ t , ψ t
      for  t = 1 to T
             g t = f t ( ω t )
             m t = ϕ ( g 1 , , g t ) and V t = I 2
             η t = α · m t
             ω t = ω t 1 η t
      endfor
Here, f t ( ω t ) is the gradient at ω t . For the sake of clarity, this paper summarizes the non-adaptive learning rate methods including Stochastic Gradient Descent (SGD) [19], Stochastic Gradient Descent with momentum (SGDM) [20], and Stochastic Gradient Descent with Nesterov momentum (NAG) [21] in Table 1. As observed in literature, there is a subtle difference between these non-adaptive learning rate methods in theoretical and implementation.

3.1. Stochastic Gradient Descent Momentum

The momentum is a typical non-adaptive learning rate technique, like SGD, which can achieve optimal convergence guarantees. The momentum technique modifies the SGD to accelerate convergence rate and to reduce oscillation. An update rule of SGD with momentum can be efficiently written as:
m t + 1 = β · m t + ( 1 β ) · g t
ω t + 1 = ω t α m t + 1
where a new hyper-parameter β [ 0 , 1 ) called the momentum parameter is an exponential discount factor. It determines how quickly the momentum buffer m t is updated and the variance of a normalized momentum buffer.
In SGDM, the update rule can also be written as:
ω t + 1 = ω t α [ ( 1 β ) · i = 0 t β i · g t i ]
Definition 1.
For β ( 0 , 1 ) , this paper defines the exponential discount function δ E X P , β as:
δ E X P , β ( i ) = ( 1 β ) β i
Definition 2.
For a discount function δ and a sequence of vectors x R d , we define a discounted sum D S δ ( x ) as:
D S δ ( x ) = i = 0 t δ ( i ) · x t i
when i = 0 t δ ( i ) = 1 for all t 0 ; this paper calls it a discounted sum average, and the exponentially weighted moving average E W M A β ( x ) is:
E W M A β ( x ) = D S δ E X P , β ( x ) = ( 1 β ) · i = 0 t β i · x t i
EWMA can be viewed as a weighted average method to estimate the expectation of random variable x = x 0 x t . The theoretical above indicates that the momentum buffer m t is precisely an exponentially weighted moving average, viz., m t = E W M A β ( 0 t ( ω 0 t ) ) .

3.2. The Proposed Method: Discount Momentum Optimizer

Inspired by EWMA, this paper extends the SGDM method to provide a distinct improvement in performance. The proposed algorithm can be regarded as a simple modification of the SGDM. Here, the details of the modifies are illustrated as follows:
Definition 3.
Similarity, the equation of the proposed discount function δ D M , v , β ( i ) and the discount weighted moving average D W M A D M , v , β ( x ) are shown as follows:
δ D M , μ , λ ( i ) = ( 1 λ ) λ i = 1 μ λ i = 0 μ ( 1 λ ) λ i i > 0
D W M A D M , μ , λ ( x ) = D S δ D M , μ , λ ( x )
= ( 1 μ ) · x 0 + μ ( 1 λ ) i = 0 t λ i · x t i
where discount momentum hyper-parameters μ R and λ [ 0 , 1 ) are constant.
Apply the update rule to the proposed algorithm:
ω t + 1 = ω t α [ ( 1 μ λ ) · g t + μ ( 1 λ ) i = 1 t λ i · g t i ]
Like SGDM, the update rule can also be equivalently written as
m t + 1 = λ · m t + ( 1 λ ) · g t
ω t + 1 = ω t α · [ ( 1 μ ) · g t + μ · m t + 1 ]
This suggested that the proposed discount momentum (DM) method is a simple modification of exponentially weighted moving average. On the condition of momentum hyper-parameter μ = 1 , the discount momentum (DM) is precisely the SGDM.

4. Results

In this section, we present our empirical study on the performance of DM method and compare it with other non-adaptive learning rate methods on plant diseases identification tasks in terms of training performance and generalization. We separate experiments into those with hyper-parameter tuning and those with CNN architectures. In the experiment, training occurs over 90 epochs (minibatch size 64). We apply the learning rate decay schedule by a factor of 0.1 at 30 epochs’ stepsize, which is commonly used in literature [22]. Each training run uses dual GPU (2*RTX 2080Ti).

4.1. The Dataset

A publicly-available and well-known database, plantvillage, is used for the training and testing of CNNs models. The plantvillage dataset contains 54,306 color leaf images with a uniform background and has 38 crop-disease pairs. These 38 classes comprise 14 crop plants and 26 different healthy or diseased plants. Some randomly selected images are shown in Figure 1. In our study, the images are divided into train and test subsets in an 80/20 ratio. It means that the training set contains 80% (43,810 images) of the total images and the remaining 20% (10,495 images) are used for the test data. In these non-adaptive learning rate approaches, we perform both model training and parameters’ optimization on these images.

4.2. Hyper-Parameter Tuning

Hyper-parameter tuning has a great influence on the quality of optimization for deep convolutional neural networks [23]. In this section, we discuss the discount momentum hyper-parameters μ and λ sensitivity in an image classification task. We set μ { 0.0 , 0.9 } and λ { 0.9 , 0.99 , 0.999 } [24]. Generalization error under several hyper-parameters setting are presented in Figure 2. As observed in sensitivity experiments, there is little difference between these discount momentum hyper-parameters μ and λ settings. Therefore, DM holds a low sensitivity to hyper-parameters.

4.3. Convolutional Neural Networks

In the experiment, the DM method is applied to a variety of models. ResNet and DenseNet are typical convolutional neural networks’ architectures, which are efficient and widely-used in literature. We consider testing the task of plant diseases classification with 50-layer ResNet and the 121-layer DenseNet. We select DM as the baseline algorithm and include comparisons with SGD, SGDM, and NAG non-adaptive learning rate methods. For DM, SGD, SGDM, and NAG, the first 30 epochs use learning rate α = 1.0 , the next 30 epochs use α = 0.1 , and the final 30 epochs use α = 0.01 . For SGDM and NAG, the momentum parameter β is directly applied to default value 0.9.
Plantvillage-ResNet50 ResNet50 has 50 layer deep CNNs with skip connections for image classification. We test our algorithm with the ResNet50 model on the plantvillage dataset. We compare the performance of DM, SGD, SGDM, and NAG. The results are shown in Figure 3, from which we can see that the DM algorithm is significantly better than SGDM and NAG. We notice that the training speed and generalization performance of DM are relatively superior SGD at the initial 30 epochs. In the later, DM and SGD share competitive results, while DM is generally slightly better.
Plantvillage-DenseNet121 DenseNet121 is a 121-layer deep CNNs with dense connections. Results of this experiment are reported in Figure 3. As is expected, the overall performance of each algorithm on ResNet50 is similar to that on DenseNet121. We can see that the DM method performs better than the non-adaptive ones in training. In addition, compared with non-adaptive learning rate methods, it converges as fast as SGD and achieves a bit higher accuracy on the plant diseases identification task.

5. Discussion

With the widespread use of deep learning solutions in plant diseases identification tasks, some limitations on model training have been highlighted. These issues mainly include plant diseases images in different categories are unevenly distributed, the diversity of images is low, and the complexity of the network model is increased. These problems caused an increased training time, poor convergence, generalization performance, and low recognition accuracy, which restricts the popularization of deep learning solutions in disease recognition tasks. Thereby, the thesis undertakes a study on the network parameter optimization aspect and proposes a DM method to enhance the overall performance of CNNs. Actually, the proposed algorithm, referred to as discount momentum, is a variant of the SGDM method. Compared with the results obtained in DM with the current state-of-the-art non-adaptive learning rate algorithms, the DM method has improved the performance of the model in identification accuracy, convergence speed, and parameter sensitivity. In addition, DM can recover SGD, SGDM, and NAG methods by assigning parameters. Here, we discuss the updated rules of these non-adaptive learning rate methods and their relationship with DM. The first is the SGD optimization algorithm. The SGD optimizer is heavily used in deep learning and performs well across image recognition domains in spite of its simplicity [25]. It considers mini batches to compute the unbiased estimate of the expected gradient. At each iteration, m t = g t . The network parameters are updated by
ω t + 1 = ω t α · g t + 1
where, when the discount momentum hyper-parameter μ = 0 , the updates of parameter ω t in the DM method are precisely (12). Therefore, the DM method can recover the SGD method with μ = 0 .
Next, we have the SGDM algorithm mentioned in Section 3.1. The details are shown in Section 3.1. Comparing the update rules of DM and SGDM, it is not difficult to find that, when the discount momentum hyper-parameter μ = 1 , the DM optimizer is precisely the SGDM method.
Last but not the least is the NAG algorithm. NAG is provided as a variant of the SGDM method. It can achieve a global convergence rate for general smooth convex functions. NAG takes inspiration from Nesterov’s accelerated gradient method and slightly ahead in the measure of loss function gradient of the momentum [26]. In fact, NAG replaces the m t term of SGDM by using the [ ( 1 λ ) · g t 1 + λ · m t ] . Therefore, the parameters are updated by
ω t + 1 = ω t α [ ( 1 λ ) · g t + λ · m t + 1 ]
The DM method recovers the NAG method with discount momentum hyper-parameter μ equal to hyper-parameter λ .

6. Conclusions

On the basis of an analysis of theoretical and experiments, we provide overwhelming evidence to the claim that the proposed algorithm is feasible to spread in deep learning fields. In non-adaptive learning rate methods, there are difficulties in the selection of hyper-parameters. This caused poor model performance and training to be difficult. We discussed the hyper-parameters setting and found that there is little accuracy performance change. Therefore, the DM algorithm is less sensitive to the change of hyper-parameters. We also discussed the adaptability of the DM method in different deep learning models. ResNet and DenseNet are typical convolutional neural networks architectures, which are representative in network layers and model architecture. We considered testing the task of plant diseases classification with 50-layer ResNet and 121-layer DenseNet. Results showed that DM has higher accuracy and is independent of the model, which is superior to state-of-the-art non-adaptive learning rate methods. We hope it is useful for the development of smart agriculture. Further studies should be needed to verify the applicability of the proposed algorithm in field experiments. In the future, we hope to integrate the proposed method into the mobile client and apply it to the field environment.

Author Contributions

Conceptualization, Y.S.; methodology, Y.S.; validation, Y.L. and H.Z.; investigation, Y.L.; data curation, Y.L. and H.H.; writing—original draft preparation, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

The subject is sponsored by the National Natural Science Foundation of P. R. China (No. 61872196, No. 61872194, No. 61902196, No. 62102194 and No. 62102196), Scientific and Technological Support Project of Jiangsu Province (No. BE2019740, No. BK20200753 and No. 20KJB520001), Major Natural Science Research Projects in Colleges and Universities of Jiangsu Province (No. 18KJA520008), Six Talent Peaks Project of Jiangsu Province (No. RJFW-111), Postgraduate Research and Practice Innovation Program of Jiangsu Province (No. KYCX19_0909, No. KYCX19_0911, No. KYCX20_0759, No. KYCX21_0787, No. KYCX21_0788 and No. KYCX21_0799).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Harvey, C.A.; Rakotobe, Z.L.; Rao, N.S.; Dave, R.; Razafimahatratra, H.; Rabarijohn, R.H.; Rajaofara, H.; MacKinnon, J.L. Extreme vulnerability of smallholder farmers to agricultural risks and climate change in Madagascar. Philos. Trans. R. Soc. B Biol. Sci. 2014, 369, 1639. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Gimenez-Gallego, J.; Gonzalez-Teruel, J.D.; Jimenez-Buendia, M.; Toledo-Moreo, A.B.; Soto-Valles, F.; Torres-Sanchez, R. Segmentation of Multiple Tree Leaves Pictures with Natural Backgrounds using Deep Learning for Image-Based Agriculture Applications. Appl. Sci. 2019, 10, 202. [Google Scholar] [CrossRef] [Green Version]
  3. Ruder, S. An Overview of Multi-Task Learning in Deep Neural Networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  4. Wu, Y.; Xu, L. Crop Organ Segmentation and Disease Identification Based on Weakly Supervised Deep Neural Network. Agronomy 2019, 9, 737. [Google Scholar] [CrossRef] [Green Version]
  5. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  6. Zhang, X.; Han, L.; Dong, Y.; Shi, Y.; Huang, W.; Han, L.; González-Moreno, P.; Ma, H.; Ye, H.; Sobeih, T. A Deep Learning-Based Approach for Automated Yellow Rust Disease Detection from High-Resolution Hyperspectral UAV Images. Remote Sens. 2019, 11, 1554. [Google Scholar] [CrossRef] [Green Version]
  7. Moreno-Revelo, M.Y.; Guachi-Guachi, L.; Gómez-Mendoza, J.B.; Revelo-Fuelagán, J.; Peluffo-Ordóñez, D.H. Enhanced Convolutional-Neural-Network Architecture for Crop Classification. Appl. Sci. 2021, 11, 4292. [Google Scholar] [CrossRef]
  8. Coulibaly, S.; Kamsu-Foguem, B.; Kamissoko, D.; Traore, D. Deep neural networks with transfer learning in millet crop images. Comput. Ind. 2019, 108, 115–120. [Google Scholar] [CrossRef] [Green Version]
  9. Mohanty, S.P.; Hughes, D.P.; Salath, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1–9. [Google Scholar] [CrossRef] [Green Version]
  10. Ghazi, M.; Yanikoglu, B.; Aptoula, E. Plant identification using deep neural networks via optimization of transfer learning parameters. Neurocomputing 2017, 235, 228–235. [Google Scholar] [CrossRef]
  11. Too, E.C.; Yujian, L.; Njuki, S.; Yingchun, L. A comparative study of fine-tuning deep learning models for plant disease identification. Comput. Electron. Agric. 2018, 161, 272–279. [Google Scholar] [CrossRef]
  12. Sambasivam, G.; Opiyo, G.D. A predictive machine learning application in agriculture: Cassava disease detection and classification with imbalanced dataset using convolutional neural networks. Egypt. Inform. J. 2020, 22, 27–34. [Google Scholar] [CrossRef]
  13. Cruz, A.C.; Luvisi, A.; De Bellis, L.; Ampatzidis, Y. X-FIDO: An Effective Application for Detecting Olive Quick Decline Syndrome with Deep Learning and Data Fusion. Front. Plant Sci. 2017, 8, 1714. [Google Scholar] [CrossRef] [PubMed]
  14. Zhang, S.; Zhang, S.; Zhang, C.; Wang, X.; Shi, Y. Cucumber leaf disease identification with global pooling dilated convolutional neural network. Comput. Electron. Agric. 2019, 162, 422–430. [Google Scholar] [CrossRef]
  15. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  17. Gehring, J.; Auli, M.; Grangier, D.; Yarats, D.; Dauphin, Y.N. Convolutional Sequence to Sequence Learning. In Proceedings of the 34th International Conference on Machine Learning, Sydney, NSW, Australia, 6–11 August 2017; pp. 1243–1252. [Google Scholar]
  18. Zhang, X.; Qiao, Y.; Meng, F.; Fan, C.; Zhang, M. Identification of Maize Leaf Diseases Using Improved Deep Convolutional Neural Networks. IEEE Access 2018, 6, 30370–30377. [Google Scholar] [CrossRef]
  19. Herbert, R.; Sutton, M. A Stochastic Approximation Method; Springer: New York, NY, USA, 1985; Volume 22. [Google Scholar]
  20. Polyak, B.T. Some methods of speeding up the convergence of iteration methods. Ussr Comput. Math. Math. Phys. 1964, 4, 1–17. [Google Scholar] [CrossRef]
  21. Nesterov, Y.E. A method for solving the convex programming problem with convergence rate. Dokl. Akad. Nauk Sssr 1983, 269, 372376. [Google Scholar]
  22. Luo, L.; Xiong, Y.; Liu, Y.; Sun, X. Adaptive Gradient Methods with Dynamic Bound of Learning Rate. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019; pp. 315–323. [Google Scholar]
  23. Ma, J.; Yarats, D. Quasi-hyperbolic momentum and Adam for deep learning. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  24. Zhou, Z.; Zhang, Q.; Lu, G.; Wang, H.; Zhang, W.; Yu, Y. AdaShift: Decorrelation and Convergence of Adaptive Learning Rate Methods. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  25. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1139–1147. [Google Scholar]
  26. Johnson, R.; Zhang, T. Accelerating Stochastic Gradient Descent using Predictive Variance Reduction. In Proceedings of the Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems, Christopher J. C. Burges and Léon Bottou and Zoubin Ghahramani and Kilian Q. Weinberger, Lake Tahoe, NV, USA, 5–8 December 2013; pp. 315–323. [Google Scholar]
Figure 1. Randomly selected images from the plantvillage dataset.
Figure 1. Randomly selected images from the plantvillage dataset.
Applsci 11 09468 g001
Figure 2. Hyper-parameters μ and λ sensitivity experiments with DenseNet121 on the plantvillage dataset.
Figure 2. Hyper-parameters μ and λ sensitivity experiments with DenseNet121 on the plantvillage dataset.
Applsci 11 09468 g002
Figure 3. Additional empirical results on convolutional neural networks. (The first figure on the left is ResNet50-plantvillage train accuracy; the first figure on the right is ResNet50-plantvillage train loss; the second figure on the left is ResNet50-plantvillage validation accuracy; the second figure on the right is Densenet121-plantvillage train accuracy; the third figure on the left is Densenet121-plantvillage train loss; and the third figure on the right is Densenet121-plantvillage validation accuracy, respectively).
Figure 3. Additional empirical results on convolutional neural networks. (The first figure on the left is ResNet50-plantvillage train accuracy; the first figure on the right is ResNet50-plantvillage train loss; the second figure on the left is ResNet50-plantvillage validation accuracy; the second figure on the right is Densenet121-plantvillage train accuracy; the third figure on the left is Densenet121-plantvillage train loss; and the third figure on the right is Densenet121-plantvillage validation accuracy, respectively).
Applsci 11 09468 g003
Table 1. An overview of non-adaptive learning rate methods.
Table 1. An overview of non-adaptive learning rate methods.
SGDSGDMNAG
m t + 1 = g t + 1 m t + 1 = β m t + ( 1 β ) g t g t + 1 = f ( ω t + 1 α · m t / V t )
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, Y.; Liu, Y.; Zhou, H.; Hu, H. Plant Diseases Identification through a Discount Momentum Optimizer in Deep Learning. Appl. Sci. 2021, 11, 9468. https://doi.org/10.3390/app11209468

AMA Style

Sun Y, Liu Y, Zhou H, Hu H. Plant Diseases Identification through a Discount Momentum Optimizer in Deep Learning. Applied Sciences. 2021; 11(20):9468. https://doi.org/10.3390/app11209468

Chicago/Turabian Style

Sun, Yunyun, Yutong Liu, Haocheng Zhou, and Huijuan Hu. 2021. "Plant Diseases Identification through a Discount Momentum Optimizer in Deep Learning" Applied Sciences 11, no. 20: 9468. https://doi.org/10.3390/app11209468

APA Style

Sun, Y., Liu, Y., Zhou, H., & Hu, H. (2021). Plant Diseases Identification through a Discount Momentum Optimizer in Deep Learning. Applied Sciences, 11(20), 9468. https://doi.org/10.3390/app11209468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop