Next Article in Journal
Taxonomy of Anomaly Detection Techniques in Crowd Scenes
Previous Article in Journal
Asymmetric Arc Routing by Coordinating a Truck and Multiple Drones
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification

1
Faculty of Computing, Universiti Malaysia Pahang, Kuantan 26300, Pahang, Malaysia
2
Department of Computer Science and Engineeering, Islamic University, Kushtia 7003, Bangladesh
3
Faculty of Computing and Informatics, Multimedia University, Cyberjaya 63100, Selangor, Malaysia
4
Faculty of Electrical and Electronics Engineering Technology, Universiti Malaysia Pahang, Pekan 26600, Pahang, Malaysia
5
Malaysian Institute of Information Technology, Universiti Kuala Lumpur, Kuala Lumpur 50250, Selangor, Malaysia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(16), 6079; https://doi.org/10.3390/s22166079
Submission received: 16 June 2022 / Revised: 7 July 2022 / Accepted: 11 July 2022 / Published: 14 August 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Automatic leaf disease detection techniques are effective for reducing the time-consuming effort of monitoring large crop farms and early identification of disease symptoms of plant leaves. Although crop tomatoes are seen to be susceptible to a variety of diseases that can reduce the production of the crop. In recent years, advanced deep learning methods show successful applications for plant disease detection based on observed symptoms on leaves. However, these methods have some limitations. This study proposed a high-performance tomato leaf disease detection approach, namely attention-based dilated CNN logistic regression (ADCLR). Firstly, we develop a new feature extraction method using attention-based dilated CNN to extract most relevant features in a faster time. In our preprocessing, we use Bilateral filtering to handle larger features to make the image smoother and the Ostu image segmentation process to remove noise in a fast and simple way. In this proposed method, we preprocess the image with bilateral filtering and Otsu segmentation. Then, we use the Conditional Generative Adversarial Network (CGAN) model to generate a synthetic image from the image which is preprocessed in the previous stage. The synthetic image is generated to handle imbalance and noisy or wrongly labeled data to obtain good prediction results. Then, the extracted features are normalized to lower the dimensionality. Finally, extracted features from preprocessed data are combined and then classified using fast and simple logistic regression (LR) classifier. The experimental outcomes show the state-of-the-art performance on the Plant Village database of tomato leaf disease by achieving 100%, 100%, 96.6% training, testing, and validation accuracy, respectively, for multiclass. From the experimental analysis, it is clearly demonstrated that the proposed multimodal approach can be utilized to detect tomato leaf disease precisely, simply and quickly. We have a potential plan to improve the model to make it cloud-based automated leaf disease classification for different plants.

1. Introduction

The detection of plant diseases is the foundation for crop disease prevention and crop quality assurance. Traditional plant disease detection systems rely heavily on human observation, resulting in low detection efficiency, generability and reliability. Farmers with a lack of technical competence and agricultural professionals are unable to serve the field at all times, causing them to overlook the most effective preventative opportunities. In recent times, image processing [1], pattern recognition [2,3], computer vision [4,5] and text [6,7,8,9] or video [6,10] analysis have fast advanced outcomes in recent years. A machine learning-based intelligent disease detection approach provides a means for effectively resolving agricultural concerns [11,12].
Tomato is considered the most significant and widely consumed crop after potato [13]. The farmland area used for horticultural crops has expanded by 164% in the last four decades. Recently, worldwide tomato production reached over 180 million tonnes [14,15]. China by far is the world’s top tomato grower, accounting for 31% of worldwide tomato yield. However, every year a vast amount of tomato plants are affected by numerous leaf diseases. This results in lower crop production, which affects human health and livelihood, as well as financial stability [16]. Therefore, it is very crucial to detect tomato leaf disease accurately in the early stage to decrease crop losses and assure optimal growth. Plant leaf diseases are traditionally classified by trained experts by doing a visual investigation of plant leaf tissues. These conventional techniques are very time-consuming, expensive, and expertise-dependent [17]. In recent years, machine-learning-based recognition techniques for crop disease identification became popular for their successful application [18]. Moreover, in the field of computer imaging, the conventional deep learning (DL) algorithms are the most widely used approach for the automated identification of plant diseases [17]. Neural Networks [19], K-nearest Neighbors (k-NN) [20], Niave Bayes [21], Logistic Regression [22], Random Forest (RF) [23], Support Vector Machines (SVM) [24], as well as adaptive boosting [20] are perhaps the most prominent conventional computer-based techniques used for plant disease categorization. Traditional machine learning approaches rely significantly on the features individuals provided. These traits are painstakingly extracted by the expert, making these procedures costly, as well as time-consuming. On the other hand, the limitations of the handcrafted features technique can be readily overcome with the automatic extraction of features utilizing deep learning (DL). Therefore, the DL methods are widely employed for plant disease categorization [25] because of its high performance. Among the wide range of DL methods, a deep convolutional neural network (CNN) is most widely used for crop disease detection [26]. An improved YOLOv5 algorithm of deep learning has been used in the detection of plant diseases [27]. However, the deep learning methods require a large amount of training data and a considerable amount of time for training. On the other hand, the performance of standard machine learning can be outperformed compared to deep learning whilst providing a limited amount of training data. Moreover, in terms of computing resources, deep learning is more expensive than traditional machine learning techniques. Actually, the machine learning algorithm is able to perform faster classification and deep learning performs well in feature extraction and classification.
For these considerations, this study has combined the DL with the ML method. In the initial stage, all image are preprocessed (Filtering and Segmentation) to generate the synthetic image using CGAN. At first, we used the attention-based dilated CNN to extract the feature from the tomato leaf disease images. During the feature extraction process, the hidden layer in deep learning models allows them to learn hierarchical representations. Deep architectures can select discriminating representations from model training that aid with exact predictions based on the training data in later classification phases. Finally, the ML method LR (Logistic Regression) is utilized to classify the extracted feature. The proposed tomato leaf disease detection approach is divided into three phases: fixed feature extractor using attention-based dilated CNN as fusion features, dimensionality reduction using normalization, and training of a logistic regression classifier. Furthermore, we also developed Logistic Regression (LR), CNN with LR (CNN-LR), and the proposed attention-based dilated CNN with LR (ADCLR) model to test the robustness and applicability of the proposed tomato leaf detection approach. The main contribution of this study is as follows:
  • In this study, we have introduced sequential image pre-processing steps. The tomato leaf images have been pre-processed using the color conversion, filtering method for denoising the images. To handle the larger features, we have used the bilateral filter method which helps to make images smoother with fine spatial parameters. Furthermore, the noises have been removed from the filtered data using the fast and simple Otsu segmentation method. Then, we use the CGAN model to generate synthetic image from the image to handle imbalance and noisy or wrongly labeled data to obtain good prediction results.
  • To extract the most informative feature in a short time, we have designed a lightweight dilated CNN architecture and attention mechanism in which the multiple hidden layers of the architecture allow them to learn hierarchical representations from the images. Then, the extracted features have been classified using the fast and simple logistic regression model.
  • To check the validation and robustness of the proposed hybrid architecture, we have also implemented eleven popular transfer learning algorithms on the same dataset and compared the performance with the proposed ADCLR model. The experimental analysis clearly shows that the proposed hybrid ADCLR provides superior performance for detecting tomato leaf disease.
The basic flowchart of our proposed approach is shown in Figure 1. At first, the input images are preprocessed using filtering and segmentation methods. Then, we use the CGAN model to generate synthetic image. Synthetic image is generated to handle imbalance and noisy or wrongly label data to obtain good prediction results. Then, the synthetic images are used in the attention-based dilated CNN layer to extract the advanced features. Finally, the logistic regression model is used to learn the extracted features and classify the tomato leaf disease images accurately.
The remainder of the paper is laid out as follows: Section 1 conducts a state-of-the-art survey and explains the study’s objective. The data and dataset preprocessing are introduced in Section 2. The experimental setup, technique, and development of all section of the proposed method are described in Section 3. The results are presented in Section 4, followed by an informative discussion, before the conclusion is presented in Section 5.

2. Related Study

Nowadays, intelligent approaches for plant leaf disease detection have shown great successful applications in early diagnosis. The researchers have developed several strategies for automatic plant disease classification. In recent years, machine learning or deep learning approaches are widely adopted methods for the early diagnosis of plant disease.

2.1. Machine Learning Methods

To obtain a high-performing model, the researchers have used several data preprocessing steps such as color conversion, edge-based segmentation, filtering and segmentation. Furthermore, image analysis, shape, size, augmentation, and color conversion are used to extract the feature from the segmented images. Then, traditional machine learning methods have been utilized to detect plant diseases efficiently [28].
For example, Hlaing et al. [29] developed a feature extraction approach based on the Johnson SB distribution and a scale-invariant feature transform (SIFT). The proposed approach was used to extract SIFT and color statistic features, which were then fed into a multi-class support vector machine classifier for categorization. For tomato disease categorization, the proposed approach got an accuracy score of 85.1%. In [30], a novel method based on the concatenation of various features was presented. The Hue moments, Haralick, and color histograms were extracted and then combined. For tomato leaf disease categorization, the retrieved feature was input into the decision tree as well as RF classifiers. They achieved a maximum of 94% classification accuracy with the random forest method. Kalyoncu et al. [31] proposed a unique plant leaf disease categorization approach based on numerous characteristics. A digital image of a leaf is used to extract the shape, textural and geometric, as well as color properties. In particular, sorted uniform LBP is presented as a novel local binary pattern (LBP) alternative for describing leaf texture. A machine learning method using the SVM algorithm got very low accuracy of 85.02% but this method is fast [29]. The discriminant classifier (LD) is being used to classify the data once it has been combined with the retrieved characteristics. This approach was tested on three different datasets: ICL, Flavia, and Swedish. The average accuracy of ICL, Flavia, and Swedish was 86.8%, 98.6%, and 99.5%, respectively. The authors of [32] proposed a semi-automated technique for soybean disease detection based on color and texture characteristics. In their approach, they used a total of 4775 images for classification with the SVM method (90% accuracy).

2.2. Deep Learning Methods

In the past couple of years, researchers have widely focused on deep learning methods due to their successful application for plant disease diagnosis. In this sub-section, tomato leaf disease detection-related studies based on the deep learning method have been highlighted. Batool et al. [33] presented a tomato leaf disease identification method. In their study, the AlexNet pre-trained model was used for feature extraction. Then, the kNN method was used to classify the extracted feature. This achieved a maximum of 76.1% testing accuracy. Another study [34] developed a transfer learning model (MobileNetV2) for tomato leaf disease classification. To improve the model performance, they utilized the fine-tuned strategy of the MobileNetV2 model, and achieved impressive performance (90% accuracy) with the fine-tuned model. Agarwal et al. [35] proposed a CNN approach for tomato disease diagnosis (91.2% accuracy). In [36], an inception method combined with dilated convolution was used to identify 26 diseases of 14 different crops. They achieved a maximum of 99.37% classification accuracy with the PlantVillage dataset. A novel CNN model with eight hidden layers was introduced in [37] for tomato plant disease detection. They achieved 98.4% classification accuracy with the PlantVillage dataset. The authors [38] proposed a nine-layer CNN model to detect 39 types of the plant leaf. To enhance the performance of their model, they used different data augmentation techniques and finally achieved 96.46% accuracy on the PlantVillage dataset of tomato leaf disease. Nithish et al. [39] developed a pre-trained deep learning method (ResNet-50). The ResNet-50 architecture was fine-tuned to successfully categorize the six classes of leaf disease and achieved a 97% average classification accuracy.

2.3. Deep Learning with Machine Learning

For the particular task classification to be more effective, the researchers have developed a hybrid model where they combined the ML-ML, ML-DL, or DL-DL methods. Due to the excellent feature learning capabilities of the DL methods, some recent studies used the DL layers to extract the feature from the data [40]. Several hidden layers of the DL model have the capabilities to select the discrimination feature more effectively. Fisher et al. [41] proposed an approach where the image’s features were extracted using the CNN network. The images were classified using the RPN and Fast-RCNN by constructed feature maps. Furthermore, VGG networks are commonly used with faster RCNN. ResNet [42] has more convolution layers than VGG16 and can use convolution to extract more object information. ResNet has a layer-skipping structure that allows it to skip one or more layers immediately. It addresses the issue of gradient disappearance caused by layer stacking. On the other hand, the VGG16 network is unable to extract detailed aspects of tomato leaf diseases [43]. A feed-forward neural network with such a residual connection is used to create the deep residual network. The identical mapping function of the skipping structure allows the output with one stage to be used as the source of a subsequent layer. The benefit of this method is that no other variables are imported, as well as the calculation time not being greatly increased. Gradient disappearance [44] induced by expanding the number of neurons in the hidden layer is prevented by employing cross-layer operations and reusing intermediary features. As a result, the deep residual network is crucial in the field of recognition. In addition, deep residual networks are frequently used in defect detection as well as fault-tolerant control [45,46]. Some studies have shown that the ML with DL approach provides slightly higher performance than conventional ML or DL method. For example, the MobileNet and NasNet feature extractor with Logistic Regression achieved 97% accuracy [28]. On the other hand, recently, the dilated CNN mechanism has become more popular because of its effective and fast feature learning capabilities [47].
Another attention-based method proposed by Devi et al. [48] that used the Salp Swarm Algorithm to classify tomato leaf disease. This method got 97.56% accuracy to predict five types of tomato leaf disease from plan village data. The limitation of this method is that its performance is not high and it faces some computational complexity. On the other hand, a method that utilized the Lightweight Attention-Based CNN mechanism [49] to classify ten types of tomato leaf disease has 99.34% accuracy but it has slightly higher time complexity than conventional methods such as CNN [37] and SVM [29]. In 2022, Zhao et al. [50] developed a method utilizing spatial attention with CNN for real time leaf disease detection. This method has 95.20% accuracy but this method needs to be adaptable by increasing its generability.

3. Materials and Methods

3.1. Data Description

In this experimental analysis, a well-known PlantVillage dataset crosscheck the citation [51] was used to detect the tomato leaf disease. Hughes and Salathe et al. [52] generated the PlantVillage collection, which includes 54,309 label images for 14 different species and 38 different types of healthy and leaf disease images. From the entire dataset, we used 15,989 images of tomato leaves grouped into ten classes. Then, the selected images were resized into 256 × 256 pixels. The resized images were then normalized. The normalized procedure helps to speed up the training procedure by reducing the computational complexity. To test the validation of the proposed architecture, we used a total of 1000 new images during testing (100 images for each class) from 54,309. The detailed data description is shown in Table 1. Figure 2 shows the basic image samples of tomato leaf disease.

3.2. Data Preprocessing

In the preprocessing phase, we added a label for each image depending on the prefix of their filename. Overall preprocessing task is presented sequentially in the Algorithm 1. At first, the images were segmented using Bilateral filtering and Ostu’s image segmentation process. Before the segmentation process, two-color converting functions were applied where the sequence of the image’s color-space was lost and only the brightness and saturation for each pixel were kept. After that, data normalization was accomplished by calculating the mean difference between each pixel and dividing the result by the standard deviation. The images were normalized to make it easier for the error function, which is typically not convex, to identify the global minimum. Decreasing the range of inputs for training variables also aids the backpropagation algorithm’s efficiency. Python’s random shuffle technique, which is a command algorithm based on an arbitrary number generator, was used to shuffle the data. The order of the images was originally sequential after the application, but it is now mixed throughout the collection.
A multilevel categorization is a strategy that uses more than two labels. Every label in this classification is not exclusive. For each sample, the classification method yields only one degree. In actuality, multi label classification of leaf disease is used to label categories of tomato leaf disease in one or more types. We focused on the image’s multilevel classification of tomato leaf disease. We conducted a multilevel study on the D data set. For further information on data collection, see Area F of this technical section. The data includes visual features of tomato leaf disease and each declaration set on the vector stage. The labeling of the tomato disease image is presented in Equation (1).
D = ( E , F ) | F I m a g e , E ( 0 , 1 ) L
Here, F is the Tomato leaf disease images, dataset E with L (Number of Tomato leaf disease class label) target disease class, and the Tomato leaf disease categories indicated are increased level L which is ten.
Algorithm 1 Preprocessing Algorithm for ADCLR Model
1:
Input: Dataset of tomato leaf disease
2:
Result:Preprocessing of tomato leaf disease image data, with number of image N
3:
for Each iteration i in range(0,N) do
4:
   Resize to dimension (256 × 256)
5:
   Pre-processing to resize the image (224 × 224)
6:
   Normalize pixel values [0, 1]
7:
   Standardize pixel values to (256 × 256)
8:
end for
9:
if Image set size is (224 × 224) then
10:
   Normalize pixel value [0, 1]
11:
   Standardize pixel values to (256 × 256)
12:
end if
13:
if Image set size is (224 × 224) then
14:
   Perform Image Filterng with Bilateral Filter
15:
   Complete Otsu segmentation
16:
end if
17:
Generation of Preprocessed image of tomato leaf

3.3. Image Filtering

The bilateral image filtering method is used in our method. In this filtering, any input image (a) is converted to a smoothed form by the bilateral filter (b). Then, most texture, noise, and small details are removed, but broad sharp edges are preserved without being blurred. A bilateral filter is a non-linear image smoothing filter that preserves edges while lowering noise [53]. It uses a weighted mean of intensity data from surrounding pixels to adjust the brightness of each pixel of the disease image. A Gaussian distribution could be used to calculate this weight. The weights are determined not just by the Euclidean distance between pixels, but also by the radiometric variations (color intensity, depth distance, etc.). Sharp edges are preserved as a result. The computation of Bilateral image filtering is utilized with the Equation (2).
The bilateral filter is defined as
F F i l t e r ( x ) = 1 W p [ C o c f i Ω F ( c o c f i ) f r ( ( | | F ( c o c f i ) F ( c o c f ) | | ) ) g s ( | | ( c o c f i c o c f ) | | ) ]
and normalization term, W p is defined as the Equation (3).
W p = C o c f i Ω f r ( | | I ( c o c f i ) F ( c o c f ) | | ) g s ( | | c o c f i c o c f | | )
Here, F F i l t e r indicates a filtered image of the tomato leaf disease data, I is considered as the original image to be filtered, cocf indicates the current pixel coordinate to filter, Ω presents a window centered in the c o c f i so c o c f i Ω is another pixel. f r indicates the smoothing intensity difference. When smoothing disparities in positions, g s is the spatiotemporal (or domain) kernel (this functional method can be a Gaussian distribution).
In the filtering of the image of tomato leaf disease, the spatial closeness (using spatially kernel g s ) and the intensity difference (using the range kernel f r ) are used to give the weight W p . Imagine a pixel at ( i , j ) that is needed to be denoised inside an image by using its neighbors, and one of its neighbors is at ( i , j ) ( k , l ) . Each assigned pixel ( k , l ) to remove the noise the pixel ( i , j ) is provided by the following Equations (4) and (5), presuming the range, as well as spatial kernels, are Gaussian kernels.
ω ( i , j , l ) = e x p ( ( i k ) 2 + ( j l ) 2 2 σ d 2 F ( i , j ) + F ( k , l ) 2 2 σ r 2 )
Here, σ d and σ r both indicate the smoothing parameters, and F ( i , j ) and F ( k , l ) presents intensity of pixels ( i , j ) and ( k , l ) .
F D ( i , j ) = k , l F ( k , l ) w ( i , j , k , l ) k , l w ( i , j , k , l )
Here, F D presents the denoised pixel intensity of the pixel ( i , j ) . The bilateral filter steadily approaches like Gaussian convolution relatively closely as the ranged parameter σ r grows. Actually, the range of Gaussian expands and compresses, which implies that it becomes nearly constant through the image’s intensity intervals. The larger features become smoother as the spatial parameter value of σ d is increased.

3.4. Image Segmentation

Object and boundary (lines, curves, etc.) inside images are often identified via the image segmentation method. Image segmentation is typically used to assign a label to each image pixel because pixels with nearly identical labels have similar characteristics. Image segments are used to reduce the complexities of an image, making subsequent processing and analysis easier. In layman’s terms, segmentation is the process of labeling pixels.
For this purpose, we have used the Otsu segmentation approach to segment the tomato leaf disease images. Automatic image thresholding is performed using Otsu’s [1]. This method generates a single intensity threshold that divides pixels into two different classes: foreground as well as background. This limit is set by reducing intraclass intensity variation. The technique looks for a threshold that minimizes intraclass variance, which is defined as the weighted combination of the classes’ variances as given in Equation (6).
σ P 2 ( t ) = P 0 ( t ) σ 0 2 ( t ) + P 1 ( t ) σ 1 2 ( t )
In the above Equation (6) P 0 and P 1 indicates the class probability with threshold value t difference and σ 0 2 and σ 1 2 presents variance.
Here, P 0 , 1 ( t ) is measured by the histogram bins L, o ( i ) indicates previous probability as computed in Equations (7) and (8):
P 0 ( t ) = i = 0 t 1 o ( i )
P 1 ( t ) = i = t L 1 o ( i )
The minimizing and maximizing of the intra-class variance is equivalent as computed in Equation (9).
σ b 2 ( t ) = σ 2 σ P 2 ( t ) = P 0 ( t ) ( μ 0 μ T ) 2 + P 1 ( t ) ( μ 1 μ T ) 2 = P 0 ( t ) P 1 ( t ) [ μ 0 ( t ) μ 1 ( t ) ] 2
Generally, Otsu’s technique initializes the value of P i and μ , then updates their levels based on threshold levels to obtain the intended σ b 2 ( t ) threshold. The Otsu threshold is indeed a fast as well as a simple algorithm that works with histograms (that are 256-element integer or float vectors).
Figure 3 clearly shows the preprocessed result image and feature extraction result using our method.

3.5. Synthetic Image Generation

This method uses the Conditional Generative Adversarial Network (CGAN) model to generate synthetic images from preprocessed images. To deal with imbalance and noisy or worn-out label data, a synthetic image is generated. This aids in training the model because it can predict well.
Initially, the Generative Adversarial Network (GAN) was developed in 2014 [54]. The GAN is built on the concept of having two neural networks fight in a zero-sum mechanism. Therefore, it is adversarial, in which the loss with one network benefits of another directly, as well as vice versa. To use picture generation for instance, there are distinct networks in this work. A generating network that makes images as well as a discriminator network that classifies the inputs as true or bogus. A gradient descent technique is used to update the gradients of each network after each train batch, as is the case with most deep learning approaches. Because the generator network’s output goes directly further into discriminator network, the combined networks’ training is automated through competition. A score can be determined using categorical cross-entropy as follows:
L o s s = E x × [ l o g ( D i s ( x ) ) ] + E z [ l o g ( 1 D i s ( G e n ( z ) ) ]
E x t e n d e d L o s s = E x × [ l o g ( D i s ( x / y ) ) ] + E z [ l o g ( 1 D i s ( G e n ( z / y ) ) ]
In the equation Equation (10), where ( E x × [ l o g ( D i s ( x ) ) ] ) indicates the recognition of real images and the ( E z [ l o g ( 1 D i s ( G e n ( z ) ) ) is used for the recognition of fake tomato leaf disease images. An additive noise input to a generator that starts as a real input from the dataset. The term D i s ( x ) is used to detect false photos E x or E z since it calculates the probability that a given piece of data is real. Because the discriminator’s source is Generator G’s result while confronted with a random source vector, z, D i s ( x ) is replaced by G e n ( z ) in the remaining half of the equation. Because the generator’s goal is to maximize or minimize the loss function with the Equation number (10), whereas the discriminator’s goal is to minimize it, this is regarded as the function to minimax loss.
A Conditional Generative Adversarial Network (CGAN) [54] is an extended version of the previous GAN model that works on given number of disease class. This mechanism of image augmentation is also used in fruit classification [55]. The generator now aims to learn to generate images belonging to one of ten classes of tomato leaf disease. Equation (10) is expanded as the Equation number (11), data objects and its label is ensured. As a result, G e n ( z / y ) is the result of the generator having random vector given class y labeling, and D i s ( x / y ) is the discriminator’s confidence that x is real provided class label y. The production of objects corresponding to several classes is enabled by the minute difference in topological from a GAN, as shown in Figure 4. When the dataset was fed into a conditional GAN, the system would learn to generate fake tomato leaf disease images by training on actual tomato leaf disease photos, requiring two networks to produce either class. The networks in question would also need to train independently.

3.6. Proposed Hybrid Classification Model

In this study, we have designed a hybrid deep learning model (ADCLR) to detect tomato leaf disease more efficiently. A new hierarchical attention network with a dilated convolutional neural network (CNN) is used with multiclass in our tomato leaf disease image categorization technique. At first, we take the publicly available dataset, followed by preprocessing of all data and then synthetic image generation. The outcome of the preprocessing is then sent to a new vector routing algorithm for extracting feature information from the deep layer of the dilated CNN and then to the attention layer. Finally, the Logistic Regression Classifier is used to classify the extracted features. Our proposed model used categorical cross-entropy with the Adam optimizer and ROC assessment approach. A multilevel disease identification module incorporates these layers. The overall sequential process flow chart of our model is shown in Figure 5. The main Algorithm 2 of our ADCLR method describes the overall process clearly.
Algorithm 2 General Algorithm of our ADCLR Model
1:
Input: Dataset of tomato leaf disease STATE Output: Prediction of tomato leaf disease image
2:
Initialization: N:= Number of image, C:= Class label
3:
for Each iteration i in range(0,N) do
4:
   Take each image i
5:
   Pre-processing to resize the image i as the resolution of (224 × 224)
6:
   Normalize pixel values [0, 1] of the image i
7:
   Standardize pixel values to (256 × 256) of the image i
8:
   Image filtering with bilateral filtering method of the image i
9:
   Segmentation of image with Otsu method of the image i
10:
     Generation of preprocessed image
11:
end for
12:
for Each iteration j in range(0,N) do
13:
  Take preprocessed image j
14:
  CGAN based Generation of synthetic image j of tomato leaf with disease label C
15:
end for
16:
Initialization of parameter
17:
DCk:= Extracted feature by Dilated CNN
18:
A k : = E x t r a c t e d r e l e v a n t f e a t u r e b y A t t e n t i o n m e c h a n i s m
19:
for Each iteration k in range(0,N) do
20:
   T a k e s y n t h e t i c i m a g e S k a n d i t s c l a s s l a b e l C
21:
   U s e d i l a t e d C N N t o e x t r a c t f e a t u r e s D C k o f t h e i m a g e s k
22:
  Use attention mechanism to capture relevant feature  A k from dilated CNN feature D C k
23:
  Use logistic regression classifier to classify image with features  A k to the target classC
24:
  Generate prediction score of each image for target class C
25:
end for
26:
Evaluate all the prediction of tomato leaf disease image
The following are the technical strengths of our proposed study:
Our method has the following steps:
  • Initially, the inputted data is preprocessed by color conversion, filtering, and denoising. Bilateral filtering is used and can handle larger features to make the image smoother with fine spatial parameters. Noise from the preprocessed filter is also removed by the fast and simple Otsu segmentation method.
  • Then, we use the Conditional Generative Adversarial Network (CGAN) model to generate synthetic image from the image those are preprocessed in previous stage. The synthetic image is generated to handle imbalance and noisy or wrongly labeled data to obtain good prediction results.
  • Then, the synthetic image is sent to our proposed ADCLR model. In the ADCLR model, the attention-based Dilated CNN is used to extract the informative feature extraction. Dilated convolution has the advantage of capturing the level of internal sequence data first by increasing the region of the convolution kernel without raising the model’s parameter amount. The attention layer simply concentrates on the memory block, instead of focusing on the entire feature space, attention mechanism has the benefit of dramatically reducing the number of parameters and sharing the weights among diverse regional places.
  • After that, the ADCLR method is trained with the training dataset and it tests the model robustness with the validation dataset. The Logistic Regression classifier is used to classify the images based on the extracted feature. Logistic regression classifier is simple, takes less time in training, and it performs well in multiclass prediction.
  • Finally, the validation of the proposed model is tested with different performance evaluation metrics and comparison on disease image.
  • To test the validation and effectiveness of the proposed approach, we also implemented eleven popular deep learning methods with the dataset, whereas our proposed method shows superior performance.

3.6.1. Dilated CNN Layer for Feature Extraction

In this phase, a few feature variables such as color features are regarded because their visual color difference recognizes whether the plant leaf is exposed to the virus by the disease or not after a human perception in the precise system. In the Dilated CNN network, the multiple hidden layers allow the model to learn the discriminatory feature more efficiently. Deep learning, unlike machine learning, learns leaves with diseases and classifiers automatically, resulting in the machine learning’s efficiency in such contexts. The methodology maintains a strategy from the demand for humans to ideas by accumulating information from experience, empowering the computer to understand complicated ideas by creating them out of smaller complexes. The outputs of the multiple layer levels of a dilated convolutional neural network and attention layer are responsible for feature extraction and selection, as shown in Figure 6.
In the ADCLR model, this stage is one of the most crucial stages. The dilated convolution layer’s deep depth tries to find hierarchical, granular quality features that can be used to describe compositional feature information. Our feature findings are pooled and delivered to a Dilated CNN layer to produce DCV output, unlike typical CNNs, which perform dilated convolution operations instantly. For the few blocks of convolution given in Figure 6. Operational steps of the dilated CNN is given in the feature extraction Algorithm 3. In Figure 6a, each green colored dot indicates that this block is the block where selected convolution is performed. We define it as follows. As a consequence, the deep CNN layer generates the following set of variables as the Equation (12):
d c v i = d c o n v 1 , d c o n v 2 , , d c o n v n R n × d
here, d c v is the output of the dilated convolution and d indicates the input sequence as (13).
D C V 1 = [ d c v 1 l , d c v 2 l , , r v n l ] R n d , l ( 1 , L )
In this Equation (13), L is the overall convolutional box, and the blocks filter has a degree of k. Let us focus on block l-th number alone.
W l R k w k , W l R k w d
This filtering matrix W employs that the operation held in k time, as well as weight w vector. The two adjacent blocks can be changed as seen in Equation (15) below.
D C V = F ( W 1 , D C V l 1 )
It is just a sliding of filtering with a window used to w-length input, where f represents linear algebra expression. Normally, d c v 1 l D C V l is computed as the Equation (18).
d c v t l = R e L U ( W l [ d c v t + 1 r l 1 ] t = 0 w 1 )
Here, the ⊕ sign indicates convolution, r presents the level of the deep layer in the dilation. The ReLU with all blocks has a length of ( w 1 ) 2 L 1 . A conventional deep convolution layer that raises it exponentially rather than increasing weights of the parameters of the network layer. Finally, hierarchical maps of D C V 1 , D C V 2 , , D C V l were obtained based on the coupling coefficients relation on upstream and downstream layers. SoftMax gives the value of the b i o set. Currently, D C V 1 = [ d c v 1 l , d c v 2 l , , d c v n l ] R n k , l ( 1 , L ) Here, l-th convolutional block output given as R n k , l ( 1 , L ) . Each k filter operation output is generated as dcv. C O V i o ’s value used as final output features. Now obtain C O V 1 = [ C O V 1 l , C O V 2 l , , C O V n l ] R M d V . The convolution terms size is dv and M indicates the amount of final convolution. Now, perform the routing D C V l to C O V l for final feature extraction and information generation. The predicting vector d c v ˜ J | l indicates the raw vector feature transformation that is calculated as the multiplication of d e v i with W j in (17).
d c v ˜ J | l = d c v i W j
By increasing small vectors and decreasing large vectors into unit vectors, this strategy improves the information exchange efficiency of the complicated routing method. To compute the medium step, we used an iteratively layered routing strategy over multilayered dilated convolution layers.
Here, softmax routing function is s r f i j and its modification with d c v is set to a i j agreement. This computation has used Equation (18).
a i j = c o v i d c v ˜ J | l + s r f i j
Typically, the dilated convolution operation enables more efficient, as well as scalable, convolution routing. Algorithm 3 describes the hierarchical routing scheme of the ADCLR model. In this stage, the autonomous final convolution layer computed as C O V 1 = [ C O V 1 l , C O V 2 l , , C O V n l ] R M d V . In a unified C O V , we summed the results of the final convolution in (19).
C O V 1 = [ C O V l , C O V l , , C O V l ]
The action will be passed throughout the hierarchical layer after it is performed. Extracted features [ C O V l , C O V l , , C O V l ] of dilated convolution will be assigned.
Algorithm 3 Feature Extraction Algorithm of DCLR Model
  • Data: Input synthetic image Tomato disease with calss label
  • Result : Extract features Fi fromTomato leaf disease data
  • Preprocess Image j to analysis
  • Process image features
  • : X = x 1 , x 2 , . . . . , x n R n d  here d is dimension
  • Get dilated convolutional output
  • Process iteration in dilated CNN
  • foreach iteration l in range(0, l)do
  •     for each iteration n in range(0, N) do
  •         D C V 1 = [ d c v 1 l , d c v 2 l , . . . . , r v n l ] R n d , l ( 1 , L )
  •        for w in range(0,w) do
  •             d c v t l = R e L U ( W l [ d c v t + 1 r l 1 ] t = 0 w 1 )
  •             W l R k w k , W l R k w d
  •            Here L is number of layers
  •        end for
  •     end for
  • end for
  • Process Dynamic convolution network
  • foreach iteration i in range(0, N)
  • do
  •     for each iteration j in range(0, N)
  •     do
  •         C O V j = i = 1 m c o v i j d c v ˜ J | l
  •         d c v ˜ J | l = d c v i W j
  •         a i j = C O V i d c v ˜ J | l + s r f i j
  •     end for
  • end for
  • Execute hierarchical attention mechanism
  • for each iteration i in range(0,N) do
  •     for each iteration j in range(0,N) do
  •        for each iteration k in range(0,N) do
  •            e i = a ( q , a i j )
  •            a i j = e x p ( e i ) k e x p ( e k )
  •            A C O V i = q T a i j
  •        end for
  •    end for
  • end for
  • Process features
  • foreach iteration i in range(0, N)do
  •     F i = i A C O V i )
  • end for

3.6.2. Hierarchical Attention Layer

This important layer provide a specified and attention aggregation real variable by using each target convolution as input. Algorithm 3 describes the hierarchical attention routing scheme of the ADCLR model. For each target convolution c o v i R d v in C O V , we evaluate attention a i j , which produced and will be utilized in the classification layer. Figure 6b presents the attention mechanism. The attention task is computed as given in Equations (20) and (21)
e i = a ( q , a i j )
a i j = e x p ( e i ) k e x p ( e k )
Wherein q is a training programme pattern vector, as well as k is the likelihood of convolution pool C O V inside the entire pool, the probability of convolution pool C O V in the entire pool is presented. After obtaining image features, the weighted total is subsequently applied to the overall target dilated convolution layers in the downstream pattern, resulting in a stationary attention aggregation variable. Figure 6b depicts process of a attention mechanism. Finally, the extracted features of the attention-based dilated convolutional are computed as the Equation (22)
A C O V i = q T a i j
According to the the Equation (22), extracted features A C O V i , A C O V 2 , , A C O V n of the attention mechanism is transformed to F 1 , F 2 , , F n for the logistic regression classifier to be classified.

3.6.3. Classification Layer

The main classification procedures presented in the Algorithm 4 after the ADCLR method was trained with a train and validation set of data.
At this point, the ADCLR model uses the LR method to classify images based on the extracted features. LR is a multi-label classification system. The tomato leaf disease target class is predicted at the end of the process. The proposed model’s validation is assessed using various performance evaluation metrics and comparisons on disease images.
Algorithm 4 Main Classification Algorithm of ADCLR Model Using Logistic Regression
  • Data: Input Fi features of attentive dialted cnn
  • Result: Prediction Tomato leaf disease
  • Get extracted features F1, F2, ...., Fi of Dilated CNN
  • Assign extracted features F1, F2, ...., Fi to x1, x2, ...., xi
  • Assign target features to y1, y2, ...., yi
  • Set max iteration as Imax
  • Set Augmented Weight Matrix, θ = 1
  • Set f unction f or costcalculation L(θ) = 0.
  • Mapping input f eatures.
  • Update a Augmented Weight Matrixθusing
  • θ l ( n ) = θ l ( n 1 ) = α × v j
  • Calculate the Cost function or average costL(θ)using
  • L ( θ ) = L ( θ ) + ( 1 m ) ( i = 1 m y i l o g ( h θ ( x ( i ) ) ) + ( 1 y ( i ) l o g ( 1 h θ ( x ( i ) ) )
  • if | l ( θ ) | ( o r ) N = N m a x  then
  •     Find optimum weight of theta
  • else
  •     Update the Augmented Weight Matirxθ
  • end if
  • Optimum weights are obtained for θ
  • Find prediction
  • o = s i g m o i d ( θ T x i )
  • o = 1 ( 1 + e θ T x i )
  • Execute prediction function with class probability
  • p ( y | I ) = p ( y | o )
The goal of this layer is to compute the probabilistic model using the formula p ( y | S ) , at which y is the class predicted. The vector o is given to the multi-layer classification via the logistic regression function for the fixed-length and care-oriented aggregates. The classification Algorithm 4 operation on tomato leaf disease image data is described in this section. This algorithm receives raw data as input and predicts tomato leaf disease and target classes. The tomato leaf disease image is initially collected and preprocessed. The fully extracted feature from the attentive dilated CNN is transmitted to the attentive hierarchical layer. In the last layer, the logistic regression is used to predict ten types of tomato leaf disease based on the attentive dilated CNN features. Here, Algorithm 4 is the LR classification algorithm that is used for tomato leaf disease classification with the ADCLR model.
First take F i preprocessed features from dilated the CNN layer. Initialize the parameters of the logistic regression classification algorithm. Calculate the ( θ T x ) value of the LR classifier then execute the sigmoid function. Then, obtain output o with the computional fuction as the Equations (23) and (25):
o = s i g m o i d ( θ T x )
o = 1 ( 1 + e θ T x )
Now, execute the prediction function with class probability as (23) follows:
p ( y | I ) = p ( y | o )
Figure 7 depicts the operating phase of our concept. From the source term through the prediction or classification algorithm, the mechanics of each layer are depicted in this figure. In one direction, the created output travels through the next processing layer input. To evaluate the performance of the model in the real world, we will use data from tomato leaf disease. The tomato leaf disease image is initially preprocessed as a raw image.

3.7. Evaluation Metrics

To evaluate our model performance, we use a performance evaluation matrix named as accuracy, precision, and recall, our used metrics equations and computation are given in Equations (26)–(28).
A.
Accuracy
The average of all true cases is used to determine the Accuracy of the prediction. It is calculated with the specified equation:
Accuracy = ( True Positive + True Negative ) True Positive + True Negative + Fasle Positive + Fasle Negative
B.
Precision
The amount of true positives divided by the total of positive predictions is known as Precision. The following equation shows the calculation of Precision.
Precision = ( True Positive ) True Positive + False Positive
C.
Recall
The Recall is a measurement of how well our model detects True Positives. As a result, Recall informs us how many tomato plants we accurately identified as having leaf disease out of those that have it.
Recall = ( True Positive ) True Positive + False Negative
D.
F1 Score
The F1 score elegantly summarizes a model’s predictive efficiency and measured by two normally competing metrics, precision and recall.
F 1 = 2 ( Precession Recall ) Precession + Recall

Experimental Setup

After the successful preprocessing procedures (BF filtering and Otsu, Segmentation method), we generated the synthetic image using CGAN. Then, we fed this to the proposed methodology developed with attention-based dilated CNN with logistic regression. We fine-tuned the proposed model in our trials to demonstrate the performance of our model. To evaluate the model, we used a binary cross-validation strategy. Indeed, we divided the dataset into 80% training and 20% validation, with 1000 images used for testing. The remaining 20% was utilized to validate the model before it was evaluated. The attention-based dilated CNN feature extraction model was trained with 100 epochs and 32 mini-batches. To reduce the loss (L), the Adam optimizer was used with only a 1 × 10 3 learning rate. We employed an L2 regularization and a dropout technique with a probability of dropping of 0.5 to offset the effect of the overfitting problem during training. The number of layers of the dilated CNN that was gradually configured to extract features from tomato leaf disease images. The model was implemented using the Python programming language as well as the Google colab framework. The categorical cross-validation was performed on a computer with five CPUs (Intel(R) 3.60 GHz), 32 GB of RAM, and Windows 8 to 10. Figure 7 and Table 2 clearly describes the model’s internal structure.

4. Result Analysis

This section explains the experimental analysis in detail, including qualitative and comparative analysis. The training set performs somewhat better than the validation set, as well as the model accumulating to a steady value, showing that the parameters used to train the model really are not excessive. In the validation model, the suggested technique achieves stable classification performance with good accuracy. In Table 3, Table 4 and Table 5 we use alphabet A to J as A: Bacterial Spot, B: Early Blight, C: Target Spot, D: Yellow Leaf Curl Virus, E: Mosaic Virus, F: Late Blight, G: Leaf Mold, H: Septoria Leaf Spot, I: Two spotted Spider Mite and J: Healthy.

4.1. Qualitative Analysis

In this qualitative analysis, a comprehensive experimental analysis is shown. Here, Table 3 shows the training performance of the proposed ADCLR model on ten disease classes of the tomato leaf dataset. The Table 3 demonstrated that our model performs the same on train and validation data.
Table 4 shows the validation performance of the proposed ADCLR model on ten disease classes of tomato leaf.
Table 5 shows the testing performance of the proposed ADCLR model on ten disease classes of tomato leaf. It is clearly shown that the testing accuracy of the proposed model is slightly lower than the train and validation performance, because we use totally different image for testing the model.
The experimental analysis demonstrated that the testing performance is slightly lower than the training and validation performance. Our ADCLR model achieved 100%, 100%, and 96.60% accuracy on train, validation, and test data, respectively. We also implemented some popular conventional methods and used the same parameter tuning in each model. Based on the results shown in the Table 6, it is clearly shown that our attention-based dilated CNN with logistic regression (ADCLR) model outperforms the other method we implemented in this study for tomato leaf disease detection. The comparative result of LR, CNN-LR and Attention based Dilated CNN-LR method is showed in the bar chart in Figure 8. This figure clearly present that our ADCLR method got higher accuracy that compared method.

4.2. Confusion Matrix

The Area Under the Curve (AUC)—receiver operating characteristic curve (ROC) curve is a performance efficiency measurement technique for multiclass classification. AUC indicates the degree or measurement of separability, whereas ROC is a probability curve. It indicates how well the model can distinguish among categories based on the training, validation, and testing data performance. The ADCLR model generates the following graphical ROC result on tomato leaf disease data. The ROC curve performances are visualized in Figure 9, Figure 10 and Figure 11 during training, validation, and testing. In these figures, the x-axis indicates a false-positive rate and the y-axis indicates a false-negative rate. Our ADCLR model shows better ROC performance over other models. The ADCLR model obtained 0.999 area value of ROC for both of training and validation, and 0.9869 on testing data.
In some contexts, the classifier can become confused when dealing with many classes with comparable features. A confusion matrix could be used to visually measure a model’s classification performance. This experiment was conducted using ten different types of tomato leaf disease. Since the leaf disease image samples are created from the leaf area and are very unstable, the low resolution and poor area selection of the images can result in noisy leaf images, causing the classifier to become confused in some circumstances. We implemented our model (Attention-Dilated CNN LR) and related models (LR and CNN-LR) on train, validation, and test data to generate a confusion matrix. Here, Figure 12, Figure 13 and Figure 14 are generated based on LR, CNN-LR, and Our (Attention dilated CNN-LR) model. All right predictions seem to be on the diagonal, while all wrong predictions are off the diagonal.
An overall description of the data analysis during training is shown in Figure 12. The LR algorithm correctly predicts 952 observations out of 1000 observations, the CNN-LR algorithm correctly predicts 960 observations out of 1000 observations. The proposed attention-based dilated CNN-LR algorithm correctly predicts 966 observations out of the same number of observations. Our proposed ADCLR algorithm performs better in prediction on validation train data as shown in Figure 12, Figure 13 and Figure 14. So, the confusion matrix also indicates that the proposed Attention-Dilated CNN-LR architecture is more accurate than the LR and CNN-LR model. It also helps to overcome the limitations of LR algorithms and works better than CNN-based feature extraction for tomato leaf disease detection.

4.3. Comparisons with State-of-the-Art Methods

This section will discuss all current methodologies, as well as the performance of our proposed method, in classifying the tomato leaf disease Plant Village database. Table 7 categorizes all of the compared methods into three groups: traditional machine learning (ML), deep learning (DL), as well as Deep learning + Machine learning (DLML). Table 7 presents different existing methods performance with their features, data, model, and evaluation metric result. From the comparative table, it is shown that SVM with SIFT features had 85% accuracy [29], RF [29] with Hue and histogram color features having 94% accuracy, the ResNet model obtained 97% accuracy [39]. On the other hand, Machine learning with a deep learning classifier achieved sightly higher accuracy than conventional the ML or DL method. MobileNet and NasNet feature extractor with Logistic Regression got 97% accuracy [28]. Fine-tuned MobileNetv2 obtained 95.6% accuracy. The SVM algorithm got very low accuracy of 85.02% but this method is fast [29]. A maximum of 94% classification accuracy is obtained with the random forest method [30]. Another attention-based method proposed by Devi et al. [48] that used the Salp Swarm Algorithm had 97.56% accuracy to predict five types of tomato leaf disease. The Lightweight Attention-Based CNN mechanism [49] to classify ten types of tomato leaf disease. This method obtained 99.34% accuracy but it has slightly higher time complexity than conventional methods such as CNN [37] and SVM [29]. In 2022, Zhao et al. [50] developed a method utilizing Spatial attention with CNN that had 95.20% accuracy but this has a weakness in generability. We have also shown the comparative performance of the state-of-the-art method in Figure 15.

Comparison of Pre-Network Recognition Accuracy

We also implemented some conventional methods with the same parameter tuning and input size to check the validation of our model. Based on this table, it is clearly shown that our ADCLR method performs better than other implemented methods. We also calculated the executing time of the widely used popular deep learning model for feature extraction. We run and check the time of execution on train, validation, and test sample data. Figure 16 shows the accuracy of different deep learning models during training. In the figure, the x-axis indicates the number of epochs, and the y-axis indicates the accuracy. The graphical line shows the performance of compared models and the proposed model. Figure 16 clearly indicates that our model performs better than the conventional method.
An overall performance (accuracy) comparison of the ADCLR model with most common and related models is clearly shown in Figure 16. Our method got higher accuracy than other implemented methods with the same parameter tuning. Additionally, Table 8 is generated based on our manual implementation with the same parameter tuning and shows that our ADCLR model performance is better compared to the conventional approach.
Figure 17 shows the loss of different deep learning models during training. In the figure, the x-axis indicates the number of epochs, and the y-axis indicates the loss. The graphical line shows the performance of compared models and the proposed model. Figure 17 clearly indicates that our model’s loss is less than the conventional method. The categorical cross entropy loss of our ADCLR method is 0.07.

4.4. Discussion

The proposed method performed image preprocessing using bilateral filtering (BF), segmentation using Otsu’s thresholding, synthetic image generation, feature extraction using attention-based dilated CNN, and classification using logistic regression. The hyper-parameter tuning on logistic regression (LR) seeks the fine change of the hyper-parameters of the attention-based dilated CNN model of feature extraction in such a way that the classification performance is improved to the highest extent possible. To ensure that the attention-based dilated CNN-LR model performs effectively, a complete simulation analysis is performed. The experimental results suggest that the ADCLR model outperforms contemporary state-of-the-art methods on a variety of measures as shown in Table 8. In the future, advanced DL-based image segmentation techniques will be used to improve the detection efficiency of the ADCLR method.
The techniques of our successive preprocessing are computed from the original normal Tomato leaf disease images. The preprocessing helps to extract more precise features from the images. Then, we use the CGAN model to generate the synthetic image to handle imbalance and noisy or wrongly labeled data to obtain good prediction results. The synthetic image is used in the attention-based dilated CNN layer for feature extraction. This aid of this technique is to reduce the misclassification issues and improve performance. In our method, the Bilateral filtering technique helps to remove the noise of the tomato leaf disease image. As a result, Otsu’s method of image segmentation is useful for handling the noise of tomato leaf disease images. Otsu’s image segmentation technique is faster and simpler than other methods [56].
In this proposed ADCLR method for feature extraction, which leverages the attention-based hybrid dilated CNN approach. By dynamically converting its hierarchical system into a deep convolution, we present a new hybrid model for optimizing learning structure, extracting features, classification and analyzing tomato leaf disease. It can automatically extract the hierarchical representations of tomato leaf disease features in order to fully leverage the features. Our hybrid neural network convolution model successfully obtains implicit and relevant feature information. The dilated convolution network can extract informative information about the features. Our hybrid approach, which uses dilated CNN and is based on a paradigm that includes hierarchical self-dilation approaches, provides a reduction in training time and a clear network structure to boost performance. The efficacy of the convolution network dynamic routing algorithm has been increased and an improved convolutional network dynamic convolution method improves the efficacy of efficient routing tuning convolution.
In our ADCLR model, the attention layer simply concentrates on the memory block instead of focusing on the entire feature space. The attention mechanism has the advantage of dramatically reducing the number of parameters and sharing the weights among diverse regional places.
The limitations of previous classification systems are well outlined in the literature section, and our novel ADCLR model is aimed to overcome some of the weaknesses of the compared method. The proposed ADCLR model is evaluated and compared to a number of existing model benchmarks. The proposed ADCLR model achieved an accuracy of 100% in training, 100% in validation, and 96.6% in testing on the PlantVillage tomato leaf disease dataset. This method predicts ten categorizations of the tomato leaf dataset. The experimental analysis of this study showed that our method outperforms over a number of competing baselines and produces a number of cutting-edge outcomes.

4.5. Real Time Test Result on New Image

To study the robustness of the proposed model in a real-time application, we used 1000 non-trained images. In Figure 18, we show the predicted class confidence of the new image of tomato leaf disease. In this figure, we show the result of our model on six images. In this figure, it is clearly shown that the prediction confidence of our model is almost 0.99 for all the new and non-trained tomato disease image samples.

4.6. Complexity Analysis

To ensure the superiority of the proposed strategy in terms of execution time, we constructed a hybrid deep learning ADCLR model that utilized attention-based dilated CNN to extract informative features. During the testing set, the model’s recorded running speed is higher than any other implemented deep learning approach. This method does not use transfer learning or a convolutional neural network (CNN) in feature extraction because its training time is long (see Figure 19). However, we apply a dilated convolution operation with a multi dilation mechanism with attention that uses only the convolutions required for deep feature extraction that helps to reduce the overall computational complexity. The training, validation set has a 122.8 and 5 s run time, respectively (see Figure 19). Figure 19 shows that our attention-based dilated CNN takes less time (122 s) in training whereas CNN takes 210 s. Similarly, our model’s attention-based dilated CNN takes less time of 8 and 5 s for feature extraction of validation and testing data, respectively. In normal convolution layers, convolution kernels are also interconnected and all are convoluted. Furthermore, the addition of attention-based dilated portions, minimizes the processing complexity of the convolutions. The added attention mechanism uses less parameters to select most important and relevant features from the selected region. The time complexity of our model is lower than compared methods.

4.7. Limitation and Future Work

Despite the outstanding performance of the proposed ADCLR model for tomato leaf disease detection, there are a few flaws in the research as well. Firstly, this study only focuses on ten types of tomato leaf disease images for classification, other category of tomato plant leaf disease images did not analyzed. Secondly, Our proposed model validity has been investigated on only the PlantVillage tomato leaf dataset. In a further study, we intend to provide larger and more diverse datasets to test the proposed model and continuously improve the network system’s generalization capacity. We will also work to refine our model so that it may be applied to other datasets including tomato or other leaf disease. Additionally, one of the primary drawbacks of the suggested detection method is that the image used in our experiment was taken in a lab setting. However, our method might be improved to support an integrative plant disease detection system that works in real-world scenarios. However, further work is needed to make this model more advanced to classify broadening categories of plant diseases and automatically recognizing the many stages of the disease, as well as complementing images of leaf diseases in real surroundings.
Figure 20 shows a possible cloud-based tomato leaf disease detection system that can be utilized on mobile phones. The cloud-based system could be able to collect and process the image of tomato leaf disease from the real-time field. The processed data will be interpreted by the cloud-based DL system, and the results are delivered to the agriculture scientist with minimal human effort. Finally, the obtained results will be sent to the farmer’s mobile phone after being verified by the practitioner. The processing costs of the DL model, as well as the data dimension, impact the system’s feasibility. We have a plan to reduce the time complexity and space complexity in future development with more adaptability and generability.

5. Conclusions

The early tomato leaf disease diagnosis method has a great effect on the quality and quantity of tomato production. Traditional methods for detecting tomato disease are time-consuming, labor-intensive, and subjective cost. This study has designed a hybrid architecture (ADCLR) based on attention and dilated convolutional layers with an LR classifier.At first tomato leaf disease images are preprocessed (using nilateral fileterng and otsu segmentation) properly and then we used our Conditional Generative Adversarial Network(CGAN) to generete synthetics tomato leaf disease image. The informative and relevant features from the images were extracted quickly using the attention-based dilated convolutional layers. Then LR has been used to classify the extracted feature. Ten types of tomato leaf disease have been analyzed in this study. Three related classifiers have also been implemented (LR, CNN-LR, and Attention-Dilated CNN-LR) in this study. The validation of the method has been tested using 1000 non-trained images sample. In comparison with other state-of-art methods (CNN, AlexNet, Efficient Net, Xception, MLP, LSTM, GRU, DenseNet, and VGG), our proposed method has achieved higher performance for the tomato leaf disease detection. Our method will work for more types of diseases of plant leaves in the future. We have a plan to reduce the time complexity and space complexity in future development. In a further study, we intend to design a cloud-based artificial intelligence (AI) system using deep learning techniques with more data variants.

Author Contributions

Conceptualization, S.S., F.A.F. and M.N.I.; Data curation, M.S.I.; Formal analysis, F.A.F., M.R., B.S.B., N.H. and M.N.H.; Funding acquisition, M.N.H.; Methodology, M.S.I., S.S., M.N.I. and M.R.; Supervision, M.R. and B.S.B.; Validation, M.S.I., M.N.I., M.R. and N.H.; Visualization, M.S.I., B.S.B., N.H. and M.N.H.; Writing—original draft, M.S.I.; Writing—review & editing, M.S.I., S.S., F.A.F., M.N.I. and M.N.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by Universiti Kuala Lumpur through the Short Term Research Grant under Grant str20017.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

We have no availability of data at this time, in future we will provide code and data publicly.

Conflicts of Interest

The authors declare that there is no conflict of interest in this paper.

References

  1. Gangadevi, G.; Jayakumar, C. Review of Deep Learning Architectures Used for Identification and Classification of Plant Leaf Diseases. In Artificial Intelligent Techniques for Wireless Communication and Networking; Wiley: Hoboken, NJ, USA, 2022; pp. 75–90. [Google Scholar]
  2. Khandokar, I.; Hasan, M.; Ernawan, F.; Islam, S.; Kabir, M. Handwritten character recognition using convolutional neural network. J. Phys. Conf. Ser. 2021, 1918, 042152. [Google Scholar] [CrossRef]
  3. Dora, S.; Terrett, O.M.; Sánchez-Rodríguez, C. Plant–microbe interactions in the apoplast: Communication at the plant cell wall. Plant Cell 2022, 34, 1532–1550. [Google Scholar] [CrossRef] [PubMed]
  4. Scharr, H.; Pridmore, T.P.; Tsaftaris, S.A. Computer Vision Problems in Plant Phenotyping, CVPPP 2017–Introduction to the CVPPP 2017 Workshop Papers. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 2020–2021. [Google Scholar]
  5. Al Farid, F.; Hashim, N.; Abdullah, J.; Bhuiyan, M.R.; Shahida Mohd Isa, W.N.; Uddin, J.; Haque, M.A.; Husen, M.N. A Structured and Methodological Review on Vision-Based Hand Gesture Recognition System. J. Imaging 2022, 8, 153. [Google Scholar] [CrossRef] [PubMed]
  6. Islam, M.S.; Sultana, S.; Kumar Roy, U.; Al Mahmud, J. A review on video classification with methods, findings, performance, challenges, limitations and future work. J. Ilm. Tek. Elektro Komput. Dan Inform. (JITEKI) 2020, 6, 47–57. [Google Scholar] [CrossRef]
  7. Islam, M.S.; Sultana, S.; Roy, U.K.; Al Mahmud, J.; Jahidul, S. HARC-New Hybrid Method with Hierarchical Attention Based Bidirectional Recurrent Neural Network with Dilated Convolutional Neural Network to Recognize Multilabel Emotions from Text. J. Ilm. Tek. Elektro Komput. Dan Inform. (JITEKI) 2021, 7, 142–153. [Google Scholar] [CrossRef]
  8. Shofiqul, M.S.I.; Ab Ghani, N.; Ahmed, M.M. A review on recent advances in Deep learning for Sentiment Analysis: Performances, Challenges and Limitations. J. Ilm. Tek. Elektro Komput. Dan Inform. (JITEKI) 2020, 9, 3775–3783. [Google Scholar]
  9. Islam, M.; Ghani, N.A. A Novel BiGRUBiLSTM Model for Multilevel Sentiment Analysis Using Deep Neural Network with BiGRU-BiLSTM. In Recent Trends in Mechatronics Towards Industry 4.0; Springer: Berlin/Heidelberg, Germany, 2022; pp. 403–414. [Google Scholar]
  10. Islam, M.S.; Sultana, S.; Islam, M.J. New Hybrid Deep Learning Method to Recognize Human Action from Video. J. Ilm. Tek. Elektro Komput. Dan Inform. 2021, 7, 306–313. [Google Scholar] [CrossRef]
  11. Islam, M.N.; Ahmed, F.; Ahammed, M.T.; Rashid, M.; Bari, B.S. Rice Disease Identification Through Leaf Image and IoT Based Smart Rice Field Monitoring System. In Enabling Industry 4.0 through Advances in Mechatronics; Springer: Berlin/Heidelberg, Germany, 2022; pp. 529–539. [Google Scholar]
  12. Bari, B.S.; Islam, M.N.; Rashid, M.; Hasan, M.J.; Razman, M.A.M.; Musa, R.M.; Ab Nasir, A.F.; Majeed, A.P.A. A real-time approach of diagnosing rice leaf disease using deep learning-based faster R-CNN framework. PeerJ Comput. Sci. 2021, 7, e432. [Google Scholar] [CrossRef]
  13. Wijesinha-Bettoni, R.; Mouillé, B. The contribution of potatoes to global food security, nutrition and healthy diets. Am. J. Potato Res. 2019, 96, 139–149. [Google Scholar] [CrossRef]
  14. Grieneisen, M.L.; Aegerter, B.J.; Scott Stoddard, C.; Zhang, M. Yield and fruit quality of grafted tomatoes, and their potential for soil fumigant use reduction. A meta-analysis. Agron. Sustain. Dev. 2018, 38, 29. [Google Scholar] [CrossRef]
  15. Meng, F.; Li, Y.; Li, S.; Chen, H.; Shao, Z.; Jian, Y.; Mao, Y.; Liu, L.; Wang, Q. Carotenoid biofortification in tomato products along whole agro-food chain from field to fork. Trends Food Sci. Technol. 2022, 124, 296–308. [Google Scholar] [CrossRef]
  16. Savary, S.; Ficke, A.; Aubertot, J.N.; Hollier, C. Crop losses due to diseases and their implications for global food production losses and food security. Food Secur. 2012, 4, 519–537. [Google Scholar] [CrossRef]
  17. Islam, M.; Dinh, A.; Wahid, K.; Bhowmik, P. Detection of potato diseases using image segmentation and multiclass support vector machine. In Proceedings of the 2017 IEEE 30th Canadian Conference on Electrical and Computer Engineering (CCECE), Windsor, ON, Canada, 30 April–3 May 2017; pp. 1–4. [Google Scholar]
  18. Azlah, M.A.F.; Chua, L.S.; Rahmad, F.R.; Abdullah, F.I.; Wan Alwi, S.R. Review on techniques for plant leaf classification and recognition. Computers 2019, 8, 77. [Google Scholar] [CrossRef]
  19. Ramya, V.; Lydia, M.A. Leaf disease detection and classification using neural networks. Int. J. Adv. Res. Comput. Commun. Eng. 2016, 5, 207–210. [Google Scholar]
  20. Kantale, P.; Thakare, S. Pomegranate disease classification using Ada-Boost ensemble algorithm. Int. J. Eng. Res. Technol. 2020, 9, 612–620. [Google Scholar]
  21. Johannes, A.; Picon, A.; Alvarez-Gila, A.; Echazarra, J.; Rodriguez-Vaamonde, S.; Navajas, A.D.; Ortiz-Barredo, A. Automatic plant disease diagnosis using mobile capture devices, applied on a wheat use case. Comput. Electron. Agric. 2017, 138, 200–209. [Google Scholar] [CrossRef]
  22. Mohanty, R.; Wankhede, P.; Singh, D.; Vakhare, P. Tomato Plant Leaves Disease Detection using Machine Learning. In Proceedings of the 2022 IEEE International Conference on Applied Artificial Intelligence and Computing (ICAAIC), Salem, India, 9–11 May 2022; pp. 544–549. [Google Scholar]
  23. Sujatha, R.; Chatterjee, J.M.; Jhanjhi, N.; Brohi, S.N. Performance of deep learning vs machine learning in plant leaf disease detection. Microprocess. Microsyst. 2021, 80, 103615. [Google Scholar] [CrossRef]
  24. Rumpf, T.; Mahlein, A.K.; Steiner, U.; Oerke, E.C.; Dehne, H.W.; Plümer, L. Early detection and classification of plant diseases with support vector machines based on hyperspectral reflectance. Comput. Electron. Agric. 2010, 74, 91–99. [Google Scholar] [CrossRef]
  25. Hasan, R.I.; Yusuf, S.M.; Alzubaidi, L. Review of the state of the art of deep learning for plant diseases: A broad analysis and discussion. Plants 2020, 9, 1302. [Google Scholar] [CrossRef]
  26. Hong, H.; Lin, J.; Huang, F. Tomato disease detection and classification by deep learning. In Proceedings of the 2020 IEEE International Conference on Big Data, Artificial Intelligence and Internet of Things Engineering (ICBAIE), Fuzhou, China, 12–14 June 2020; pp. 25–29. [Google Scholar]
  27. Chen, Z.; Wu, R.; Lin, Y.; Li, C.; Chen, S.; Yuan, Z.; Chen, S.; Zou, X. Plant disease recognition model based on improved YOLOv5. Agronomy 2022, 12, 365. [Google Scholar] [CrossRef]
  28. Al-gaashani, M.S.; Shang, F.; Muthanna, M.S.; Khayyat, M.; Abd El-Latif, A.A. Tomato leaf disease classification by exploiting transfer learning and feature concatenation. IET Image Process. 2022, 16, 913–925. [Google Scholar] [CrossRef]
  29. Hlaing, C.S.; Zaw, S.M.M. Tomato plant diseases classification using statistical texture feature and color feature. In Proceedings of the 2018 IEEE/ACIS 17th International Conference on Computer and Information Science (ICIS), Singapore, 6–8 June 2018; pp. 439–444. [Google Scholar]
  30. Basavaiah, J.; Arlene Anthony, A. Tomato leaf disease classification using multiple feature extraction techniques. Wirel. Pers. Commun. 2020, 115, 633–651. [Google Scholar] [CrossRef]
  31. Kalyoncu, C.; Toygar, Ö. GTCLC: Leaf classification method using multiple descriptors. IET Comput. Vis. 2016, 10, 700–708. [Google Scholar] [CrossRef]
  32. Kaur, S.; Pandey, S.; Goel, S. Semi-automatic leaf disease detection and classification system for soybean culture. IET Image Process. 2018, 12, 1038–1048. [Google Scholar] [CrossRef]
  33. Batool, A.; Hyder, S.B.; Rahim, A.; Waheed, N.; Asghar, M.A.; Fawad. Classification and identification of tomato leaf disease using deep neural network. In Proceedings of the 2020 IEEE International Conference on Engineering and Emerging Technologies (ICEET), Lahore, Pakistan, 22–23 February 2020; pp. 1–6. [Google Scholar]
  34. Zaki, S.Z.M.; Zulkifley, M.A.; Stofa, M.M.; Kamari, N.A.M.; Mohamed, N.A. Classification of tomato leaf diseases using MobileNet v2. IAES Int. J. Artif. Intell. 2020, 9, 290. [Google Scholar] [CrossRef]
  35. Agarwal, M.; Singh, A.; Arjaria, S.; Sinha, A.; Gupta, S. ToLeD: Tomato leaf disease detection using convolution neural network. Procedia Comput. Sci. 2020, 167, 293–301. [Google Scholar] [CrossRef]
  36. Wang, L.; Sun, J.; Wu, X.; Shen, J.; Lu, B.; Tan, W. Identification of crop diseases using improved convolutional neural networks. IET Comput. Vis. 2020, 14, 538–545. [Google Scholar] [CrossRef]
  37. Agarwal, M.; Gupta, S.K.; Biswas, K. Development of Efficient CNN model for Tomato crop disease identification. Sustain. Comput. Inform. Syst. 2020, 28, 100407. [Google Scholar] [CrossRef]
  38. Geetharamani, G.; Pandian, A. Identification of plant leaf diseases using a nine-layer deep convolutional neural network. Comput. Electr. Eng. 2019, 76, 323–338. [Google Scholar]
  39. Nithish Kannan, E.; Kaushik, M.; Prakash, P.; Ajay, R.; Veni, S. Tomato leaf disease detection using convolutional neural network with data augmentation. In Proceedings of the 2020 5th IEEE International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 10–12 June 2020; pp. 1125–1132. [Google Scholar]
  40. Hussain, N.; Khan, M.A.; Tariq, U.; Kadry, S.; Yar, M.A.E.; Mostafa, A.M.; Alnuaim, A.A.; Ahmad, S. Multiclass Cucumber Leaf Diseases Recognition Using Best Feature Selection. Comput. Mater. Contin. 2022, 70, 3281–3294. [Google Scholar] [CrossRef]
  41. Fisher, M.S.; Arab, N. Apparatus, System, and Method for Image Normalization Using a Gaussian Residual of Fit Selection Criteria. U.S. Patent 9,245,169, 22 January 2016. [Google Scholar]
  42. Ahmad, I.; Hamid, M.; Yousaf, S.; Shah, S.T.; Ahmad, M.O. Optimizing pretrained convolutional neural networks for tomato leaf disease detection. Complexity 2020, 2020, 8812019. [Google Scholar] [CrossRef]
  43. Aversano, L.; Bernardi, M.L.; Cimitile, M.; Iammarino, M.; Rondinella, S. Tomato diseases Classification Based on VGG and Transfer Learning. In Proceedings of the 2020 IEEE International Workshop on Metrology for Agriculture and Forestry (MetroAgriFor), Virtual, 4–6 November 2020; pp. 129–133. [Google Scholar]
  44. Andrychowicz, M.; Denil, M.; Gomez, S.; Hoffman, M.W.; Pfau, D.; Schaul, T.; Shillingford, B.; De Freitas, N. Learning to learn by gradient descent by gradient descent. In Advances in Neural Information Processing Systems 29 (NIPS 2016); Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R., Eds.; Neural Information Processing Systems (NeurIPS): San Diego, CA, USA, 2016; pp. 3981–3989. [Google Scholar]
  45. Wu, Y.; Jiang, B.; Lu, N. A descriptor system approach for estimation of incipient faults with application to high-speed railway traction devices. IEEE Trans. Syst. Man Cybern. Syst. 2017, 49, 2108–2118. [Google Scholar] [CrossRef]
  46. Wu, Y.; Jiang, B.; Wang, Y. Incipient winding fault detection and diagnosis for squirrel-cage induction motors equipped on CRH trains. ISA Trans. 2020, 99, 488–495. [Google Scholar] [CrossRef] [PubMed]
  47. Lei, X.; Pan, H.; Huang, X. A dilated CNN model for image classification. IEEE Access 2019, 7, 124087–124095. [Google Scholar] [CrossRef]
  48. Devi, S.N.; Muthukumaravel, A. A Novel Salp Swarm Algorithm With Attention-Densenet Enabled Plant Leaf Disease Detection And Classification In Precision Agriculture. In Proceedings of the 2022 IEEE International Conference on Advanced Computing Technologies and Applications (ICACTA), Coimbatore, India, 4–5 March 2022; pp. 1–7. [Google Scholar]
  49. Bhujel, A.; Kim, N.E.; Arulmozhi, E.; Basak, J.K.; Kim, H.T. A lightweight Attention-based convolutional neural networks for tomato leaf disease classification. Agriculture 2022, 12, 228. [Google Scholar] [CrossRef]
  50. Zhao, Y.; Sun, C.; Xu, X.; Chen, J. RIC-Net: A plant disease classification model based on the fusion of Inception and residual structure and embedded attention mechanism. Comput. Electron. Agric. 2022, 193, 106644. [Google Scholar] [CrossRef]
  51. Mohanty, S.P.; Hughes, D.P.; Salathé, M. Using deep learning for image-based plant disease detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef]
  52. Hughes, D.; Salathé, M. An open access repository of images on plant health to enable the development of mobile disease diagnostics. arXiv 2015, arXiv:1511.08060. [Google Scholar]
  53. Chen, B.H.; Cheng, H.Y.; Tseng, Y.S.; Yin, J.L. Two-pass bilateral smooth filtering for remote sensing imagery. IEEE Geosci. Remote. Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  54. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  55. Bird, J.J.; Barnes, C.M.; Manso, L.J.; Ekárt, A.; Faria, D.R. Fruit quality and defect image classification with conditional GAN data augmentation. Sci. Hortic. 2022, 293, 110684. [Google Scholar] [CrossRef]
  56. Trivedi, V.K.; Shukla, P.K.; Pandey, A. Automatic segmentation of plant leaves disease using min-max hue histogram and k-mean clustering. Multimed. Tools Appl. 2022, 81, 20201–20228. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed approach.
Figure 1. The overall framework of the proposed approach.
Sensors 22 06079 g001
Figure 2. The sample images of the tomato leaf disease.
Figure 2. The sample images of the tomato leaf disease.
Sensors 22 06079 g002
Figure 3. (a) Original image; (b) Gray scale image; (c) BF filtered image; (d) Otsu’s image segmented image; (e) extracted features at first dilated convolution layer; (f) extracted features at the second hidden layer of dilated convolution; (g) extracted feature at the third hidden layer of dilated convolution; and (h) extracted features at attention layer.
Figure 3. (a) Original image; (b) Gray scale image; (c) BF filtered image; (d) Otsu’s image segmented image; (e) extracted features at first dilated convolution layer; (f) extracted features at the second hidden layer of dilated convolution; (g) extracted feature at the third hidden layer of dilated convolution; and (h) extracted features at attention layer.
Sensors 22 06079 g003
Figure 4. Synthetic Image Generation using GAN.
Figure 4. Synthetic Image Generation using GAN.
Sensors 22 06079 g004
Figure 5. The overall procedure of the DCLR model.
Figure 5. The overall procedure of the DCLR model.
Sensors 22 06079 g005
Figure 6. (a) Dilated CNN and (b) Attention Mechanism.
Figure 6. (a) Dilated CNN and (b) Attention Mechanism.
Sensors 22 06079 g006
Figure 7. Overall operational process diagram of our ADCLR model.
Figure 7. Overall operational process diagram of our ADCLR model.
Sensors 22 06079 g007
Figure 8. The training, validation, and testing performance of ADCLR model.
Figure 8. The training, validation, and testing performance of ADCLR model.
Sensors 22 06079 g008
Figure 9. ROC curve during training.
Figure 9. ROC curve during training.
Sensors 22 06079 g009
Figure 10. ROC curve during validation.
Figure 10. ROC curve during validation.
Sensors 22 06079 g010
Figure 11. ROC curve during testing.
Figure 11. ROC curve during testing.
Sensors 22 06079 g011
Figure 12. Confusion matrix of ADCLR method during testing.
Figure 12. Confusion matrix of ADCLR method during testing.
Sensors 22 06079 g012
Figure 13. Confusion matrix of CNN-LR method during testing.
Figure 13. Confusion matrix of CNN-LR method during testing.
Sensors 22 06079 g013
Figure 14. Confusion matrix of LR method during testing.
Figure 14. Confusion matrix of LR method during testing.
Sensors 22 06079 g014
Figure 15. Comparison of state of the art method [28,29,34,37,39,48,50].
Figure 15. Comparison of state of the art method [28,29,34,37,39,48,50].
Sensors 22 06079 g015
Figure 16. Accuracy comparison of the implemented deep learning models during training.
Figure 16. Accuracy comparison of the implemented deep learning models during training.
Sensors 22 06079 g016
Figure 17. Loss comparison of the implemented deep learning models during training.
Figure 17. Loss comparison of the implemented deep learning models during training.
Sensors 22 06079 g017
Figure 18. Test Result in Leaf disease detection by our DCLR method.
Figure 18. Test Result in Leaf disease detection by our DCLR method.
Sensors 22 06079 g018
Figure 19. Comparative feature extraction time during (a) Training, (b) Validation, (c) Testing.
Figure 19. Comparative feature extraction time during (a) Training, (b) Validation, (c) Testing.
Sensors 22 06079 g019
Figure 20. Cloud based Leaf disease detection.
Figure 20. Cloud based Leaf disease detection.
Sensors 22 06079 g020
Table 1. The types of Tomato leaf disease.
Table 1. The types of Tomato leaf disease.
Leaf Disease ClassAmountPercentage %
Bacterial spot212613.29
Early blight10006.25
Target Spot14038.77
Yellow Leaf Curl Virus310719.43
Mosaic virus3722.32
Late blight200512.54
Leaf Mold9515.94
Septoria leaf spot176011
Two spotted spider mite167510.48
Healthy15909.94
Total15,989100
Table 2. Network parameter of the proposed model.
Table 2. Network parameter of the proposed model.
Layer NoLayer NameLayer InformationImage SizeOutput Size
012d ConvolutionDilated Convolution256 × 256 × 3(None, 256, 256, 64)
02ActivationReLU256 × 253 × 3(256, 64)
03NormalizationBatch Normalization (256)256 × 253 × 3 (256, 256, 64)
04PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (256, 256, 64)
052d ConvolutionDilated Convolution256 × 253 × 3(128, 128, 64)
06ActivationReLU256 × 253 × 3(128, 64)
07NormalizationBatch Normalization (128)256 × 253 × 3 (128, 128, 64)
08PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (128, 128, 128)
092d ConvolutionDilated Convolution256 × 253 × 3(64, 64, 64)
10ActivationReLU256 × 253 × 3(64, 64)
11NormalizationBatch Normalization (64)256 × 253 × 3 (64, 64, 64)
12PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (64, 64, 256)
132d ConvolutionDilated Convolution256 × 253 × 3(32, 32, 64)
14ActivationReLU256 × 253 × 3(32, 64)
15NormalizationBatch Normalization (32)256 × 253 × 3(32, 32, 64)
16PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (32, 32, 512)
172d ConvolutionDilated Convolution256 × 253 × 3(16, 16, 64)
18ActivationReLU256 × 253 × 3(16, 64)
19NormalizationBatch Normalization (16)256 × 253 × 3 (16, 16, 64)
20PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (16, 16, 1024)
212d ConvolutionDilated Convolution256 × 253 × 3(8, 8, 64)
22ActivationReLU256 × 253 × 3(8, 64)
23NormalizationBatch Normalization (16)256 × 253 × 3 (8,8, 64)
24PoolingMaxPoool2D (Kernel 2, Stride 2, Dilation 1)256 × 253 × 3 (8, 8, 2048)
25ActivationReLU256 × 253 × 3(256, 64)
26PoolingAdaptiveMaxPool2d256 × 253 × 3(8, 512)
27DropoutDrop out (0.5)256 × 253 × 3
28AttentionAttention256 × 253 × 3(100,10)
29FlatternFlatten Layer256 × 253 × 3(64 × 10)
30Logistic RegressionLogistic Regression (LR) Classifier
(N = 10, Number of features,
Random State = 100)
256 × 253 × 310
Table 3. Training performance of ADCLR model.
Table 3. Training performance of ADCLR model.
Model LRMetricsTraining Performance
ABCDEFGHIJ
LRAccuracy0.960.950.960.970.980.960.970.950.970.96
Precession0.980.950.940.981.000.960.980.930.960.99
Recall0.960.950.960.970.980.960.920.980.980.98
F10.960.980.930.960.990.960.950.960.970.98
CNN-LRAccuracy0.970.970.960.970.970.970.980.980.970.96
Precession0.990.950.940.981.000.960.980.930.960.99
Recall0.990.930.981.000.970.960.940.980.990.98
F10.980.950.960.990.980.960.940.960.970.98
Attention-Dilated CNN- LRAccuracy1.001.001.001.001.001.001.001.001.001.00
Precession1.001.001.001.001.001.001.001.001.001.00
Recall1.001.001.001.001.001.001.001.001.001.00
F11.001.001.001.001.001.001.001.001.001.00
Training supports as 80% data 1701800112324862981604761141613401272
Table 4. Validation performance of ADCLR model.
Table 4. Validation performance of ADCLR model.
ModelMetricsValidation Performance
ABCDEFGHIJ
LRAccuracy0.970.960.970.970.980.960.970.950.970.96
Precession0.990.960.950.991.000.960.980.940.960.99
Recall0.970.980.960.970.990.960.920.980.980.98
F10.960.980.940.980.990.970.980.970.970.98
CNN-LRAccuracy0.980.970.970.980.980.980.980.970.970.96
Precession0.990.960.980.991.000.970.970.970.980.99
Recall0.990.940.991.000.970.990.980.980.990.98
F10.980.950.960.990.990.980.990.990.980.98
Attention-Dilated CNN-LRAccuracy1.001.001.001.001.001.001.001.001.001.00
Precession1.001.001.001.001.001.001.001.001.001.00
Recall1.001.001.001.001.001.001.001.001.001.00
F11.001.001.001.001.001.001.001.001.001.00
Validation supports as 20% data 42520028062174401190354335318
Table 5. Testing performance of ADCLR model.
Table 5. Testing performance of ADCLR model.
ModelMetricsTesting Performance
ABCDEFGHIJ
LRAccuracy0.960.940.940.950.970.960.970.930.950.94
Precession0.980.910.90.960.980.960.980.950.960.97
Recall0.970.890.940.980.980.930.90.920.940.96
F10.950.920.950.980.950.940.950.990.960.94
CNN- LRAccuracy0.980.950.950.960.980.970.960.940.960.95
Precession0.990.940.930.970.990.950.970.940.950.98
Recall0.970.920.950.990.990.940.910.940.960.97
F10.960.930.950.970.970.950.940.980.980.95
Attention-Dilated CNN- LRAccuracy0.990.970.960.970.970.960.970.950.970.96
Precession0.980.950.940.981.000.960.980.930.960.99
Recall0.980.930.971.000.970.960.920.980.980.98
F10.990.950.960.990.980.960.950.960.970.98
Supports 100 new image per class 100100100100100100100100100100
Table 6. The performance comparison of three related models (training, validation and testing).
Table 6. The performance comparison of three related models (training, validation and testing).
Training Performance
ModelInput SizeSampleAccuracy
LR256 × 256 × 312,80196.30%
CNN-LR256 × 256 × 312,80197.00%
Attention-Dilatd CNN LR(ADCLR)256 × 253 × 312,801100.0%
Validation Performance
ModelInput SizeSampleAccuracy
LR256 × 256 × 3319896.6%
CNN-LR256 × 253 × 3319897.4%
Attention-Dilatd CNN LR(ADCLR)256 × 253 × 33198100.00%
Testing Performance
ModelInput SizeSampleAccuracy
LR256 × 256 × 3100095.20%
CNN-LR256 × 253 × 3100096.00%
Attention-Dilatd CNN LR(ADCLR)256 × 256 × 3100096.60%
Table 7. The performance comparison of the related studies with the proposed approach. DL: Deep Learning, ML = Machine Learning, DLML = Deep learning with Machine Learning.
Table 7. The performance comparison of the related studies with the proposed approach. DL: Deep Learning, ML = Machine Learning, DLML = Deep learning with Machine Learning.
TypeAuthorMethodFeaturesClassSamplesDataPerformance
MLHlaing et al. [29]SVMSIFT and color conversion Features73535Plant Village (Tomato)Accuracy 85.02%
Basavaiah et al. [29]Random forest and decision treeHu Moments, pattern and colour histograms.5300Plant Village (Tomato)Accuracy 94%(RF) Accuracy 90%(DT)
DLAgarwal et al. [37]CNNCNN model1010,100Plant Village (Tomato)Accuracy 91.2%
Nitish et al. [39]ResNetResNet-50 model612,206Plant Village (Tomato)Accuracy 97%
MLDLMedhar et al. [28]MobileNetV2 or NASNetMobile and Logistic RegressionMobileNetV2 or NASNetMobile feature extractor61,152Plant Village (Tomato)Accuracy 97% (MobileNetV2) Accuracy 97% (NASNetMobile)
Zaki et al. [34]MobileNetV2Fine-tune MobileNetV243471Plant Village (Tomato)Accuracy 95.6%
Devi et al. [48]Attention mechanismDense net with Attention59281Plant Village (Tomato)Accuracy 97.56%
Bhujel et al. [49]Lightweight Attention-Based CNNAttentive CNN1019,510Plant Village (Tomato)Accuracy 99.34%
Zaho et al. [50]Spatial attention with CNNFully connected layer1018,160Plant Village (Tomato)Accuracy 95.20%
Proposed ADCLR (Our)Attention-Dilated CNN and Logistic Regression with synthetic imageAttention-based Dilated CNN1015,989Plant Village (Tomato)Accuracy 100.00% F1 100.00% Precession 100.00% Recall 100.00%
Table 8. Performance comparison of the implemented conventional models with the proposed ADCLR model.
Table 8. Performance comparison of the implemented conventional models with the proposed ADCLR model.
ModelInput SizeAccuracyPrecessoinRecallF1-Score
CNN256, 256, 388.70%86.76%88.71%87.30%
AlexNet256, 256, 391.87%89.93%91.88%90.47%
EfficientNet256, 256, 392.25%90.31%92.26%90.75%
Xception256, 256, 397.61%95.67%88.70%96.21%
Inception-Resnet-V2256, 256, 397.80%95.86%95.87%96.41%
MLP256, 256, 397.99%96.05%97.99%96.59%
LSTM256, 256, 398.50%96.56%98.50%97.11%
GRU256, 256, 398.74%96.80%98.75%97.34%
DenseNet256, 256, 398.88%96.94%98.89%97.48%
VGG256, 256, 399.00%97.06%99.01%97.61%
Dilated CNN-RNN256, 256, 399.15%97.21%99.15%97.75%
ADCLR (Our)256, 256, 3100.00%100.00%100.00%100.00%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Islam, M.S.; Sultana, S.; Farid, F.A.; Islam, M.N.; Rashid, M.; Bari, B.S.; Hashim, N.; Husen, M.N. Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification. Sensors 2022, 22, 6079. https://doi.org/10.3390/s22166079

AMA Style

Islam MS, Sultana S, Farid FA, Islam MN, Rashid M, Bari BS, Hashim N, Husen MN. Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification. Sensors. 2022; 22(16):6079. https://doi.org/10.3390/s22166079

Chicago/Turabian Style

Islam, Md Shofiqul, Sunjida Sultana, Fahmid Al Farid, Md Nahidul Islam, Mamunur Rashid, Bifta Sama Bari, Noramiza Hashim, and Mohd Nizam Husen. 2022. "Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification" Sensors 22, no. 16: 6079. https://doi.org/10.3390/s22166079

APA Style

Islam, M. S., Sultana, S., Farid, F. A., Islam, M. N., Rashid, M., Bari, B. S., Hashim, N., & Husen, M. N. (2022). Multimodal Hybrid Deep Learning Approach to Detect Tomato Leaf Disease Using Attention Based Dilated Convolution Feature Extractor with Logistic Regression Classification. Sensors, 22(16), 6079. https://doi.org/10.3390/s22166079

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop