1. Introduction
The agricultural sector is crucial in ensuring global food security and economic stability. Among the myriads of crops grown worldwide, tomatoes hold significant importance due to their high nutritional value and widespread use in culinary practices. However, tomato cultivation faces numerous challenges, notably from various diseases that severely impact yield and quality. Common diseases such as Leaf blight, Leaf curl, Septoria leaf spot, and Verticillium wilt can lead to significant crop losses and deteriorate the biochemical quality of tomatoes, further compounding the challenges in meeting consumer demands. With the global population projected to reach 9.7 billion by 2050, the demand for tomatoes is expected to rise significantly, prioritizing effective disease management for sustaining production and meeting food supply needs [
1]. Traditional disease detection methods rely heavily on manual inspection and expert knowledge and have several inherent limitations. These methods are often time-consuming, labor-intensive, and prone to human error. Moreover, they do not scale well, making it difficult to address the rising global demand for efficient and accurate tomato disease detection. Given the potential for severe economic losses in the agricultural sector, a more scalable and correct solution is required. Recent advancements in machine learning (ML) and artificial intelligence (AI) offer promising avenues to address these challenges. In particular, convolutional neural networks (CNNs) have shown exceptional capabilities in image recognition tasks, including plant disease detection [
2]. However, the practical deployment of these models in agricultural settings is often hindered by the limited computational resources of edge devices used in the field. Edge AI emerges as a viable solution to this challenge by processing data locally on devices deployed in the field rather than relying on cloud-based services. This approach reduces latency and bandwidth consumption, enhances data privacy, and enables real-time decision-making. These aspects are crucial for timely disease management, where delays can result in widespread crop damage. Additionally, edge AI presents a cost-effective alternative, particularly for farmers in regions with limited internet connectivity, thus extending the potential reach of precision agriculture technologies.
In the rapidly advancing field of precision agriculture, accurately detecting crop diseases is pivotal for optimizing yield and quality, particularly for high-value crops such as tomatoes. Tomato plants are susceptible to various leaf diseases, including early blight, late blight, and leaf mold, drastically affecting productivity and economic viability [
3]. Traditional disease detection methods, often reliant on manual inspection or laboratory testing, are labor-intensive and time-consuming. This paper addresses these limitations by leveraging Edge Impulse to deploy multiple deep learning models on a single edge device, creating a solution that integrates seamlessly into practical agricultural settings [
4]. Our study begins with assembling an extensive dataset comprising thousands of tomato leaf images, categorized into distinct classes representing healthy leaves and those afflicted by specific diseases. This dataset is meticulously curated and enhanced through preprocessing techniques such as resizing, normalization, and augmentation (including rotation, flipping, and color adjustments) to ensure a comprehensive and robust resource for a model train.
We then focus on designing and training convolutional neural network (CNN) architectures using TensorFlow, aiming for high disease detection accuracy while optimizing the edge deployment models. Various CNN architectures, including MobileNet, Inception, ShuffleNet, Squeezenet, EfficientNet, and a custom Deep Neural Network (DNN), are explored and fine-tuned to balance predictive performance and computational efficiency. To ensure compatibility with edge devices, the trained models are converted to TensorFlow Lite format and subjected to int8 quantization, significantly reducing their size and enhancing inference speed. This step is crucial for deploying models on resource-constrained edge hardware with limited processing power and memory. The deployment phase involves generating the necessary files and configuring the Edge Impulse platform to enable the simultaneous operation of multiple models on a single device. This setup not only facilitates versatile disease detection but also demonstrates the potential of edge AI in practical agricultural applications.
Performance evaluations on edge hardware platforms, such as mobile phones and microcontroller units, provide insights into the models’ operational efficacy in real-world conditions. Key metrics such as inference speed, accuracy, and resource utilization are analyzed to validate the models’ reliability and efficiency [
5].
The contributions of this paper include the following:
Curated and categorized a comprehensive dataset of tomato leaf images, capturing various disease types, to support advanced model training.
Engineered and trained various convolutional neural network (CNN) models using TensorFlow for tomato leaf disease identification.
Optimized CNN models for efficient edge deployment by converting them to TensorFlow Lite and implementing int8 quantization, significantly enhancing their suitability for resource-constrained environments.
Successfully deployed multiple deep learning models on a single edge device utilizing the Edge Impulse platform, demonstrating the feasibility of real-time, on-device disease detection.
Conducted rigorous performance evaluations of the deployed models on edge hardware, analyzing key metrics such as inference speed, accuracy, and resource utilization to validate their operational effectiveness in practical agricultural scenarios.
The rest of this paper is organized as follows: The article presents a comprehensive approach to tackling tomato leaf diseases, which contribute to significant global yield losses in
Section 2.
Section 3 details the curation of an extensive dataset of tomato leaf images, the application of advanced preprocessing and augmentation techniques, and the design of efficient CNNs using TensorFlow and Edge Impulse.
Section 4 describes the deployment process involving generating files via Edge Impulse and configuring the platform for simultaneous model operation on a single device.
Section 5 explores the practical integration of edge AI in agriculture, including a specific case study in Ghana.
Section 6 presents the paper conclusion by summarizing the findings and emphasizing the potential impact of this technology on crop management and sustainable farming practices.
2. Review of Literature
Tomato leaf diseases significantly threaten agricultural productivity, impacting both the yield and quality of tomato crops. Effective disease management hinges on early detection and accurate diagnosis, crucial for minimizing losses and optimizing crop health. Recent advancements in computer vision and deep learning have facilitated the development of automated systems capable of detecting plant diseases from leaf images with high precision. Convolutional neural networks (CNNs), in particular, have emerged as a powerful tool in this domain, enabling the accurate classification of disease types based on visual data. However, the practical deployment of these models in real-world agricultural settings faces computational efficiency and resource constraint challenges, especially for edge devices used in the field. With the growing need for real-time disease detection, edge computing has become an essential component of these systems. Platforms like Edge Impulse allow for deploying multiple deep learning models on a single edge device, significantly enhancing the ability to perform real-time analysis in the field. This review synthesizes current tomato leaf disease detection methodologies and highlights existing approaches’ limitations.
Recent studies have explored various CNN architectures and their deployment on edge devices to improve detection accuracy and operational efficiency. For instance, Jayanthi G. et al. implemented a Raspberry Pi-based system for real-time detection, achieving a training accuracy of 98% with the PlantVillage dataset [
6]. Despite these promising results, their approach was limited by a single-CNN architecture and potential overfitting issues, with a testing accuracy of 88.17%. Our study contrasts this by exploring multiple CNN architectures to enhance model generalization and evaluate performance across edge devices to ensure broader applicability. Additionally, our approach includes rigorous testing and validation processes to mitigate overfitting and ensure that the models perform well on unseen data. This comprehensive evaluation helps select the best-performing and most efficient architectures for practical use.
Brindha R. et al. conducted a comparative analysis of transfer learning models, including VGG16 and MobileNet, for disease detection [
7]. Although transfer learning improved classification accuracy, the computational demands of some models posed challenges for edge deployment. We address this by employing a systematic approach to model selection, focusing on accuracy and computational efficiency. By converting models to TensorFlow Lite and applying quantization, our research makes computationally intensive models feasible for edge deployment in resource-constrained environments. This enhances the model’s applicability and ensures that the system remains cost-effective and accessible to farmers and agricultural workers in developing regions.
Dataset quality and diversity are pivotal for training effective detection models, as highlighted by another study that achieved high accuracy with a limited dataset. Our research expands on this by preparing a comprehensive dataset with diverse disease representations. We emphasize model size optimization and inference speed, facilitating seamless deployment on edge devices while maintaining high accuracy. In their study on model interpretability, Tanjim Mahmud et al. employed CNNs and pre-trained models like EfficientNetB3, Xception, and MobileNetV2 for tomato leaf disease recognition, achieving an accuracy of 0.993 with EfficientNetB3 [
8]—their research combined machine learning techniques with image processing methods to enable prompt disease diagnosis. However, the system was not fully ready for real-world agricultural deployment, as further work was needed to improve its trustworthiness. Our approach integrates comprehensive testing and validation phases to ensure our models are reliable and effective in real-world conditions.
A recent research paper used deep learning to detect diseases on tomato plant leaves and aimed to run the algorithm in real time on a robot in the field [
9]. They tested AlexNet and SqueezeNet architectures on an Nvidia Jetson TX1 and used the PlantVillage dataset. The study was limited by the number of architectures tested and the dataset’s diversity, highlighting the need for broader evaluation. Our research builds on this by testing a more comprehensive range of CNN architectures and utilizing more diverse datasets to ensure the models are robust and generalizable across different scenarios. This broader evaluation is crucial for developing systems that can be effectively used in varied agricultural environments. M. S. Alzahrani et al. evaluated three deep learning models—DenseNet121, ResNet50V2, and ViT—for early detection and recognition of tomato leaf diseases using a dataset from Kaggle [
10]. Their study’s limitations included the dataset size and the need for further testing on diverse datasets to ensure model generalizability. Our research addresses these limitations by expanding the dataset and conducting extensive tests to validate the model’s performance across different data samples. By focusing on early detection and recognition, we aim to develop models to identify diseases at their initial stages, allowing for timely interventions and reducing crop losses.
V. Gonzalez-Huitron et al. trained and evaluated lightweight CNN models for tomato leaf disease detection on a Raspberry Pi 4 [
11]. They used a subset of the PlantVillage dataset and employed depthwise separable convolution architectures suitable for low-power devices. The study’s limitations included focusing on low-cost devices rather than novel techniques and using a subset of an existing dataset. Our research incorporates novel techniques and a more comprehensive dataset to improve detection accuracy and efficiency. By optimizing the models for low-power devices, we ensure that the system remains accessible and practical for widespread use, particularly in resource-limited settings.
P. Anh et al. benchmarked deep learning models for multi-leaf disease detection on edge devices, identifying MobileNet V3 as the most suitable model for deployment on a Raspberry Pi 3 [
12]. Their study highlighted the challenges of deploying models on resource-constrained devices and emphasized the need for balancing accuracy, inference time, and memory usage. Our research takes this further by benchmarking multiple models and optimizing them for various edge devices to ensure they can operate efficiently within the constraints of limited computational resources. This balance between accuracy and resource efficiency is crucial for developing practical solutions for real-time disease detection in the field. Paarth Bir et al. used transfer learning with pre-trained models like EfficientNetB0, MobileNetV2, and VGG19 for tomato leaf disease detection on mobile devices [
13]. Their study noted the computational and memory requirements of these models, which may limit their use on mobile devices, and their relatively small dataset size as limitations. Our research overcomes these limitations by using a larger, more diverse dataset and optimizing the models for mobile deployment. By reducing the computational and memory footprint of the models, we ensure they can be effectively used on mobile devices, making them more accessible to farmers and agricultural workers.
H. N. N. Kumar et al. integrated machine learning and the IoT for real-time tomato leaf disease detection, using a three-stage Stacked Deep Convolutional Autoencoder to optimize CNN performance [
14]. Their study was not an actual real-time system and required further research for real-world deployment. Our research builds on this by developing an accurate real-time system operating efficiently in real-world conditions. By integrating the IoT and advanced machine learning techniques, we create a system that can provide timely and accurate disease detection, enabling prompt interventions and reducing crop losses. Rajeev Karothia et al. proposed CNN models for identifying nine common tomato leaf diseases, with InceptionV3 achieving the highest accuracy of 99.64% [
15]. Their study planned to expand to other crops but was limited by the dataset’s diversity and the performance of different models. Our research addresses these limitations by using a more diverse dataset and exploring multiple CNN architectures to identify the most effective models for disease detection. By expanding the scope to include other crops, we aim to develop a versatile system that can be used for disease detection across different agricultural settings.
M. Afify et al. developed a deep learning system for detecting nine tomato crop diseases, deploying the final model on a smartphone for real-time classification [
16]. Their study faced limitations with the dataset, generalization issues, and a significant drop in performance on new, unseen images. Our research addresses these issues using a larger, more diverse dataset and implementing techniques to improve model generalization. By ensuring that the models perform well on new, unseen data, we create a robust system that can be effectively used for real-time disease detection on smartphones and other mobile devices.
This comprehensive review of the existing literature identifies key advancements and limitations in tomato leaf disease detection. While significant progress has been made, the need for improved CNN architectures, diverse datasets, and optimized deployment strategies on edge devices remains. Our research addresses these challenges by enhancing model performance, ensuring practical applicability in real-world agricultural settings, and pushing the boundaries of current methodologies to deliver more robust, efficient, and accessible solutions for tomato leaf disease detection.
Table 1 provides a summary of the key findings from the literature.
3. Methodology
To address the challenges of real-time tomato leaf disease detection in agricultural settings, we developed a comprehensive and well-structured methodology that integrates deep learning techniques with edge computing capabilities. Our approach ensures high model accuracy and guarantees efficient real-time performance in resource-constrained environments, such as the fields where these systems will be deployed. By leveraging TensorFlow for model training and Edge Impulse for streamlined deployment (
Figure 1), the methodology is designed to enable models to perform real-time inference on edge devices without relying heavily on cloud infrastructure.
Our methodology is divided into two primary segments: TensorFlow and Edge Impulse. The TensorFlow segment begins with the critical step of dataset preparation. A diverse and extensive dataset of tomato leaf images is curated, representing healthy leaves and those affected by various diseases. We further employ data augmentation techniques such as image rotation, flipping, and brightness adjustments to enhance the model’s generalizability to unseen conditions. These techniques ensure that the model can handle various lighting and environmental conditions, improving its robustness in real-world scenarios. Following this, we design and train various convolutional neural network (CNN) architectures tailored to tomato leaf disease detection. During training, hyperparameters such as learning rates, batch sizes, and network depths are meticulously tuned to achieve optimal performance. Once trained, the models are converted to TensorFlow Lite (TFLite) format and undergo 8-bit quantization. This step is crucial for reducing the model size and computational demands, making them suitable for real-time inference on low-powered edge devices.
In the Edge Impulse segment, we focus on deploying the trained models onto edge devices, ensuring they maintain high accuracy and efficiency when used in agricultural fields. The deployment involves generating the necessary deployment files and configuring Edge Impulse to handle multiple models on a single edge device. By integrating the models with edge devices like Raspberry Pi or microcontrollers, we enable real-time, on-device disease detection without cloud processing, reducing latency and bandwidth usage. In this phase, the models undergo rigorous testing in real-world agricultural environments, where we evaluate vital metrics such as accuracy, inference speed, and power consumption to ensure their practicality for field use. The individual steps involved in the methodology are illustrated in
Figure 2.
3.1. TensorFlow
3.1.1. Dataset Acquisition and Augmentation
The dataset for training the deep learning models was a comprehensive collection of tomato leaf images meticulously curated by P. K. Mensah et al. [
17]. Originally collected from local farms in Ghana, this dataset represented various tomato leaf disease states and healthy conditions. The varieties of tomatoes used included Cherry, Techiman, Petomech, Power Rano, and Jaguar F1 tomatoes. We focused exclusively on the tomato portion of the dataset, which comprised a total of 5435 raw images, further augmented to create 27,178 images categorized into multiple disease classes. Sample images of the dataset can be seen in
Figure 3. For this study, we selected a subset of 16,947 images categorized into five specific classes: Healthy (2500 images), Leaf Blight (3645 images), Leaf Curl (2582 images), Septoria Leaf Spot (4356 images), and Verticillium Wilt (3864 images).
Figure 4 shows the distribution of the images in the dataset.
This selection ensured a balanced representation across different disease states, which is essential for training robust deep-learning models. To further enhance the robustness of our models, we employed a series of data augmentation techniques, including rotation, flipping, scaling, and color jittering. These techniques artificially expanded the dataset’s size and variability, simulating the variations likely encountered in real-world agricultural settings. This augmentation process was crucial in improving the models’ generalization capabilities, enabling them to classify new, unseen data accurately during real-time inference. The dataset was divided into training and validation subsets, usually with an 80/20 split ratio. The training subset was utilized to train the model, while the validation subset was leveraged to assess the model’s performance during training and fine-tune hyperparameters.
3.1.2. Models Design and Training
In the model design phase, we explored several state-of-the-art CNN architectures, each selected for its balance of accuracy, computational efficiency, and suitability for deployment on resource-constrained edge devices. The architectures included MobileNet, Inception, ShuffleNet, SqueezeNet, EfficientNet, and a custom DNN explicitly designed for this task.
MobileNet: We utilized MobileNet, a lightweight neural network designed for mobile and embedded devices featuring depthwise separable convolutions to reduce computational complexity. The architecture illustrated in
Figure 5 includes depthwise and pointwise convolution layers, batch normalization, ReLU activations, global average pooling, and dropout regularization. MobileNet’s efficiency makes it ideal for resource-constrained applications like maize crop disease identification [
18].
Inception: The Inception architecture, introduced in the GoogLeNet model, revolutionized deep learning by using Inception modules that apply multiple convolutional filters simultaneously, capturing features at different scales. This multi-path approach allows the network to learn richer, more complex representations while maintaining computational efficiency.
Figure 6 illustrates the architecture of the Inception model. Inception was pivotal in achieving high accuracy in image classification tasks, particularly in the ImageNet challenge [
19].
ShuffleNet: ShuffleNet is a lightweight convolutional neural network designed for mobile devices. It employs pointwise group convolutions and channel shuffling to reduce computational complexity while maintaining accuracy [
20]. The channel shuffle operation ensures efficient feature learning across groups, making ShuffleNet highly suitable for resource-constrained environments without sacrificing performance on tasks like image classification.
Figure 7 shows the architecture of the ShuffleNet model.
SqueezeNet: SqueezeNet is a compact neural network architecture that achieves AlexNet-level accuracy with significantly fewer parameters. It uses “squeeze” and “expand” layers to reduce the number of parameters, optimizing efficiency [
21]. This design allows SqueezeNet to deliver high performance in image classification tasks while being highly suitable for deployment on devices with limited memory and computational power.
Figure 8 shows the architecture of SqueezeNet.
EfficientNet: EfficientNet, a deep-learning architecture, optimizes network depth, width, and resolution using Neural Architecture Search (NAS) and model scaling. Scaling uniformly improves performance across tasks like ImageNet classification, object detection, and semantic segmentation [
22]. A dropout layer was added to check the model’s overfitting.
Figure 9 illustrates the architecture of EfficientNet.
Custom DNN: The model architecture comprises a series of convolutional (Conv2D) and max-pooling (MaxPooling2D) layers, followed by a flattened layer to transition from convolutional to fully connected layers. The convolutional layers employ varying filter sizes and numbers to extract hierarchical features from the input data, and are then down-sampled using max-pooling operations to reduce spatial dimensions while preserving important features. The flattened layer reshapes the output from the convolutional layers into a vector suitable for input into the subsequent fully connected layers [
23,
24]. This architecture is designed for classification tasks, particularly for tomato crop disease identification, where the model learns to classify images into one of five classes based on the features extracted from the input images.
Figure 10 shows the architecture of the model.
3.1.3. Model Training and Performance Evaluation
Each model was trained using the prepared dataset, with hyperparameters meticulously tuned to optimize performance. Training was conducted using techniques, including learning rate annealing, early stopping, and batch normalization, to prevent overfitting and ensure convergence to an optimal solution. The images were resized to 96 × 96 dimensions. The hyperparameters used for training include a learning rate of 0.0005, batch size of 32, and epochs of 20.
Post-training, the performance of each model was evaluated using standard metrics, including accuracy, precision, recall, and F1-score, calculated on a reserved validation set. This evaluation allowed us to identify the most effective models for conversion and deployment. The models were then converted to TensorFlow Lite (TFLite) format, a critical step in ensuring compatibility with edge devices. We also applied 8-bit integer quantization during this conversion process, significantly reducing the model’s size and computational requirements. While reducing model precision, this quantization process is essential for maintaining high inference speeds and low power consumption on edge devices.
3.2. Edge Impulse
Edge Impulse is a powerful platform designed to deploy machine learning models on edge devices. It enables efficient model conversion, optimization, and deployment, making it ideal for real-time applications in resource-constrained environments like agriculture.
Figure 11 represents the flow cycle of training and deploying machine learning models on Edge Impulse.
3.2.1. Deployment File Generation
The next phase of our methodology involved preparing the models for deployment using Edge Impulse, a platform designed to facilitate the deployment of machine learning models on edge devices. We used Edge Impulse to generate deployment files after selecting the best-performing models from the TensorFlow segment. This process included converting the models into formats optimized for the specific hardware of our target edge devices, such as microcontrollers or single-board computers, and configuring the deployment platform involved, setting up the edge devices with the necessary software environments, including the deployment of Edge Impulse’s firmware and runtime environment. This configuration ensured the edge devices could run the TFLite models efficiently, with all dependencies correctly managed.
3.2.2. Model Testing and Performance Evaluation
Once the deployment files were generated and the platform was configured, we rigorously tested the models on the target edge devices. This testing phase was crucial to validate the models’ performance in real-world scenarios, where factors such as varying lighting conditions, device power limitations, and processing delays could impact inference accuracy and speed. Performance metrics, including inference time, power consumption, and on-device accuracy, were collected to assess the models’ viability in practical applications. The entire test dataset was also classified using the deployed models to evaluate their real-time performance under typical agricultural conditions.
3.2.3. On-Device Performance Evaluation
The final step in our methodology involved an in-depth evaluation of the models’ behavior directly on the edge devices. This phase included monitoring the models during continuous operation and assessing their stability, responsiveness, and accuracy over extended periods. The models’ behavior was analyzed under various operational conditions, such as fluctuating power supply and environmental noise, to ensure robust performance in real-world settings.
Any anomalies or performance issues detected during the on-device evaluation were addressed by iterating on the model design or adjusting the deployment configuration. The trained models were then uploaded to the Edge Impulse platform for further refinement, if necessary, ensuring that the final deployment was reliable and efficient.
By adhering to this rigorous methodology, we aim to provide a practical solution for tomato leaf disease detection that is accurate and deployable in resource-constrained environments, empowering farmers with the tools to protect their crops and optimize yields.
5. Discussion
The experimental evaluation of various deep learning models for tomato leaf disease detection, deployed using the Edge Impulse platform, provides valuable insights into their on-device performance and practical applicability.
MobileNet demonstrated robust training accuracy (96.78%) with a low loss (0.0969). However, the substantial drop in validation accuracy to 67.78% suggests overfitting, limiting its generalization ability. Despite a compact model size (3.40 MB) and efficient training time (11,306.58 s), MobileNet’s latency on low-end MCUs (125,160 ms) highlights challenges in real-time deployment. High-end MCUs (2562 ms) and AI accelerators (427 ms) improved performance, indicating their potential for use on more capable devices. Inception achieved a training accuracy of 70.77% and a validation accuracy of 61.25%, reflecting balanced performance. However, its large model size (21.40 MB) and high latency on low-end MCUs (451,659 ms) suggest limited suitability for resource-constrained environments. Its improved performance on high-end MCUs (9244 ms) and AI accelerators (1541 ms) makes Inception more viable where computational resources are less restricted. ShuffleNet displayed lower accuracy (65.80% training, 53.22% validation) but excelled in efficiency with a small model size (52.50 KB) and minimal latency (3811 ms on low-end MCUs, 13 ms with AI accelerators). Its compact design positions ShuffleNet as a suitable candidate for edge deployment in constrained environments. SqueezeNet offered a training accuracy of 73.76% and a validation accuracy of 59.70%. Its small model size (166.40 KB) and moderate latency on high-end MCUs (1458 ms) and AI accelerators (243 ms) make it a viable option for deployment in resource-constrained settings. However, its accuracy is lower compared to other models. EfficientNet exhibited high training accuracy (97.12%) but suffered from significant overfitting, with its validation accuracy dropping to 11.09%. The model’s large size (4.60 MB) and extended training time (53,194.23 s) pose challenges for practical deployment. While latency is reduced on high-end MCUs (1448 ms) and AI accelerators (242 ms), further optimization is needed to address generalization issues. The custom DNN balanced training (87.28%) and validation (61.85%) accuracy with moderate overfitting. Its model size (338.80 KB) and latency on low-end MCUs (43,664 ms) and AI accelerators (149 ms) suggest it is a promising option for edge deployment, combining efficiency with reasonable performance.
Use Case
Ghanaian farmers, particularly rural farmers, face significant challenges in accurately identifying pests and diseases affecting their crops. This often leads to delayed and ineffective responses to outbreaks. To address this issue, an AI-powered mobile application integrated with the AGHUB EXTENSION PORTAL [
23] is currently in its developmental stages to assist farmers in the early detection of these threats.
This mobile application will allow farmers to capture images of affected crops using mobile phones. These images are then uploaded to the AGHUB EXTENSION PORTAL, where an advanced AI system analyzes them to identify potential pests or diseases. Once the AI has processed the image, it provides the farmer with a preliminary diagnosis, including the suspected pest or disease and its characteristic features.
The image data and AI-generated diagnosis are transmitted to the Ghana Plant Protection and Regulatory Services Directorate (PPRSD) through the AGHUB EXTENSION PORTAL to enhance accuracy and inform appropriate response strategies. PPRSD experts carefully examine the information, validating the AI’s findings and conducting further investigations as needed. Based on the confirmed pest or disease, PPRSD assesses the outbreak’s potential impact and risk level.
In collaboration with other relevant government agencies, PPRSD develops comprehensive response strategies. These strategies include recommendations and mitigation measures, which are then communicated to farmers and other stakeholders through the AGHUB EXTENSION PORTAL and other available channels. The various steps involved in an emergency response scenario are illustrated in
Figure 20.
By empowering farmers with AI-driven tools and facilitating rapid response through the AGHUB EXTENSION PORTAL, this system aims to improve the overall management of pests and diseases in Ghana’s agricultural sector, leading to increased crop yields and enhanced food security.
6. Conclusions
This work addresses the challenge of crop disease detection by utilizing Edge Impulse to deploy multiple deep-learning models, including MobileNet, Inception, ShuffleNet, Squeezenet, EfficientNet, and a custom Deep Neural Network (DNN) on a mobile device, facilitating the rapid and precise detection of tomato leaf diseases (Leaf blight, Leaf curl, Septoria leaf spot, and Verticillium wilt) in the field.
The results demonstrate that deploying multiple deep learning models on a single edge device for tomato leaf disease detection is feasible, with models like MobileNet and ShuffleNet showing strong potential due to their efficiency and performance. However, models such as Inception and EfficientNet require further optimization to address issues of accuracy and latency. At the same time, the custom DNN provides a favorable balance of performance and efficiency for practical deployment. Future work should optimize these models through techniques like pruning and quantization to enhance their real-time performance and reduce latency. Improving dataset quality by expanding and diversifying training data to cover different geographic regions and conditions will bolster model robustness and generalizability. Also, we plan to collaborate with the Ghana Ministry of Food and Agriculture to deploy the system on large farms and measure the model’s accuracy, power consumption, and performance in varying environmental conditions, which will reveal some practical challenges in its deployment. Exploring real-time adaptation mechanisms and integrating models with edge devices will ensure better applicability and performance. Moreover, developing user-friendly interfaces for end-users can facilitate practical deployment, offering real-time feedback and actionable insights to enhance disease management in agricultural settings.