1. Introduction
Citrus trees in Florida are currently facing wilting due to multiple factors such as a range of plant diseases, hurricanes, winter freeze, etc. In 2022–2023, citrus production in Florida totalled 18.1 million boxes, which marked a decrease of 60% from the 2021–2022 season’s 45.3 million boxes [
1]. Citrus diseases, affecting the health of both leaves and fruits, continue to pose a threat to the productivity and economic stability of citrus growers across the United States due to their symptomatic effects. Huanglongbing (HLB), or citrus greening disease, being lethal to citrus trees and lacking a cure, poses a greater concern to Florida than to other states, and is transmitted by the Asian citrus psyllid [
2]. Citrus canker was detected in Florida a few decades ago and has re-emerged several times, yielding destruction of trees [
3]. It stands out as a prevalent threat, caused by a bacterial pathogen, inducing necrotic lesions on leaves, stems, and fruit [
4]. Infected leaves exhibit irregular shaping, stunting, or puckering. Severe cases of infection can lead to adverse outcomes such as leaf loss, premature fruit shedding, twig deterioration, overall tree debilitation, and blemishes on the fruit. Greasy spot, triggered by
Mycosphaerella citri, and citrus melanose, induced by
Diaporthe citri, also contribute to the array of challenges faced by citrus trees [
5]. Scab, initiated by the fungus
Elsinoe fawcettii, poses a notable threat to lemon varieties, Rangpur lime, and rough lemon rootstocks, resulting in irregular outgrowths resembling scabs or warts on fruit, leaves, and twigs, often accompanied by leaf distortion [
6]. In Florida, zinc deficiency manifests initially as small yellow blotches between green veins on leaves, progressing to overall yellowing, except for green venous areas, with worsening deficiency [
7]. Not only do citrus diseases impact production, but adverse weather conditions such as hurricanes, winter freezes, light snow, sleet, and freezing rain, and shortage of labor also contribute to its reduction. Consequently, safeguarding the profitability and marketability of citrus hinges on effective control and identification of infected fruits. It is crucial to detect disease infections with reliability and accuracy, as these can be prevented by human intervention, unlike weather conditions which are uncontrollable. However, the general detection methods rely on manual inspection, which is prone to human error, or conventional methods such as electron microscopy or polymerase chain reaction, which are time consuming, expensive, and need laboratory skills. When considering a long-term final tactic for disease detection, automated systems capable of accurately and tirelessly classifying citrus diseases at faster speeds are necessary.
Most studies on leaf disease detection methods focus on visual or color digital imaging by utilizing image classification and recognition techniques. HSI systems have proven to be reliable, accurate, and non-destructive methods, allowing for in-depth spectral analysis with high spectral resolution. Moreover, fluorescence imaging systems have been widely employed to evaluate the physiological characteristics of plant leaves under environmental stress or disease conditions. When the specific wavelength excites the chlorophyll molecules, chlorophyll emits lights at the different ranges of wavelengths. These unique spectral signatures characterizing disease presence across a broad spectrum allow for the performance of disease detection and classification tasks due to the HSI’s ability [
8,
9]. Previous research has explored multispectral detection to assess the effects of citrus diseases on grapefruits, achieving a classification accuracy of 95.7% [
10]. The research demonstrated the potential of two-band multispectral imaging, though it acknowledged the risk of missing small-sized canker lesions. Thomas et al. also explored a combined reflectance and transmission hyperspectral imaging system for investigating plant pathogens [
11]. Their findings suggested that reflectance-based measurements facilitate early detection, while transmission measurements provide supplementary insights into understanding and quantifying the complex spatio-temporal dynamics of plant–pathogen interactions. Bauriegel and Herppich utilized hyperspectral and chlorophyll fluorescence imaging systems for early detection of Fusarium head blight in wheat [
12]. They measured the Fusarium using a hyperspectral imaging system in a laboratory environment, and using a chlorophyll fluorescence imaging system in the field. The existing hyperspectral and fluorescence systems require complex setup and operate individually, resulting in longer data collection times and the need to manually align data, leading to lower accuracy.
In addition, accurately distinguishing between healthy and infected leaves or between leaf diseases with similar visible symptoms remains a challenging task. In recent years, machine learning methods, including support vector machine (SVM), k-nearest neighbor (KNN), artificial neural networks (ANN), and convolutional neural networks (CNN), have expedited and advanced automated systems, ensuring promising outcomes in disease detection tasks [
13,
14]. In the agricultural field, HSI combined with machine learning has been demonstrated as a highly accurate method for the detection and classification of diseases on various leaves such as citrus [
15,
16], grapevines [
17], mangroves [
18], squash [
19], etc. Xie et al. utilized HSI to detect diseases on tomato leaves and applied extreme learning machine (ELM) models with the full spectrum, as well as a successive projections algorithm (SPA)-ELM for selected wavelengths [
20]. The ELM model achieved a higher accuracy (100%) compared to the SPA-ELM model (97.1%) in the testing data. The SPA-ELM method features a simplified model, shorter calculation time, and the potential for developing multispectral-based detection instruments. Wu et al. employed reflectance HSI with ELM, SVM, and KNN for the detection of gray mold on strawberry leaves [
21]. Three machine learning classifiers achieved high accuracies of over 90% in the test set. Liu et al. also demonstrated the potential of reflectance HSI with machine learning models for assessing apple mosaic disease-related leaf chlorophyll content on apple leaves [
22]. The highest accuracy (98.89%) from validation data sets was achieved using a random forest model, based on several wavelengths, average leaf chlorophyll content, and their combinations. Chen et al. investigated a fluorescence imaging system with machine learning models, a convolutional neural network for transfer learning employing ResNet50 for the identification of leaves infected by cucumber downy mildew [
23]. The chlorophyll fluorescence parameters resulting from plant stress due to disease provide valuable information surrounding the physiological characteristics of plants. The parameters were utilized as inputs for deep learning models, with the enhanced ResNet model achieving an accuracy of 94.76% and facilitating early disease detection. Weng et al. performed a chlorophyll fluorescence measurement on leaves using three LEDs at 650 nm for citrus HLB [
24]. The least squares SVM achieved an accuracy of 95.0% for Navel oranges and 96% for Satsuma.
While there have been numerous studies investigating solely reflectance or fluorescence based hyperspectral measurement, there has been limited research on the potential use of simultaneous reflectance and fluorescence imaging in a single system for identifying leaf diseases. Employing simultaneous hyperspectral reflectance and fluorescence imaging could offer advantages in distinguishing diseases (e.g., high classification accuracy). This study explored the potential of a portable hyperspectral reflectance and fluorescence imaging system combined with machine learning to distinguish various diseases on citrus leaves. The front and back sides of leaves infected with six different diseases, as well as a healthy control set, were tested using the hyperspectral imaging system. The hyperspectral leaf images were investigated with machine learning classifiers.
2. Materials and Methods
2.1. Portable Hyperspectral Imaging System
A portable hyperspectral imaging system was developed for this study. The system consists of two LED sources: a visible and near-infrared (VNIR) broadband light for reflectance, and an ultraviolet-A (UV-A) excitation light for fluorescence (Metaphase Technologies, Bristol, PA, USA) (
Figure 1). The wavelengths of the VNIR LEDs are set at 428, 650, 810, 850, 890, 910, and 940 nm, whereas the wavelength of the UV-A is set at 365 nm. The LED intensities can be adjusted using two digital dimming controllers, each managing three channels. The first controller handles wavelengths of 365, 428, and 650 nm, while the second controller manages wavelengths of 810, 850, 890, 910, and 940 nm. The line beams are generated by the rod focal lens, focusing on the leaf sample on the sample holder, and are overlapped by tilting 6° from the vertical position. The sample holder was designed with dimensions of 254 × 197 × 15 mm
3 and printed by a 3D printer (F370, Stratasys, Eden Prairie, MN, USA) with black thermoplastic. A customized reflectance standard panel with a size of 254 × 32 × 15 mm
3 (Labsphere, North Sutton, NH, USA) was side mounted to the sample holder for flat-field correction of the reflectance images. A miniature line-scan hyperspectral camera (Nano-Hyperspec VNIR, Headwall Photonics, Bolton, MA, USA), together with a wide-angle lens with 5 mm focal length (Edmund Optics, Barrington, NJ, USA) and a long-pass (>400 nm) gelatin filter (Kodak, Rochester, NY, USA), consisting of an imaging spectrograph and a CMOS focal plane array detector (12-bit and 1936 × 1216 pixels) was used to acquire reflectance and fluorescence signals. The long pass filter was used to eliminate second-order effects from the UV-A excitation. To block the ambient light, an aluminum-framed enclosure with dimensions of 56 × 36 × 56 cm
3 with black aluminum composite boards was used to enclose the LED lights, the camera, the sample holder, the reflectance panel, and the moving stage, whereas power supplies and controllers were mounted outside. The components (i.e., two lights, a camera, and a stage) were connected to the laptop through a powered four-port USB hub. The entire system was mounted on a compact 45 × 60 cm
2 optical breadboard, which was suitable for on-site and field experiments.
2.2. System Software and Operation
The system software developed by LabVIEW (v2022, National Instruments, Austin, TX, USA) on a Windows 11 laptop computer is shown in
Figure 2. To operate the system, software development kits (SDKs) from hardware manufacturers with LabVIEW were applied to implement parameterization and data transfer functions. The SDKs included User Datagram Protocol (UDP) for LED light control, Universal Serial Bus (USB) for camera control, serial communication for stage movement control, and the LabVIEW Vision Development Module (VDM) for image and spectrum display. During a measurement, the VNIR line light was turned on for 10 s to stabilize LED output. The translation stage then moved the sample holder leftward, while the hyperspectral camera collected line-scan reflectance signals from the standard panel and passed samples. When the sample holder completed its pass, reflectance image acquisition ended, and the VNIR light was turned off. Subsequently, the UV-A line light was turned on for 10 s to stabilize. The camera then acquired line-scan fluorescence signals as the stage returned to its starting position, turning off the UV-A light upon completion, concluding a full imaging cycle.
The spatial resolution along the translation direction for a predetermined scan distance was dependent on the stage moving speed and the number of total scans. For instance, when the number of total scans was 250 lines over a distance of 250 mm, with the stage moving at 3.3 mm/s, it took about 76 s, resulting in an approximate spatial resolution of 1 mm/pixel. The stage moving speed was adjusted to synchronize continuous line-scan image acquisition with the translation stage movement, based on the exposure time of the camera. Empirically, a reciprocal relationship between moving speed (V in mm/s) and exposure time (T in s) was determined (V = 0.99/T), leading to speeds of 3.3 mm/s for an exposure time of 0.3 s and 1.65 mm/s for an exposure time of 0.6 s. In addition to the continuous moving mode, the hyperspectral system was able to operate in an incremental step-by-step line scanning mode. The software displayed reflectance and fluorescence images along with an original spectrum and spatial profile, updating them line by line to show real-time scan progress while the image was being acquired. After completion, each of the images for reflectance and fluorescence were saved into separate data files in standard band interleaved by line (BIL) format.
2.3. Sample Preparation and Measurement
Mature leaves from Valencia orange trees, exhibiting symptoms of canker, HLB, greasy spot, melanose, scab, and zinc deficiency—each with only one condition and excluding any with multiple infections—were sampled at the Citrus Research & Education Center (CREC) of the University of Florida in Lake Alfred, FL. The leaf samples were handpicked and placed in zipper bags in a cooler, which were then transferred to the laboratory within 4 h. These leaves were refrigerated and imaged within 48 h of collection. The diseases were identified based on visual inspection of the symptoms on each leaf by a plant pathologist. Depending on leaf size, 4 to 16 leaves from the same class were imaged simultaneously. Both the front and back sides were scanned. The total numbers of leaves in the control, canker, HLB, greasy spot, melanose, scab, and zinc deficiency classes were 107, 105, 101, 114, 117, 73, and 125, respectively. The citrus leaf samples were placed on the sample holder. The sample holder and the reflectance panel were moved by a linear motorized stage (FUYU Technology, Chengdu, China) within the scope of the hyperspectral camera for line-scan image acquisition. The lens-to-sample distance was fixed at 285 mm, where the spatial resolution along the camera scanning line direction was 0.33 mm/pixel. Each camera frame captured a region of interest (ROI) of 810 × 348 pixels (spatial × spectral) to address a 270 mm long instantaneous field of view and a spectral region from 395 to 1005 nm.
2.4. Hyperspectral Image Processing and Machine Learning
Hyperspectral reflectance and fluorescence images of each leaf were acquired from the standard panel using 810 spatial pixels and 250-line scans over a 250 mm distance. This resulted in the creation of two 810 × 250 × 348 (X × Y × λ) raw hypercubes. A flowchart summarizing data processing and machine learning procedures is shown in
Figure 3. Initial data smoothing was achieved by averaging groups of three neighboring pixels in both spectral dimension (λ) and spatial dimension (X) along the camera’s scanning line, producing two 270 × 250 × 116 reduced hypercubes with a spatial resolution of approximately 1 mm/pixel in both the X and Y dimensions and a spectral interval of 5.3 nm. VNIR reflectance values from the standard reference panel (i.e., 100%) were used for flat-field correction to convert the intensity values in the averaged image to relative reflectance values (0–100%). A mask image was generated from a single-band reflectance image at 815 nm where leaves exhibited high reflectance, using a single threshold value computed with Otsu’s method. This mask image was then utilized to remove the background from both reflectance and fluorescence images acquired from the same leaf samples. The hyperspectral image processing procedures were executed using in-house programs developed in MATLAB (R2024a, MathWorks, Natick, MA, USA). Following this, all leaf pixels in the masked reflectance image at 815 nm were grouped into 5 × 5-pixel windows to reduce spectral variation and the amount of data required for machine learning. For each window, the mean (M) and standard deviation (STD) of the leaf pixel intensities were calculated. Any window where more than 10% of the 25 pixels fell outside the range of M ± 3STD was excluded from further analysis. The remaining pixel windows had their spectra averaged in the spatial domain while maintaining full spectral resolution. These mean spectra were then used for machine learning classification. This segmentation method produced the reflectance, the fluorescence, and the combined reflectance and fluorescence, which were used to develop algorithms for leaf diseases classification. The spectral pixel-based data were labeled according to each type of leaf disease and input into the Classification Learner app in MATLAB (R2024, MathWorks, Natick, MA, USA) to assess the effectiveness of each spectral measurement in disease detection. Furthermore, the averages of the spectral pixel-based data for individual leaves, called leaf-based, were labeled as each type of leaf disease to compare the pixel-based results. To reduce data dimensionality and enhance computational efficiency, principal component analysis (PCA), consisting of the first ten PCA components, was applied. Both the full spectra and the scores from the PCA were used for machine learning classification, and their accuracies were compared. Optimizable machine learning classifiers from nine general categories (including naive Bayes, efficient linear, trees, kernel, ensemble, KNN, discriminant, neural network, and SVMs) were tested to evaluate classification performance. For example, the SVMs utilized hyperplanes to divides the classes with a large margin and used kernel to handle complex dataset. The larger margin with few support vectors decreased the classification performance, resulting in underfitting. To streamline the evaluation of misclassification costs and model training, equal penalties were assigned to all leaf disease misclassifications, and default hyperparameters from the MATLAB Classification Learner app were used. For each class, a 70/30 split was used to create the training and test datasets. For the training dataset, five-fold cross-validation was employed to minimize overfitting and assess the models’ generalization abilities. Each spectral dataset was randomly split into five equal-sized folds, with models trained on four folds and validated on the other. This method ensured reliable evaluation of classification performance by averaging the errors across the five folds.
2.5. Performance Assessment
The confusion matrix plot was utilized to evaluate the performance of the currently selected classifier across each class. Accuracy, precision, recall, and F1 score for each class were applied as parameters that indicated how well the model was functioning in distinguishing between classes. The precision indicated the accuracy of positive predictions, showing how many of the predicted positive cases were positive. Recall measured the ability of the classifier to correctly identify positive cases, showing how many of the actual positives were identified. F1 score is a harmonic mean of precision and recall, providing a single metric that balances both measures, particularly useful in uneven class distribution. It also provides class-wise performance metric rather than the overall performance of all the classes. Accuracy, precision, recall, and F1 score can be represented by the formulas for these metrics in terms of false positives (FP), true positives (TP), false negatives (FN), and true negatives (TN).
4. Discussion
In this research, the portable hyperspectral imaging system provided the reflectance and fluorescence spectral data for the leaf samples to analyze the citrus disease infections. The absorption in the reflectance spectra occurs at 675 nm and the emission peaks in the fluorescence spectra arise at 691 nm and 744 nm for the front side and 738 nm for the back side, where chlorophyll A absorbs most [
25]. These findings align with those reported by Chappelle et al. (1992), who showed the chlorophyll content in the reflectance [
26]. The variation between the emission peaks in fluorescence at 744 nm and 738 nm might be because of the different chlorophyll distribution caused by the different amounts of sunlight reaching the front and back side of the leaves. Furthermore, 686 nm is related to photosystem II, and the 700–750 nm range is correlated with both photosystems II and I (PSII and PSI), which are integral pigment–protein complexes crucial in the primary stages of photosynthesis [
27]. The infected leaves, except those afflicted with scab, showed reduced intensities around 750 nm due to the disturbance of PSI. Between 400 and 650 nm, there was a slightly higher peak around 550 nm in the infected leaves as compared to the control due to the onset of disease.
The machine learning classifiers were applied to analyze the hyperpectral data and classify the infected citrus leaves. The PCA analysis shows the different patterns compared to the full spectra analysis. The fluorescence data for the front and back side of the leaf demonstrated higher classification accuracy than the combined reflectance and fluorescence data using most cases of the classifiers except naïve Bayes and efficient linear and discriminant analysis. The classification accuracies using the reflectance data were lower than those using the combined reflectance and fluorescence data. Overall, the gap between the lowest and highest accuracies in PCA analysis was not as large as in the full spectra. The PCA analysis also exhibited a similar pattern, manifesting as a low gap between the accuracies from the classifiers among the pixel-based results. Leaf-based classification accuracies provide higher accuracy, and may be beneficial for faster computation and improved generalization, despite the potential bias and loss of detail.
The combined reflectance and fluorescence showed superiority over individual reflectance and fluorescence, regardless of pixel-based and leaf-based analysis. The full spectra for both pixel-based and leaf-based analyses provided higher accuracies than PCA. This suggests that more bands provide more information and insights about the leaf conditions not visible in solely the reflectance or fluorescence images. Furthermore, the best accuracies of 90.7% and 94.5% for pixel-based and leaf-based analysis on the back side of the leaves were slightly higher than those of 88.0% and 90.9% on the front side, respectively. Depending on the type of disease, the impact of the infection on either the front or back side of the leaves varies. Specifically, the back sides of leaves have high humidity and moisture. Therefore, they are more vulnerable to specific bacteria, viruses, or pathogens that favor those conditions, and are possibly exposed to a higher severity of conditions than the front side. Most of the diseases, including greasy spot, are more prevalent in humid and wet conditions. The results imply that the back side of a leaf encounters higher disease severity than the front side, resulting in higher accuracy in classification.
The classification of healthy and infected plant leaves using machine learning and deep learning methods has gained attention, reflecting the growth of AI in this field. Our findings showed that the machine learning models provided accuracies between 90% and 95% for the classification of citrus leaf diseases. Previous work by Singh et al. used deep learning based models to classify bean rust disease and angular leaf spot disease, showing 91.74% accuracy [
28]. Paul et al. utilized convolutional neural network (CNN) models and transfer learning based-models VGG-16 and VGG-19 to classify tomato leaf diseases, reporting a highest accuracy of 95% [
29]. Nikith et al. compared CNN models with KNN and SVM models for leaf diseases such as bacterial blight, brown spot, powdery mildew, etc. [
30]. The CNN model showed an accuracy of 96%, whereas the KNN and the SVM models achieved 64% and 76%, respectively. Overall, the CNN method often provides a higher degree of accuracy than machine learning models, such as SVM. However, both machine learning and deep learning methods are powerful tools for leaf disease classification. Deep learning methods are suitable for complicated high-dimensional data, whereas machine learning methods and the application of PCA are straightforward and have greater interpretability. Therefore, machine learning models are not too complicated, but remain effective in interpreting leaf disease classification.