Next Article in Journal
Analysis of the Effectiveness of a Freight Transport Vehicle at High Speed in a Vacuum Tube (Hyperloop Transport System)
Next Article in Special Issue
A Multi-Stage Method for Logo Detection in Scanned Official Documents Based on Image Processing
Previous Article in Journal
Computing RF Tree Distance over Succinct Representations
Previous Article in Special Issue
A Lightweight Graph Neural Network Algorithm for Action Recognition Based on Self-Distillation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effect of the Light Environment on Image-Based SPAD Value Prediction of Radish Leaves

1
United Graduate School of Agricultural Science, Tokyo University of Agriculture and Technology, Tokyo 183-8509, Japan
2
Institute of Agriculture, Tokyo University of Agriculture and Technology, Tokyo 183-8509, Japan
3
Research Center for Agricultural Information Technology, National Agriculture and Food Research Organization, Tokyo 105-0003, Japan
*
Author to whom correspondence should be addressed.
Algorithms 2024, 17(1), 16; https://doi.org/10.3390/a17010016
Submission received: 2 November 2023 / Revised: 22 December 2023 / Accepted: 26 December 2023 / Published: 29 December 2023
(This article belongs to the Special Issue Recent Advances in Algorithms for Computer Vision Applications)

Abstract

:
This study aims to clarify the influence of photographic environments under different light sources on image-based SPAD value prediction. The input variables for the SPAD value prediction using Random Forests, XGBoost, and LightGBM were RGB values, HSL values, HSV values, light color temperature (LCT), and illuminance (ILL). Model performance was assessed using Pearson’s correlation coefficient (COR), Nash–Sutcliffe efficiency (NSE), and root mean squared error (RMSE). Especially, SPAD value prediction with Random Forests resulted in high accuracy in a stable light environment; CORRGB+ILL+LCT and CORHSL+ILL+LCT were 0.929 and 0.922, respectively. Image-based SPAD value prediction was effective under halogen light with a similar color temperature at dusk; CORRGB+ILL and CORHSL+ILL were 0.895 and 0.876, respectively. The HSL value under LED could be used to predict the SPAD value with high accuracy in all performance measures. The results supported the applicability of SPAD value prediction using Random Forests under a wide range of lighting conditions, such as dusk, by training a model based on data collected under different illuminance conditions in various light sources. Further studies are required to examine this method under outdoor conditions in spatiotemporally dynamic light environments.

1. Introduction

In the cultivation of root crops such as radish and carrots, it is difficult to control the production environment while checking the condition of roots. Therefore, production environment control for root crops requires the estimation of subsurface information from aboveground information. It has been proven that chlorophyll content in leaves is related to nitrogen content in crops and can be used as an indicator to control the production environment during crop cultivation [1,2,3]. Conventional chemical analysis of chlorophyll content is labor-intensive, destructive, and limited to the small measurement area. In contrast, the chlorophyll value (SPAD value) measured by a chlorophyll meter (SPAD-502Plus, Konica Minolta, Inc., Tokyo, Japan) can accurately evaluate the chlorophyll content in plant leaves in a nondestructive manner. SPAD values can not only indirectly measure the nitrogen content in crops but are also related to plant growth indices and yield [3,4,5], and it has been reported that water stress during crop cultivation causes SPAD values to increase [6]. However, the measurement area for SPAD values using existing SPAD meters is still limited, and only one leaf can be measured at a time. Therefore, a method to predict leaf SPAD values on a large scale and in a short time would greatly contribute to the optimization of production environment control and cultivation management.
With the development of color image acquisition technology, color information obtained from digital cameras is now widely used to predict the physiological characteristics of crops, such as the vegetation index (VI) [7,8,9], shape analysis [10], and leaf nitrogen and chlorophyll content estimation [11]. For example, VI based on the RGB color system has been integrated with crop height information and used for yield prediction [9]. HSV and HSL color systems, which are obtained by converting from the RGB color system, are also used for semantic segmentation [12], such as georeferencing and crop quality monitoring [13]. There are also studies using multispectral cameras [14,15] and hyperspectral cameras [16,17], but these methods require high equipment costs.
In previous studies, algorithms such as Linear Regression [7], convolutional neural networks (CNN) [18], Support Vector Machines [3], Random Forests [3,19], XGBoost [19], and Generalized Linear Models (GLMs) [20] have been applied to predict SPAD values using color values measured by images and color chromatometers. Although leaf colors vary from crop to crop and direct comparisons of model performances are not possible, some algorithms have achieved high model performance in predicting SPAD values. Therefore, if we can appropriately link leaf color to SPAD values using machine learning algorithms, we can accurately estimate SPAD values from leaf color in real time. To achieve this, it is important to select appropriate input variables for each algorithm.
Images acquired with a digital camera are easily affected by light sources such as brightness and color temperature [8]. Therefore, it is desirable to acquire images using digital cameras under stable light conditions. However, because the intensity of sunlight, for example, changes from moment to moment, the pixel values of an image may change spatiotemporally, even for the same subject. In response to this problem, machine learning was applied to predict SPAD values from pixel values using a photographic device that creates a stable light environment [18]. However, similar to conventional SPAD meters, this device can only measure the SPAD values of individual leaves. In agriculture, there is a need to efficiently acquire a large amount of information about crops in the fields or greenhouses. Therefore, new technologies are needed to measure SPAD values on many leaves at once. However, when measuring crops using images using UAVs for instance, it is often affected by the light environment. Therefore, by quantifying and clarifying the influence of the light environment on images during crop measurement, novel methods to measure the SPAD value of many leaves at once can contribute to the development of efficient and productive crop production.
In this study, we constructed a SPAD value prediction model using Random Forests [21], XGBoost [22], and LightGBM [23] based on RGB values obtained from images of radish leaves, as well as HSV and HSL values obtained by conversion from RGB values, and color temperature and illuminance as input variables. We focus on the effect of the light environment at the time of photography on the pixel values of digital images and model the relationship between leaf color and SPAD values under varying photographic conditions with respect to the color temperature and illuminance of the light source. Furthermore, based on information such as the variable importance calculated by Random Forests, we examined the influence of photographic conditions on the prediction of SPAD values for agriculture where domain knowledge remains qualitative.

2. Materials and Methods

2.1. Plant Material

The material tested was leaves cut at 1.5 cm in diameter (leaf patches) from a radish produced by the JA Toyohashi Omura gardening club purchased at a supermarket in Tsuzuki Ward, Yokohama City, Kanagawa Prefecture, Japan, on 1 April 2023.

2.2. Photographic Environment and Data Acquisition Methods

The images were taken using a box-type device with a height of 900 mm, width of 460 mm, and depth of 460 mm, constructed on a flat room floor. A digital camera (DC-GX7MK3, Panasonic Corp., Tokyo, Japan) was fixed on a tripod so that the camera could capture images of leaves from above for expected applications in image-based SPAD value predictions in an open field using a UAV or the use of a fixed-point camera for the greenhouse. The camera parameters were F-value 7.1, ISO 200, and focal length 12 mm. In this photographic environment, the color temperature and light intensity can be controlled by switching the lighting using a dimmer. In this study, halogen bulbs (color temperature: 3000 K) and LED bulbs (color temperature: 5000 K) were used, with nine levels of illuminance for the halogen bulbs (3, 50, 100, 300, 500, 800, 1000, 2000, and 2700 lx) and nine levels of illuminance for the LED bulbs (120, 300, 500, 800, 1000, 2000, 3000, 4000, and 5440 lx). Figure 1 shows a schematic diagram of the experimental device to capture images of radish leaves.
In this study, leaf patches of radishes were prepared, and the color information was obtained for an image of a leaf patch with a circularity greater than 0.85 (Figure A1). The minimum, mean, median, and maximum values of the RGB, HSL, and HSV color systems (see Section 2.3 for detail) were calculated for each leaf patch. The RGB, HSL, and HSV values of the center of gravity of the leaf patches were calculated for each leaf patch and used as input variables in the Random Forests analysis (described in Section 2.4), together with the color temperature and illuminance of the photographic environment, to predict the SPAD values.

2.3. Color Information Collection Methods

Color information (RGB values, as well as HSL and HSV values calculated based on RGB values) was obtained from the images captured from the leaf patches. HSL values are values in the HSL color system, where H, S, and L represent hue, saturation, and lightness, respectively. The values of H and S in the HSL color system are reported to be indices that are theoretically unaffected by the intensity of illuminance, because illuminance information is concentrated only in L [13]. Therefore, it is possible to reduce the influence of the light source on the photographic environment.
The HSL color system can be converted from the RGB color system in a cubic space through the icosahedral hue, saturation, and intensity (HSI) color system, which is similar to the HSL color system and the HSV color system. The difference between the HSL and HSV color systems lies in the color when S (saturation) is reduced. The HSL color system ranges from 0 to 360 for H, 0 to 100 for S, and 0 to 100 for L. The HSV color system ranges from 0 to 360 for H, 0 to 100 for S, and 0 to 100 for V. In this study, HSL and HSV values were calculated from each RGB value using a colormap in a Python library and normalized to the range of 0 to 1 in each value range.

2.4. Modeling with Random Forests, XGBoost, and LightGBM

Random Forests [21] is an ensemble learning algorithm that integrates and analyzes multiple classification and regression trees. First, the algorithm extracts a large number of bootstrap samples. Then, it generates a decision tree for each bootstrap sample. These decision trees only use a small number of randomly selected variables as features. The final output is the mean value for regression and the majority vote for classification, as the results are obtained from multiple decision trees. Decision trees have the problem that, the deeper the tree, the more complex the structure becomes, and the more prone to overfitting. Random Forests, however, have better generalization ability than decision trees because of bagging, which alleviates the problem of overfitting. XGBoost [22] and LightGBM [23] are both algorithms based on gradient boosting trees. First, a decision tree is created, and the error between the predicted and observed values is calculated from the initial prediction. Subsequently, a new decision tree is created so that the loss is reduced using the gradient descent method. The final output is the sum of the weights of the leaves to which the predicted data belong to each decision tree. XGBoost uses a level-wise search, whereas LightGBM uses a leaf-wise search. XGBoost is computationally slow, whereas LightGBM has the disadvantage of overfitting. Importantly, Random Forests, XGBoost, and LightGBM allow an evaluation of the importance of each input variable for model interpretation (see [24] for Random Forests).
In this study, we employed Random Forests, XGBoost, and LightGBM algorithms to build data-driven models with 17 different sets of input variables to predict SPAD values based on environmental information (color temperature and illuminance) (Figure 2). The 17 models consisted of eight models that used data collected under halogen and LED light sources (HAL+LED), three models that used data collected under a halogen light source only (HAL), three models that used data collected under a LED light source only (LED), and three models that used data collected under a LED light source based on illuminance information (LED_RES). The LED_RES condition is a condition where the indoor lighting environment has low light intensity compared to sunlight, and thus, it was set up to verify the applicability of the prediction models for the extrapolated conditions.
The input variables to Random Forests, XGBoost, and LightGBM were the color information and the light environment in which the image was captured. The color information is the minimum, mean, median, and maximum values of each color component in the RGB, HSL, and HSV color systems, and the values at the center of gravity of the leaf patch. Color information GB is a model that uses the minimum, mean, median, and maximum values of the G and B values in the RGB color system and the values at the center of gravity of the leaf patch. The reason for constructing a model that uses only GB values as input variables for color information is to exclude the R values for red light that can be absorbed by water molecules. The input variables for the lighting environment were the color temperature of the light source (LCT) and illuminance of the light source (ILL).
In this study, we implemented Random Forests computation using “Scikit-learn” [25], XGBoost computation using “xgboost”, and LightGBM computation using “lightgbm” in Python. We used the default hyper-parameters of Random Forests, XGBoost, and LightGBM, except for random_state. In this study, a five-fold cross-validation was conducted by changing the random seed (i.e., random_state) 50 times to evaluate the variability of the model structure originating from the randomization process in the computation. Note that a five-fold cross-validation was not performed for LED_RES, because the training and validation data were divided according to the explicit objectives. The model performance was evaluated based on the Pearson’s correlation coefficient (COR), Nash–Sutcliffe coefficient (NSE) [26], and root mean squared error (RMSE) between the observed SPAD values and model output values. COR evaluates the correlation between the observed and predicted SPAD values of the model (Equation (1)). NSE considers the balance between the sum of the squared errors and the variance of observed data (Equation (2)). RMSE evaluates the error between the predicted and observed SPAD values at the dimension of the observed SPAD value (Equation (3)).
C O R = N i = 1 N ( Y o , i Y p , i ) i = 1 N Y o , i i = 1 N Y p , i N i = 1 N Y o , i 2 ( i = 1 N Y o , i ) 2 N i = 1 N Y p , i 2 ( i = 1 N Y p , i ) 2
N S E = 1 i = 1 N ( Y o , i Y p , i ) 2 i = 1 N ( Y o , i Y o ¯ ) 2
R M S E = 1 N i = 1 N ( Y o , i Y p , i ) 2
where Y o , i and Y p , i are the observed and predicted SPAD values for the data point i   ( = 1 ,   2 ,     ,   N ) , Y o ¯ is the mean observed SPAD value, and N is the size of the dataset. Specifically, the NSE takes a value in the range [ , 1 ] , of which a NSE value of 1 means a perfect fit. For interpreting the model of the best score in all three algorithms, SHapley Additive exPlanations (SHAP) [27], Partial Dependence (PD), and Individual Conditional Expectation (ICE) were employed to visualize the model structures in this study. SHAP decomposes the difference between the predictions for a particular instance and the mean prediction into the contribution of each feature based on the Shapley value concept of cooperative game theory. However, it cannot explain “how the predictions react to changes in feature values”, and thus, PD and ICE were used together to compensate this part of the interpretability. While SHAP evaluates the importance of the input variables in SPAD value prediction, PD and ICE can be used to interpret how the data-driven SPAD value prediction model responds to a different light environment.

3. Results

3.1. Modeling Results

Random Forests, XGBoost, and LightGBM could predict the SPAD values by capturing the color changes in different light environments (Figure 2). Among the three algorithms, Random Forests showed the highest accuracy in SPAD value prediction. Therefore, we will focus on the results from the Random Forests computations in the subsequent analyses. Figure 3 shows a scatter plot between the observed and predicted values of the best model with Random Forests for the five-fold cross-validation, and Figure 4 shows the results of the best model with Random Forests for illuminance-limited training data (LED_RES). The accuracy of both models was high, with a COR of 0.85 or higher.
In the analysis integrating halogen and LED light source data (HAL+LED), the model (HAL+LED_RGB+LCT+ILL) with the minimum, mean, median, and maximum values in the RGB color system and color temperature and illuminance as the model input had the highest accuracy in predicting the SPAD values (Figure 3b). In the same model, the prediction accuracy was improved by adding color temperature and illuminance information as the model input. However, the underestimation of high SPAD values did not improve. In the analysis of the halogen light source (HAL) data, the model using RGB values and illuminance as the model input (HAL_RGB+ILL) exhibited the highest prediction accuracy (Figure 3l). The model using the HSL values and illuminance as the model input (LED_HSL+ILL) had the highest accuracy (Figure 3j). Among the models constructed in this study, the model with the HSL value and illuminance as the model input (LED_RGB+ILL) using data obtained using LED light sources had the highest accuracy when COR was used as the performance measure (Figure 3j).
In this study, the prediction accuracy was high, with a COR of 0.85 or higher even in the case of limited illuminance, and the applicability was verified under extrapolated conditions (LED_RES). In this case, the highest prediction accuracy was obtained when the HSL value was used as the model input, as in the case of data obtained under LED light sources (Figure 4b).

3.2. Variable Importance of the Random Forests

Figure 5 shows the top 10 mean absolute SHAP values for data collected under halogen and LED light sources (HAL+LED), with RGB values, illuminance, and color temperature as the model input (HAL+LED_RGB+LCT+ILL; Figure 5a); for data taken under LED light sources (LED) with RGB values and illuminance as the input values (LED_RGB+ILL; Figure 5b); and for data taken under HSL values and illuminance (LED_HSL+ILL; Figure 5c). The figure compares the differences in light source information and color information among the three models from the top five CORs of the Random Forests in Figure 2, excluding the two models using illuminance-limited data and the model using GB values. In the HAL+LED_RGB+LCT+ILL model, the mean value of G contributed the most (Figure 5a). The LED_RGB+ILL model indicated that the median value of B had the highest contribution to the SPAD value prediction, followed by the mean value of B (Figure 5b). The LED_HSL+ ILL model exhibited the median value of H had the highest contribution to the model (Figure 5c).
For the HAL+LED_RGB+LCT+ILL, LED_RGB+ILL, and LED_HSL+ILL models, the PD and ICE for the top three variables of the average absolute SHAP values for each model are shown in Figure 6. PD and ICE—out of importance, not shown in Figure 6—are shown in Figure A2. SPAD values tended to be higher when the median value of B was low (Figure 6d). SPAD values were higher when G was low (Figure 6a,b), and SPAD values were higher when R was high (Figure A2a,b). For illuminance above 4000 lx, the SPAD values tended to increase slightly, when the illuminance was high (Figure A2c). In addition, the SPAD value tended to be high when the H value was low (Figure 6g). This change in the H value indicates that the SPAD value decreases as the leaf color changes from green to blue in the hue circle. The S and L values did not change significantly with respect to the SPAD values (Figure 6h,i).

4. Discussion

In this study, it was found that the accuracy of SPAD value prediction was enhanced by using color information (e.g., RGB, HSL, and HSV) obtained from digital images and photographic environment information (color temperature and illuminance) as the model input. Previous studies have proposed methods to stabilize the light environment during photographic [18] and color-corrected images using color charts as color standards [28]. However, a method to stabilize the light environment at the time of photography can only predict SPAD values on an individual leaf basis, although it can provide a highly accurate SPAD value prediction. In addition, the color correction of an image is often difficult under certain conditions. In contrast, our method presented the potential of the sensor fusion of a color meter or an illuminance sensor for image-based SPAD value prediction that can be applied for a UAV [7,8,29] or fixed camera [30,31] without limiting the photographic conditions.
No significant decrease in prediction accuracy was observed when the data obtained under both halogen and LED light sources were combined for training. However, the SPAD value prediction by the HAL model was less accurate than the LED model. This suggests that the color temperature affects the image-based SPAD value prediction using Random Forests. The result also suggests that SPAD values can be predicted even under non-white conditions, such as dusk, by collecting and learning data under different illuminance conditions in various light source environments. The high accuracy of the SPAD value prediction for the LED model suggests that the LED light source is a more suitable as a light source for image-based SPAD value prediction. However, if the object is in shadow, the SPAD value prediction may be difficult, or the error in the SPAD value prediction may be large. Therefore, it is necessary to consider a method that uses the ratio of shadows as a model input or a color correction method for acquired images when shadows exist on an object [31,32]. The SPAD value is used as an index for environmental control and as a guideline for harvests when producing crops. Although this study focused on radishes, the proposed experimental method can be applied to other crops as well. However, it is unclear whether the model performance can be higher for different varieties or crops other than radishes, as the leaf colors and SPAD values are different among cultivars. Therefore, a direct comparison cannot be examined, because the radish cultivar used in this study is unknown, as it was purchased at a supermarket. Thus, the application of this study to other cultivars and crops is a future challenge.
Following a previous study where [8] reported the results of a similar analysis excluding the R value, which is a wavelength absorbed by water molecules, we conducted modeling using only GB values, excluding the R values in this study. As a result, the GB-based model also showed high accuracy, with a COR of 0.85 or higher. However, the GB-based model showed a lower model performance than the RGB-based models. This suggests that the GB-based models can exclude the effect of water molecules, but the R values may contain important factors for the SPAD value prediction.
The HSL value obtained under the LED light source, which produced models with the highest model performance in this study, is converted using RGB values. Therefore, the color component may be affected by water molecules in the leaves and may not be an appropriate color space when considering the effect of water molecules in the leaves. However, because the HSL color space and HSV color space can theoretically aggregate illuminance information into L [13] or V, the HSL color system and the HSV color system are considered useful in environments with continuously changing lighting conditions. In contrast, in halogen lighting conditions, the color becomes closer to orange, and thus, it may be difficult to separate color and illumination information in the HSL color space or HSV color space, resulting in lower accuracy of the SPAD value prediction.
In the HAL+LED model (Figure 5a), which was constructed from data under halogen and LED light sources, the G values were important, whereas the B values were important in the LED model (Figure 5b). Previous studies have reported that the blue and green bands contribute significantly to SPAD value prediction [19]. The importance of these variables in this study differed depending on the light source, implying that the important variables for SPAD value prediction vary depending on the light environment. Therefore, it is important to incorporate light environment information, such as the color temperature and illuminance of the light source, for image-based SPAD prediction.
Although this study was conducted in a closed indoor environment, there exists a wide range of fields where image-based SPAD value prediction can be conducted, ranging from fields [28] to greenhouses [30]. The results of this study can be used in environments where the leaf color can be obtained under a stable light source, as in previous studies [18]. However, when images are acquired under unstable light environments, such as in the fields or in greenhouses, the accuracy of the SPAD value prediction becomes an issue. In addition, the maximum illuminance in this study was 5000 lx, whereas the outdoor environment was highly illuminant, ranging from 10,000 to 100,000 lx. Further study is therefore needed to collect observation data for building data-driven models specifically in outdoor environments for the proposed method to be applied to a variety of lighting conditions. Furthermore, although this study predicted the SPAD value using Random Forests, XGBoost, and LightGBM, the accuracy of SPAD value prediction may be improved by applying other algorithms, such as Deep Neural Networks. Furthermore, the accuracy changes depending on the compatibility between the input data and the algorithm. In this study, unlike the results of previous studies [19], the Random Forests algorithm achieved the highest accuracy. Further studies on SPAD value prediction using novel algorithms on other datasets may illustrate how different algorithms perform well on datasets with different features and descriptive statistics.

5. Conclusions

This study was conducted to clarify the influence of the light environment on SPAD value prediction using Random Forests, XGBoost, and LightGBM under different light source environments. Leaf patches with a diameter of 1.5 cm were prepared using Japanese radishes (Raphanus sativus L. var. sativus) photographed under an established stable light environment and used for extracting variables for image-based SPAD value prediction. The variables obtained by the image analysis were the RGB values, which were pixel values, and the HSL and HSV values obtained by color space conversion from the RGB values. In addition, color temperature and illuminance, which were measured separately, were used as input variables. The results showed that, under a stable LED light source, the SPAD values could be predicted using Random Forests with high accuracy (COR 0.929 for RGB, color temperature, and illuminance as input variables; COR 0.922 for HSL, color temperature, and illuminance as input variables). It was also found that image-based SPAD value prediction was possible with relatively high accuracy, even under a halogen light source with a color temperature close to that of dusk (COR 0.895 for RGB and illuminance as the model input, and COR 0.876 for HSL and illuminance as the model input). The results suggest that SPAD values can be predicted under a wide range of lighting conditions, such as dusk, by training a data-driven model using data collected under different illuminance conditions in various light sources. Under LED light sources, the use of HSL values resulted in accurate SPAD value prediction by separating the illumination information into L (COR = 0.972, NSE = 0.944, and RMSE = 2.07). In addition, focusing on the H value, the H value tended to change from blue to green as the SPAD value increased. In this study, the SPAD value prediction was performed in a stable light environment in a closed indoor environment; therefore, the development of a method for outdoor conditions, where the light environment is spatiotemporally dynamic, is a future challenge.

Author Contributions

Y.K. conceptualized this study and conducted the experiment and analysis. Y.K. drafted this paper, and S.F. finalized it. S.F. supervised this research. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the FLOuRISH Fellowship Program for Next-Generation Researchers at Tokyo University of Agriculture and Technology.

Data Availability Statement

Data are contained within the article. We are planning to prepare the dataset available after publishing the entire achievement of the project.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. Examples of leaf patch images photographed under different lighting conditions.
Figure A1. Examples of leaf patch images photographed under different lighting conditions.
Algorithms 17 00016 g0a1

Appendix B

Figure A2. Partial dependence and individual conditional expectations in the best model for SPAD value prediction with Random Forests using RGB values in halogen and/or LED light sources—out of importance, not shown in Figure 6: (a) HAL+LED_RGB+LCT+ILL_R_median, (b) LED_RGB+ILL_R_median, and (c) LED_RGB+ILL_Illuminance value.
Figure A2. Partial dependence and individual conditional expectations in the best model for SPAD value prediction with Random Forests using RGB values in halogen and/or LED light sources—out of importance, not shown in Figure 6: (a) HAL+LED_RGB+LCT+ILL_R_median, (b) LED_RGB+ILL_R_median, and (c) LED_RGB+ILL_Illuminance value.
Algorithms 17 00016 g0a2

References

  1. Wang, Y.; Wang, D.; Shi, P.; Omasa, K. Estimating rice chlorophyll content and leaf nitrogen concentration with a digital still color camera under natural light. Plant Methods 2014, 10, 36. [Google Scholar] [CrossRef] [PubMed]
  2. Netto, A.T.; Campostrini, E.; de Oliveira, J.G.; Bressan-Smith, R.E. Photosynthetic pigments, nitrogen, chlorophyll a fluorescence and SPAD-502 readings in coffee leaves. Sci. Hortic. 2005, 104, 199–209. [Google Scholar] [CrossRef]
  3. Guo, Y.; Chen, S.; Li, X.; Cunha, M.; Jayavelu, S.; Cammarano, D.; Fu, Y. Machine Learning-Based Approaches for Predicting SPAD Values of Maize Using Multi-Spectral Images. Remote Sens. 2022, 14, 1337. [Google Scholar] [CrossRef]
  4. Sakamoto, M.; Komatsu, Y.; Suzuki, T. Nutrient Deficiency Affects the Growth and Nitrate Concentration of Hydroponic Radish. Horticulturae 2021, 7, 525. [Google Scholar] [CrossRef]
  5. Kandel, B.P. Spad value varies with age and leaf of maize plant and its relationship with grain yield. BMC Res. Notes 2020, 13, 475. [Google Scholar] [CrossRef] [PubMed]
  6. Stagnari, F.; Galieni, A.; D’Egidio, S.; Pagnani, G.; Pisante, M. Responses of radish (Raphanus sativus) to drought stress. Ann. Appl. Biol. 2018, 172, 170–186. [Google Scholar] [CrossRef]
  7. Liu, Y.; Hatou, K.; Aihara, T.; Kurose, S.; Akiyama, T.; Kohno, Y.; Lu, S.; Omasa, K. A Robust Vegetation Index Based on Different UAV RGB Images to Estimate SPAD Values of Naked Barley Leaves. Remote Sens. 2021, 13, 686. [Google Scholar] [CrossRef]
  8. Liu, Y.; Hatou, K.; Aihara, T.; Kurose, S.; Akiyama, T.; Kohno, Y.; Lu, S.; Omasa, K. Assessment of naked barley leaf SPAD values using RGB values under different growth stages at both the leaf and canopy levels. Eco-Eng. 2021, 33, 31–38. [Google Scholar] [CrossRef]
  9. Sumesh, K.C.; Ninsawat, S.; Somard, J. Integration of RGB-based vegetation index, crop surface model and object-based image analysis approach for sugarcane yield estimation using unmanned aerial vehicle. Comput. Electron. Agric. 2021, 180, 105903. [Google Scholar] [CrossRef]
  10. Iwata, H.; Niikura, S.; Matsuura, S.; Takano, Y.; Ukai, Y. Evaluation of variation of root shape of Japanese radish (Raphanus sativus L.) based on image analysis using elliptic Fourier descriptors. Euphytica 1998, 102, 143–149. [Google Scholar] [CrossRef]
  11. Lu, J.; Cheng, D.; Geng, C.; Zhang, Z.; Xiang, Y.; Hu, T. Combining plant height, canopy coverage and vegetation index from UAV-based RGB images to estimate leaf nitrogen concentration of summer maize. Biosyst. Eng. 2021, 202, 42–54. [Google Scholar] [CrossRef]
  12. Waldamichael, F.G.; Debelee, T.G.; Ayano, Y.M. Coffee disease detection using a robust HSV color-based segmentation and transfer learning for use on smartphones. Int. J. Intell. Syst. 2022, 37, 4967–4993. [Google Scholar] [CrossRef]
  13. Motonaga, Y.; Kameoka, T.; Hashimoto, A. Constructing Color Image Processing System for Managing the Surface Color of Agricultural Products. J. Jpn. Soc. Agric. Mach. 1997, 59, 13–22. [Google Scholar] [CrossRef]
  14. Cao, Y.; Li, G.L.; Luo, Y.K.; Pan, Q.; Zhang, S.Y. Monitoring of sugar beet growth indicators using wide-dynamic-range vegetation index (WDRVI) derived from UAV multispectral images. Comput. Electron. Agric. 2020, 171, 105331. [Google Scholar] [CrossRef]
  15. Sulemane, S.; Matos-Carvalho, J.P.; Pedro, D.; Moutinho, F.; Correia, S.D. Vineyard Gap Detection by Convolutional Neural Networks Fed by Multi-Spectral Images. Algorithms 2022, 15, 440. [Google Scholar] [CrossRef]
  16. Kanning, M.; Kühling, I.; Trautz, D.; Jarmer, T. High-Resolution UAV-Based Hyperspectral Imagery for LAI and Chlorophyll Estimations from Wheat for Yield Prediction. Remote Sens. 2018, 10, 2000. [Google Scholar] [CrossRef]
  17. Pourdarbani, R.; Sabzi, S.; Dehghankar, M.; Rohban, M.H.; Arribas, J.I. Examination of Lemon Bruising Using Different CNN-Based Classifiers and Local Spectral-Spatial Hyperspectral Imaging. Algorithms 2023, 16, 113. [Google Scholar] [CrossRef]
  18. Tan, L.; Zhou, L.; Zhao, N.; He, Y.; Qiu, Z. Development of a low-cost portable device for pixel-wise leaf SPAD estimation and blade-level SPAD distribution visualization using color sensing. Comput. Electron. Agric. 2021, 190, 106487. [Google Scholar] [CrossRef]
  19. Sudu, B.; Rong, G.; Guga, S.; Li, K.; Zhi, F.; Guo, Y.; Zhang, J.; Bao, Y. Retrieving SPAD Values of Summer Maize Using UAV Hyperspectral Data Based on Multiple Machine Learning Algorithm. Remote Sens. 2022, 14, 5407. [Google Scholar] [CrossRef]
  20. Zhang, J.; Zhang, D.; Cai, D.; Wang, L.; Wang, J.; Sun, L.; Fan, X.; Shen, S.; Zhao, J. Spectral technology and multispectral imaging for estimating the photosynthetic pigments and SPAD of the Chinese cabbage based on machine learning. Comput. Electron. Agric. 2022, 195, 106814. [Google Scholar] [CrossRef]
  21. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  22. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD’16), New York, NY, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef]
  23. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A highly efficient gradient boosting decision tree. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), New York, NY, USA, 4–9 December 2017; pp. 3149–3157. Available online: https://dl.acm.org/doi/10.5555/3294996.3295074 (accessed on 27 December 2023).
  24. Cutler, D.R.; Edwards, T.C., Jr.; Beard, K.H.; Cutler, A.; Hess, K.T.; Gibson, J.; Lawler, J.J. Random Forests for Classification in Ecology. Ecology 2007, 88, 2783–2792. [Google Scholar] [CrossRef] [PubMed]
  25. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2021, 12, 2825–2830. [Google Scholar] [CrossRef]
  26. Nash, J.E.; Sutcliffe, J.V. River flow forecasting through conceptual models part I-A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  27. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), New York, NY, USA, 4–9 December 2017; pp. 4768–4777. Available online: https://dl.acm.org/doi/10.5555/3295222.3295230 (accessed on 27 December 2023).
  28. Mohan, P.J.; Gupta, S.D. Intelligent image analysis for retrieval of leaf chlorophyll content of rice from digital images of smartphone under natural light. Photosynthetica 2019, 57, 388–398. [Google Scholar] [CrossRef]
  29. Guo, Y.; Yin, G.; Sun, H.; Wang, H.; Chen, S.; Senthilnath, J.; Wang, J.; Fu, Y. Scaling Effects on Chlorophyll Content Estimations with RGB Camera Mounted on a UAV Platform Using Machine-Learning Methods. Sensors 2020, 20, 5130. [Google Scholar] [CrossRef] [PubMed]
  30. Yang, H.; Hu, Y.; Zheng, Z.; Qiao, Y.; Hou, B.; Chen, J. A New Approach for Nitrogen Status Monitoring in Potato Plants by Combining RGB Images and SPAD Measurements. Remote Sens. 2022, 14, 4814. [Google Scholar] [CrossRef]
  31. Putra, B.T.W.; Soni, P. Improving nitrogen assessment with an RGB camera across uncertain natural light from above-canopy measurements. Precis. Agric. 2020, 21, 147–159. [Google Scholar] [CrossRef]
  32. Yuan, Y.; Wang, X.; Shi, M.; Wang, P. Performance comparison of RGB and multispectral vegetation indices based on machine learning for estimating Hopea hainanensis SPAD values under different shade conditions. Front. Plant Sci. 2022, 13, 928953. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of the experimental device to capture images of radish leaves.
Figure 1. Schematic diagram of the experimental device to capture images of radish leaves.
Algorithms 17 00016 g001
Figure 2. Model performance of the Random Forests, XGBoost, and LightGBM algorithms for SPAD value prediction on the test datasets: (a) COR, (b) NSE, and (c) RMSE.
Figure 2. Model performance of the Random Forests, XGBoost, and LightGBM algorithms for SPAD value prediction on the test datasets: (a) COR, (b) NSE, and (c) RMSE.
Algorithms 17 00016 g002
Figure 3. Results of SPAD value prediction with Random Forests using color information collected under different lighting conditions: (a) HAL+LED_RGB, (b) HAL+LED_RGB+LCT+ILL, (c) HAL+LED_HSL, (d) HAL+LED_HSL+LCT+ILL, (e) HAL+LED_HSV, (f) HAL+LED_HSV+LCT+ILL, (g) HAL+LED_GB, (h) HAL+LED_GB+LCT+ILL, (i) LED_RGB+ILL, (j) LED_HSL+ILL, (k) LED_GB+ILL, (l) HAL_RGB+ILL, (m) HAL_HSL+ILL, and (n) HAL_GB+ILL.
Figure 3. Results of SPAD value prediction with Random Forests using color information collected under different lighting conditions: (a) HAL+LED_RGB, (b) HAL+LED_RGB+LCT+ILL, (c) HAL+LED_HSL, (d) HAL+LED_HSL+LCT+ILL, (e) HAL+LED_HSV, (f) HAL+LED_HSV+LCT+ILL, (g) HAL+LED_GB, (h) HAL+LED_GB+LCT+ILL, (i) LED_RGB+ILL, (j) LED_HSL+ILL, (k) LED_GB+ILL, (l) HAL_RGB+ILL, (m) HAL_HSL+ILL, and (n) HAL_GB+ILL.
Algorithms 17 00016 g003
Figure 4. Results of SPAD value prediction with Random Forests using limited illuminance data (training data: <300 lux; test data: >300 lux): (a) LED_RES_RGB+ILL, (b) LED_RES_HSL+ILL, and (c) LED_RES_GB+ILL.
Figure 4. Results of SPAD value prediction with Random Forests using limited illuminance data (training data: <300 lux; test data: >300 lux): (a) LED_RES_RGB+ILL, (b) LED_RES_HSL+ILL, and (c) LED_RES_GB+ILL.
Algorithms 17 00016 g004
Figure 5. The top 10 mean absolute SHAP values of the best model in the SPAD value prediction with the Random Forests using halogen and/or LED light source data: (a) HAL+LED_RGB+LCT+ILL, (b) LED_RGB+ILL, and (c) LED_HSL+ILL.
Figure 5. The top 10 mean absolute SHAP values of the best model in the SPAD value prediction with the Random Forests using halogen and/or LED light source data: (a) HAL+LED_RGB+LCT+ILL, (b) LED_RGB+ILL, and (c) LED_HSL+ILL.
Algorithms 17 00016 g005
Figure 6. Partial dependence and individual conditional expectations in the best model for SPAD value prediction with Random Forests using RGB values in halogen and/or LED light sources: (a) HAL+LED_RGB_G_mean, (b) HAL+LED_RGB_G_median, (c) HAL+LED_RGB_G_mu, (d) LED_RGB_B_median, (e) LED_RGB_B_mean, (f) LED_RGB_B_mu, (g) LED_HSL_H_median, (h) LED_HSL_H_mean, and (i) LED_HSL_S_mean.
Figure 6. Partial dependence and individual conditional expectations in the best model for SPAD value prediction with Random Forests using RGB values in halogen and/or LED light sources: (a) HAL+LED_RGB_G_mean, (b) HAL+LED_RGB_G_median, (c) HAL+LED_RGB_G_mu, (d) LED_RGB_B_median, (e) LED_RGB_B_mean, (f) LED_RGB_B_mu, (g) LED_HSL_H_median, (h) LED_HSL_H_mean, and (i) LED_HSL_S_mean.
Algorithms 17 00016 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kamiwaki, Y.; Fukuda, S. Effect of the Light Environment on Image-Based SPAD Value Prediction of Radish Leaves. Algorithms 2024, 17, 16. https://doi.org/10.3390/a17010016

AMA Style

Kamiwaki Y, Fukuda S. Effect of the Light Environment on Image-Based SPAD Value Prediction of Radish Leaves. Algorithms. 2024; 17(1):16. https://doi.org/10.3390/a17010016

Chicago/Turabian Style

Kamiwaki, Yuto, and Shinji Fukuda. 2024. "Effect of the Light Environment on Image-Based SPAD Value Prediction of Radish Leaves" Algorithms 17, no. 1: 16. https://doi.org/10.3390/a17010016

APA Style

Kamiwaki, Y., & Fukuda, S. (2024). Effect of the Light Environment on Image-Based SPAD Value Prediction of Radish Leaves. Algorithms, 17(1), 16. https://doi.org/10.3390/a17010016

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop