Next Article in Journal
Inflorescence Trait Diversity and Genotypic Differentiation as Influenced by the Environment in Elymus nutans Griseb. from Qinghai–Tibet Plateau
Next Article in Special Issue
Synchronous Retrieval of LAI and Cab from UAV Remote Sensing: Development of Optimal Estimation Inversion Framework
Previous Article in Journal
Soil-Water Effects of Good Agricultural and Environmental Conditions Should Be Weighed in Conjunction with Carbon Farming
Previous Article in Special Issue
Identification of Soybean Planting Areas Combining Fused Gaofen-1 Image Data and U-Net Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimation of Fv/Fm in Spring Wheat Using UAV-Based Multispectral and RGB Imagery with Multiple Machine Learning Methods

1
College of Agronomy, Inner Mongolia Agricultural University, Huhhot 010019, China
2
Bayannaoer Academy of Agricultural and Animal Sciences, Bayannaoer 015000, China
3
School of Biological Science and Technology, Baotou Teachers’ College, Baotou 014031, China
4
Chifeng Forest and Grassland Protection and Development Center, Chifeng 024005, China
*
Authors to whom correspondence should be addressed.
Agronomy 2023, 13(4), 1003; https://doi.org/10.3390/agronomy13041003
Submission received: 20 February 2023 / Revised: 25 March 2023 / Accepted: 28 March 2023 / Published: 29 March 2023

Abstract

:
The maximum quantum efficiency of photosystem II (Fv/Fm) is a widely used indicator of photosynthetic health in plants. Remote sensing of Fv/Fm using MS (multispectral) and RGB imagery has the potential to enable high-throughput screening of plant health in agricultural and ecological applications. This study aimed to estimate Fv/Fm in spring wheat at an experimental base in Hanghou County, Inner Mongolia, from 2020 to 2021. RGB and MS images were obtained at the wheat flowering stage using a Da-Jiang Phantom 4 multispectral drone. A total of 51 vegetation indices were constructed, and the measured Fv/Fm of wheat on the ground was obtained simultaneously using a Handy PEA plant efficiency analyzer. The performance of 26 machine learning algorithms for estimating Fv/Fm using RGB and multispectral imagery was compared. The findings revealed that a majority of the multispectral vegetation indices and approximately half of the RGB vegetation indices demonstrated a strong correlation with Fv/Fm, as evidenced by an absolute correlation coefficient greater than 0.75. The Gradient Boosting Regressor (GBR) was the optimal estimation model for RGB, with the important features being RGBVI and ExR. The Huber model was the optimal estimation model for MS, with the important feature being MSAVI2. The Automatic Relevance Determination (ARD) was the optimal estimation model for the combination (RGB + MS), with the important features being SIPI, ExR, and VEG. The highest accuracy was achieved using the ARD model for estimating Fv/Fm with RGB + MS vegetation indices on the test sets (Test set MAE = 0.019, MSE = 0.001, RMSE = 0.024, R2 = 0.925, RMSLE = 0.014, MAPE = 0.026). The combined analysis suggests that extracting vegetation indices (SIPI, ExR, and VEG) from RGB and MS remote images by UAV as input variables of the model and using the ARD model can significantly improve the accuracy of Fv/Fm estimation at flowering stage. This approach provides new technical support for rapid and accurate monitoring of Fv/Fm in spring wheat in the Hetao Irrigation District.

1. Introduction

Spring wheat is an important crop in northern China, with a wide planting range and high yields. The Hetao Irrigation District is one of the major production areas for spring wheat in China, and it plays a significant role in ensuring food security [1]. Chlorophyll fluorescence has proven to be a useful indicator of photosynthetic system health and is widely employed in assessing photosynthesis [2]. Fv/Fm, a commonly used chlorophyll fluorescence parameter, represents the ratio of variable fluorescence (Fv) to maximum fluorescence (Fm) of chlorophyll. Fv represents the fluorescence emitted by open PSII reaction centers, whereas Fm represents the maximum fluorescence emitted by fully open PSII reaction centers under saturating light conditions. Fv/Fm reflects the efficiency of energy transfer within the PSII antenna and the proportion of open PSII reaction centers [3,4]. The Fv/Fm parameter is critical in understanding the physiological status of plants and is a measure of a plant’s capacity to convert light energy into chemical energy, which can offer insights into the health and productivity of crops [5]. In the case of spring wheat, Fv/Fm can be utilized to monitor crop growth and identify potential stress factors, such as water and nutrient deficiencies, that may adversely impact crop yield [6]. Traditionally, Fv/Fm has been estimated using portable instruments such as pulse amplitude modulated (PAM) fluorometers, which necessitate darkening the leaf with a leaf clamp for 15–20 min before measurement. PAM measurements can be laborious and time-consuming, particularly when large areas require monitoring. To overcome these limitations, researchers have explored the use of remote sensing technology.
Remote sensing is a powerful tool for monitoring plant growth. Remote sensing approaches retrieve chlorophyll fluorescence, which is excited by the absorption of sunlight, using spectral reflectance. Since the fluorescence emission spectrum is superimposed on leaf or canopy reflectance that can be obtained by handheld, ground-mounted, aerial, or space-borne sensors, remote sensing technique opens a new way for upscaling chlorophyll fluorescence from leaf to landscape levels. Fv/Fm has been estimated in many studies. Zhao et al. [7] collected spectral data and Fv/Fm values from potato leaves using a hyperspectral imaging system and a closed chlorophyll fluorescence imaging system, decomposed the spectral data by continuous wavelet transform (CWT), and developed an estimation model using partial least squares. Yi et al. [8] used hyperspectral and PAM fluorescence data along with correlation and regression analyses to develop Fv/Fm estimation models for aspen and cherry leaves. Jia et al. [9] calculated vegetation index using hyperspectral data to estimate Fv/Fm for wheat through linear regression. With these Fv/Fm estimation studies, the time-consuming issue of traditional fluorescence determination methods were addressed, but ground-based hyperspectral data collection was not only expensive but also incapable of estimating Fv/Fm at high spatial and temporal resolution. Unmanned aerial vehicle (UAVs) equipped with RGB sensors and multispectral sensors may solve the problem at low cost [10]. Most research on estimating Fv/Fm has primarily relied on ground-based hyperspectral measurements, with few studies employing unmanned aerial vehicles equipped with multispectral and visible sensors. This study fills a critical gap in the literature by demonstrating the feasibility of using UAV-based remote sensing to estimate Fv/Fm, offering new opportunities for efficient and high-throughput monitoring of plant health and productivity.
Machine learning techniques have revolutionized the field of data analysis by identifying complex patterns and trends that are often challenging to detect using traditional methods. In recent years, the application of machine learning methods to analyze data acquired by UAVs has gained significant traction [11,12,13,14,15]. However, the majority of previous studies have focused on using a limited number of machine learning methods (one to four) to estimate the desired parameters, with few investigations comparing the performance of more than twenty different techniques. Given the subtle variation of Fv/Fm and the limited spectral resolution of multispectral and RGB sensors compared to hyperspectral, it is crucial to employ multiple machine learning methods to achieve higher accuracy. This approach allows for the exploitation of the complementarity of various algorithms and enables robust and comprehensive estimation of Fv/Fm from remote sensing data. In this study, the goal is to estimate Fv/Fm in spring wheat using UAV-acquired RGB and MS remote sensing data by multiple machine learning methods to improve accuracy, which is important for rapid and accurate detection of wheat stress and timely adjustment of field management measures.

2. Materials and Methods

2.1. Study Site and Experimental Design

During the wheat flowering stage, both RGB and MS remote sensing images were obtained, resulting in the calculation of 51 vegetation indices (comprising 25 RGB and 26 multispectral). Following this, critical spectral features were extracted, while multicollinearity was eliminated, and feature selection was conducted to estimate the Fv/Fm values. An array of 26 machine learning techniques were utilized, with their respective performances assessed based on accuracy, stability, and interpretability. In conclusion, a high-precision UAV remote sensing monitoring model for the Fv/Fm of spring wheat in the Hetao Irrigation District was developed, thus providing a robust scientific foundation and theoretical underpinning for local agricultural advancement.
The study was conducted from 2020 to 2021 at the experimental field of the Bayannur Institute of Agriculture and Animal Husbandry Science, located in the Inner Mongolia Autonomous Region at 40°04′ N, 10°03′ E and an altitude of 1038 m above sea level (Figure 1). The soil type at the experimental site was loam, and the baseline fertility information is presented in Table 1. A split-plot design was utilized, with nitrogen (N) fertilizer application methods serving as the main plot and cultivar as the subplot. The main plot, which included five levels (CK, N1, N2, N3, and N4), featured various N application methods. N1 (0.8/0.2), N2 (0.7/0.3), N3 (0.5/0.5), and N4 (0.3/0.7) had the same N application rate of 180 kg/ha but differed in seeding fertilizer rates and follow-up fertilizer rates, while CK had no fertilizer applied. The subplot comprised three cultivars of spring wheat: “Baimai 13”, “Nongmai 730”, and “Nongmai 482”. The experiment included a total of 15 treatments with three replications, resulting in 45 experimental plots, each measuring 12 m2. The plots were arranged in randomized groups. The sowing rate was set at 300 kg/ha. Phosphorus fertilizer was applied as a basal fertilizer during sowing, and no potassium fertilizer was applied during the entire reproductive phase. Three flood irrigations were performed at the tillering, heading, and grain filling stages, each with a volume of 900 m3/ha.

2.2. UAV Multispectral Data Acquisition and Processing

Remote images were obtained during the flowering stage of the wheat plant (12 June 2020; 15 June 2021) using a DJI Phantom 4 multispectral drone (Da-Jiang Innovations, Shenzhen, China). The drone (P4M, Figure 2) features 6 CMOS, including 1 color sensor (ISO: 200–800) for visible (RGB) imaging and 5 monochrome sensors (Table 2) for multi-spectral (MS) imaging. The images were acquired on clear and windless days, with a fixed and consistent takeoff location. The D-RTK 2 (Da-Jiang Innovations, Shenzhen, China) high-precision GNSS mobile station was utilized to assist in the positioning of the UAV and enhance its positioning accuracy. Prior to takeoff, the UAV was manually placed directly above three reflectivity gray plates of 20%, 40%, and 60%, and reflectivity plate photos were taken. The flight path was automatically planned by DJI GS Pro (Da-Jiang Innovations, Shenzhen, China) after calculating the current solar azimuth, with a flight altitude of 30 m, the ground sampling distance was 1.59 cm/pixel, a heading overlap of 85%, and a collateral overlap of 80%. Following the flight, DJI Terra (Da-Jiang Innovations, Shenzhen, China) was used to perform radiometric correction for multispectral images, followed by image stitching to generate a single-band reflectivity orthophoto. RGB images were stitched to produce color ortho images without a radiation correction step.

2.3. Construction and Selection of Spectral Indices

The digital number values for each RGB band and the reflectance of each MS band in each treatment plot were extracted using the zonal statistics function of ENVI. Subsequently, two types of vegetation indices (VIs) were computed, as presented in Table 3 and Table 4.

2.4. Fluorescence Data Acquisition and Processing

The collection of fluorescence data was synchronized with the UAV flight, and the fluorescence information of wheat canopy leaves was obtained using the Handy PEA plant efficiency analyzer (Hansatech Instruments Ltd., Norfolk, UK). In each plot, 20 leaves were randomly collected, and the average value was adopted as a representative value of the plant. Prior to collection, the target leaves were subjected to a dark treatment for 20 min using the leaf clips that were provided with the instrument.

2.5. Construction of Regression Model

In this study, 26 machine learning regression models (listed in Table 5) were developed using PyCaret to estimate Fv/Fm. PyCaret is a user-friendly, open source, low-code machine learning library in Python that enables users to easily prepare data, train and evaluate machine learning models, and deploy models to production. PyCaret offers various features for data preparation, feature engineering, model training and evaluation, model interpretation, and model deployment. Additionally, it includes built-in visualizations and interactive plots that help users to interpret model results.
To prepare the data, a normalization technique was applied to transform the data into a fixed range between 0 and 1, thereby ensuring that all features were on the same scale. Multicollinearity, which refers to high correlation between multiple features, was addressed by removing highly correlated features to ensure data stability.
Feature selection was performed to select key features and reduce noise, thereby enhancing the accuracy and efficiency of the algorithm. Once all the models were built, the model with the highest accuracy was selected based on the accuracy ranking, and hyperparameter optimization was conducted to further improve model accuracy. This study used the feature selection scheme of the embedding method, implemented by calling the SelectFromModel API in sklearn, relying on the algorithmic model of random forests.

2.6. Segmentation of Dataset and Accuracy Evaluation

The 90 samples were randomly partitioned into a training set and a test set in a 0.7/0.3 ratio; a K-fold cross-validation (K = 5) was employed to train and optimize the model. Seven indicators were utilized to assess the accuracy of the model in the test set:
MAE (Mean Absolute Error) is a measure of the average magnitude of the errors in a set of predictions, without considering their direction. It measures the average absolute difference between the actual and predicted values. A smaller MAE indicates a more accurate prediction;
MSE (Mean Squared Error) is a measure of the average magnitude of the errors in a set of predictions, considering both the magnitude and direction of the errors. It measures the average of the squared differences between the actual and predicted values. A smaller MSE indicates a more accurate prediction;
R2 (Coefficient of Determination) is a statistical measure that represents the proportion of the variance in the dependent variable that is predictable from the independent variables. In regression analysis, R2 is used to evaluate the goodness of fit of the model. Generally, it ranges from 0 to 1, with a higher value indicating a better fit. An R2 of 1 indicates that the model perfectly predicts the target variable, while an R2 of 0 indicates that the model does not explain any variance in the target variable;
RMSE (Root Mean Squared Error) is the square root of MSE. It is a measure of the average magnitude of the errors in a set of predictions, considering both the magnitude and direction of the errors. A smaller RMSE indicates a more accurate prediction;
RMSLE (Root Mean Squared Logarithmic Error) is similar to RMSE, but instead of taking the difference between the actual and predicted values, it takes the logarithmic difference. It is used in cases where the target variable has a skewed distribution;
MAPE (Mean Absolute Percentage Error) is a measure of the average error as a percentage of the actual values. It measures the average percentage difference between the actual and predicted values. A lower MAPE indicates a more accurate prediction;
TT (Total Time) is defined as the amount of time spent on building the machine learning model. The value of TT represents the computational cost of constructing the model, including the time spent on training and validating the model. A smaller value of TT indicates that the model has a lower computational cost and can be built more efficiently. This can be beneficial in scenarios where the model needs to be constructed in a timely manner, or when computational resources are limited. A lower TT value also indicates that the model may be more scalable and can be trained on larger datasets without excessive computational cost.
M A E = 1 n i = 1 n   y i y ^
M S E = i = 1 n   y i y ^ i 2 n
R M S E = i = 1 n   y i y ^ i 2 n
R 2 = 1 i = 1 n   y i y ^ i 2 i = 1 n   y i y i 2
RMSLE   = 1 n i = 1   ( l o g ( y i + 1 ) l o g ( y ^ i + 1 ) ) 2
M A P E = 100 % n i = 1 n   y ^ i y i y i
where n is the number of samples, y i is the observed value, y i is the mean of the observed value, and y ^ i is the predicted value.

3. Results

3.1. Basic Statistical Information of the Fv/Fm Dataset

As shown in Table 6, the basic statistics of the measured Fv/Fm, indicating a range of values with a minimum of 0.550, a maximum of 0.848, a mean of 0.773, a standard deviation of 0.081, and a coefficient of variation (CV%) of 10.4. These statistics suggest that the Fv/Fm measurements demonstrate moderate variability within the total dataset. Both the training and test sets were derived from the total dataset and displayed similar ranges of Fv/Fm values. Specifically, the training set showed a minimum of 0.551, a maximum of 0.846, a mean of 0.775, a standard deviation of 0.077, and a CV% of 9.900, while the test set demonstrated a minimum of 0.550, a maximum of 0.848, a mean of 0.768, a standard deviation of 0.090, and a CV% of 11.700. The distributions of Fv/Fm values in the training and test sets were comparable, with a slightly higher mean and lower CV for the training set than the test set. These findings indicate that the training and test sets are representative of the total dataset and can be utilized for model development and validation.

3.2. Correlation Analysis of Fv/Fm with Multispectral and RGB Vegetation Indices

The correlation coefficients between Fv/Fm values and MS, RGB vegetation indices are shown in Table 7. The highest correlation coefficients of multispectral vegetation indices were NDVI, followed by SIPI, EVI, and MSAVI2. The highest correlation coefficients of RGB vegetation indices were RGBVI, followed by VDVI, GLI, and CIVE. Most of the multispectral vegetation indices and half of RGB vegetation indices showed good correlation with Fv/Fm. Overall, the correlation between the multispectral vegetation indices and Fv/Fm was higher than that of the RGB vegetation indices.

3.3. Important Features Selected after Data Pre-Processing

The process of feature selection was conducted to identify the most crucial variables that significantly influence the estimation of the chlorophyll fluorescence parameter Fv/Fm in spring wheat using UAV remote sensing. Table 8 illustrates the significant features that were selected after the application of data pre-processing techniques, including the visible light vegetation indices (RGB), multispectral vegetation indices (MS), and a combination of both (RGB + MS). The pre-processing methodology involved eliminating feature collinearity and implementing tree-based feature selection methods. The table highlights that RGBVI and ExR were the vital features for the RGB dataset, MSAVI2 was the only critical feature for the MS dataset, and SIPI, ExR, and VEG were the essential features for the RGB + MS dataset.

3.4. Model Based on RGB VIs Development and Evaluation

Table 9 presented the results of the performance evaluation of 26 machine learning models for estimating vegetation indices using RGB vegetation indices. The models were ranked according to their R2 accuracy scores, with the top performing model listed first. Evaluation metrics such as mean absolute error (MAE), mean squared error (MSE), root mean squared error (RMSE), relative mean squared logarithmic error (RMLSE), mean absolute percentage error (MAPE), and computation time (TT) were used to assess model performance. The gradient boosting regression (GBR) model achieved the highest accuracy with an R2 score of 0.800, followed closely by the random forest (RF) model with an R2 score of 0.795. Most of the other models performed relatively poorly, with R2 scores ranging from 0.789 to −116.050. Notably, the worst performing models, including lasso, elastic net, least angle regression (LLAR), dummy, support vector machine (SVM), and kernel ridge regression (KR), had negative R2 scores. In addition to the R2 scores, the evaluation metrics indicated that the best performing models also had the lowest MAE, MSE, RMSE, RMLSE, and MAPE scores, demonstrating the models’ ability to make accurate predictions with low error rates. However, there was considerable variation in computation time among the models, with some taking significantly longer than others. In conclusion, the results suggest that GBR models are the most accurate and efficient for estimating Fv/Fm using RGB vegetation indices.

3.5. Model Based on MS VIs Development and Evaluation

Table 10 presented an accuracy assessment of 26 machine learning models for estimating vegetation indices using multi-spectral data. The top seven models with the highest R2 scores, ranging from 0.860 to 0.849, were Huber, LR, Ridge, LAR, OMP, BR, ARD, and TR. These models had the lowest MAE, MSE, RMSE, and RMLSE values among all the models, indicating that they produced the most accurate estimates of vegetation indices. The computation time for these models ranged from 0.004 to 0.316 s, with the LR model having the longest computation time. The next group of models, with R2 scores ranging from 0.794 to 0.684, included KNN, RF, CatBoost, ADA, GBR, ET, and PAR. These models had higher MAE, MSE, RMSE, and RMLSE values than the top models, indicating that they produced less accurate estimates of vegetation indices. The computation time for these models ranged from 0.006 to 0.190 s. The last group of models, with R2 scores ranging from 0.668 to −0.567, included XGB, DT, RANSC, LGBM, SVM, Lasso, EN, LLAR, Dummy, MLP, and KR. These models had the lowest R2 scores and the highest MAE, MSE, RMSE, and RMLSE values, indicating that they produced the least accurate estimates of Fv/Fm. The computation time for these models ranged from 0.004 to 0.298 s, with the KR model having the longest computation time. Overall, the Huber model was the most accurate for estimating Fv/Fm using multispectral vegetation indices.

3.6. Model Based on RGB and MS VIs Development and Evaluation

Table 11 presented an accuracy assessment of 26 machine learning models for the estimation of vegetation indices (VIs) using both RGB and multispectral data. The ARD and OMP models achieved the highest R2 accuracy scores of 0.868, followed closely by the Ridge, LR, LAR, BR, and Huber models, all with R2 values of 0.858. The Tr model obtained an R2 of 0.849, while RANSC, KNN, RF, ET, ADA, Catboost, and XGB models had R2 values ranging from 0.830 to 0.723. The GBR and DT models had R2 values of 0.721 and 0.690, respectively. The PAR model had an R2 of 0.593, indicating lower accuracy than the previous models. On the other hand, the LGMB, Lasso, EN, LLAR, and Dummy models had negative R2 values, indicating poor accuracy. Moreover, the SVM and MLP models had low R2 values of −0.073 and −0.292, respectively. The KR model had the worst performance with high MAE, MSE, RMSE, and RMLSE values and a very low R2 value of −118.159. In summary, the ARD and OMP models demonstrated the same highest accuracy for estimating Fv/Fm using both MS and RGB data. However, the computation time of OMP was found to be higher than that of ARD. Therefore, the optimal model is ARD.

3.7. Hyperparameter Optimization of the Highest Accuracy Model in Three Modes

Table 12 and Figure 3 show the hyperparameter-optimized performance of the most accurate machine learning models in three different image types (RGB, MS and MS + RGB). The results show that the performance of these models varies according to the different image types and data sets. Notably, the ARD model achieves the highest accuracy on the test set of MS + RGB VIs mode (test set MAE = 0.019, MSE = 0.001, RMSE = 0.024, R2 = 0.925, RMSLE = 0.014 and MAPE = 0.026).

4. Discussion

Chlorophyll fluorescence is a potent indicator of photosynthesis, where Fv/Fm serves as a measure of the maximum photochemical efficiency of photosystem II in chloroplasts and can be utilized as an indicator of plant health [75]. For many plant species, the optimal Fv/Fm range is approximately between 0.79 and 0.84, with lower values indicating higher plant stress [76]. The wheat flowering period is a critical stage of growth and development, as it signifies the transition of wheat plants from vegetative to reproductive growth stages, where the focus of growth shifts from leaves and roots to flowers and seeds [77]. During this period, wheat plants require significant amounts of nutrients and energy to support the normal development and maturation of flowers and seeds [78]. Therefore, providing sufficient water and nutrients is crucial for wheat plants during the flowering period. Stress during this phase can significantly affect the yield and quality of wheat [79]. Monitoring Fv/Fm during the flowering period is crucial in detecting stress and implementing timely agricultural management practices to optimize wheat production and achieve high yields.
In this study, both multispectral vegetation indices and RGB vegetation indices were found to be effective in estimating Fv/Fm values. The R2 of test set was 0.920 for multispectral vegetation indices and 0.838 for RGB vegetation indices, indicating that the former had a higher accuracy in estimating Fv/Fm values. Multispectral images acquired by UAVs can provide information about the spectral reflectance of vegetation, which changes simultaneously in the canopy when stress occurs [80], and this can be used to estimate Fv/Fm. The key idea of this approach is that Fv/Fm is related to the fluorescence yield of photosystem II, which in turn is related to the chlorophyll content of leaves. The chlorophyll content can be estimated from the reflectance in the red and near-infrared (NIR) bands [81]. The important spectral index MSAVI2 extracted from the multispectral estimation in this study was calculated precisely from the red and NIR bands. In addition to multispectral images, RGB images can also be used to estimate Fv/Fm values by color information. In general, the color of plants changes when they are under stress, and although some color changes are difficult for the human eye to observe, color values can be quantified by computer technology [82]. However, due to the lack of vegetation-sensitive red edge and infrared bands, the RGB images do not provide enough spectral information, so the estimated Fv/Fm is not as accurate as the MS images.
This study shows that the machine learning models have high accuracy and stability and can effectively use RGB and MS data to estimate Fv/Fm, which is consistent with the findings of other studies [83] using machine learning for remote sensing estimation. The effects of different machine learning models and different types of vegetation indices on the accuracy of Fv/Fm estimation are obvious. Firstly, different types of vegetation indices can provide different information in estimating Fv/Fm, and different machine learning models have different adaptation and fitting ability to different datasets [84]. The MS + RGB model exhibits superior accuracy compared to both the MS and RGB models in estimating Fv/Fm. This suggests that the integration of RGB and MS data provides benefits in enhancing the accuracy of the estimation. Additionally, the findings suggest that utilizing multiple data sources can enhance the accuracy of the model as compared to relying solely on single-source data. These results are consistent with previous studies that have reported improved accuracy in multi-source data estimation [85,86]. However, the improvement in accuracy when combining (MS + RGB) is only marginally better than using MS alone in this study. The ARD model cannot be quantified for the percentage contribution of RGB and MS, so the random-forest-model-based feature importance assessment was implemented, the result was shown in Figure 4. The importance of multispectral vegetation index in the model was higher than that of RGB vegetation indices, and the RGB contributed less valid information.
Zhao et al. [7] developed two estimation models for Fv/Fm of potato leaves with test set R2 of 0.807 and 0.822 and RMSE of 0.018 and 0.017, respectively. Yi et al. [8] developed a generalized estimation model of Fv/Fm for poplar leaves and cherry leaves with R2 = 0.88. In a study [9] estimating Fv/Fm in winter wheat, training set R2 = 0.50, RMSE = 0.012, test set R2 = 0.55, RMSE = 0.014. In these previous estimation studies of Fv/Fm based on hyperspectral, the highest R2 was of 0.88, and the highest accuracy combined model without hyperparameter optimization in this study had R2 of 0.868, which shows that the estimation accuracy of multispectral and RGB is not as good as hyperspectral. However, Yang et al. [87] showed that hyperparameter optimization can be helpful in improving the estimation accuracy of the model. In this study, the test set R2 of both the MS estimation and the estimation of the MS and RGB combination reaches 0.92 after hyperparameter optimization, indicating that the accuracy gap caused by the sensors can be narrowed or even surpassed by algorithm optimization. Negative values of R2 were observed in all three estimation models, which are usually uncommon, indicating that the prediction results using the model are worse than the estimation results using the mean, because the mean reflects the central tendency of the data, while the prediction results of the model deviate more from the true value than the mean [88]. Notably, the performance of linear regression (LR), ridge regression (Ridge), least angle regression (LAR), and orthogonal matching pursuit (OMP) models was almost identical. Moreover, the reason for this was that only MSAVI2 was screened in the multispectral vegetation indices and LR, Ridge, LAR, and OMP are all linear models, indicating that different linear models may not have a significant effect on the accuracy of estimation when a single vegetation index was used as a characteristic variable.
The present study has important practical implications for agriculture and environmental monitoring. By demonstrating the feasibility of using UAV-based remote sensing to estimate Fv/Fm in spring wheat, our research offers new opportunities for efficient and high-throughput monitoring of plant health and productivity. This can help to improve crop management strategies, enhance yield and quality of agricultural products, and reduce the environmental impact of farming practices. In addition, the integration of machine learning methods with multispectral and RGB imagery can further enhance the accuracy and reliability of Fv/Fm estimation, enabling more precise and targeted interventions in crop management.

5. Conclusions

Integration of multiple machine learning methods with multispectral and RGB imagery acquired from UAV-based remote sensing can improve the accuracy and reliability of Fv/Fm estimation in spring wheat during the flowering stage. The important features and the optimal Fv/Fm estimation models for different types remote sensing images were different: with gradient boosting regressor (GBR) as the optimal estimation model for RGB, the important features were RGBVI and ExR; with Huber as the optimal estimation model for MS, the important feature was MSAVI2; and automatic relevance determination (ARD) as the optimal estimation model for combination (MS + RGB), the important features were SIPI, ExR, VEG. The highest accuracy was achieved using the ARD model for estimating Fv/Fm with RGB + MS vegetation indices (Test set MAE = 0.019, MSE = 0.001, RMSE = 0.024, R2 = 0.925, RMSLE = 0.014, MAPE = 0.026). Based on the results of this study, there is great potential for the use of remote sensing and machine learning for efficient and sustainable plant health monitoring and management.
In conclusion, while the present study provides a valuable contribution to the use of remote sensing and machine learning for estimating Fv/Fm in wheat during the flowering stage, there are several limitations. Firstly, the study was conducted at a single ecological site over a two-year period, which may limit the generalizability of the findings to other regions and ecosystems. Future studies conducted at multiple sites with varying environmental conditions would provide a more comprehensive understanding of the applicability of the proposed methods. Secondly, the study focused solely on the flowering stage of wheat growth, which is only one phase of the crop’s development. It is possible that the performance of the proposed remote sensing and machine learning methods may differ at other stages of wheat growth, such as germination, tillering, or grain filling. Further research is needed to examine the feasibility and accuracy of the methods across multiple growth stages. Thirdly, while the proposed methods show promise for estimating Fv/Fm in wheat, it should be noted that these methods may not be directly applicable to other plant species or crops. The optimal settings and parameters for remote sensing and machine learning may vary depending on the physiological and structural characteristics of the target plants. Future studies should explore the generalizability and adaptability of the methods to other crops and plant species.

Author Contributions

Conceptualization, Q.W. and Y.Z.; methodology, Q.W.; software, M.X.; validation, Z.Z. and J.L.; formal analysis, Z.Z.; investigation, D.H.; resources, L.Y.; data curation, D.H. and J.L; writing—original draft preparation, Q.W.; writing—review and editing, Y.Z.; visualization, M.X.; supervision, Y.Z.; project administration, Y.Z.; funding acquisition, Y.Z. and M.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Inner Mongolia “science and technology” action focus on special “Yellow River Basin durum wheat industrialization capacity enhancement” (NMKJXM202201-4); Inner Mongolia “science and technology” action focus on special “Research and Application of Key Technologies for Production and Processing of Durum Wheat and Products in Hetao irrigation area” (NMKJXM202111-3) and Inner Mongolia Natural Science Foundation of China “Research on nitrogen nutrition diagnosis of spring wheat in Hetao irrigation area based on UAV mapping technology” (2021MS03089).

Data Availability Statement

Not applicable.

Acknowledgments

We sincerely appreciate the assistance of the Wheat Research Institute of the Inner Mongolia Bayannur Academy of Agriculture and Animal Science.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, Z.; Li, M.; Wu, Q.; Zhang, Y. Effects of Different Soil Moisture-Holding Strategies on Growth Characteristics, Yield and Quality of Winter-Seeded Spring Wheat. Agronomy 2022, 12, 2746. [Google Scholar] [CrossRef]
  2. Baker, N.R. Chlorophyll Fluorescence: A Probe of Photosynthesis In Vivo, Annu. Rev. Plant Biol. 2008, 59, 89–113. [Google Scholar] [CrossRef] [Green Version]
  3. Maxwell, K.; Johnson, G.N. Chlorophyll fluorescence–A practical guide. J. Exp. Bot. 2000, 51, 659–668. [Google Scholar] [CrossRef]
  4. Baker, N.R.; Rosenquist, E. Applications of chlorophyll fluorescence can improve crop production strategies: An examination of future possibilities. J. Exp. Bot. 2004, 55, 1607–1621. [Google Scholar] [CrossRef] [Green Version]
  5. Genty, B.; Briantais, J.M.; Baker, N.R. The relationship between the quantum yield of photosynthetic electron transport and quenching of chlorophyll fluorescence. Biochim. Biophys. Acta (BBA)-Gen. Subj. 1989, 990, 87–92. [Google Scholar] [CrossRef]
  6. Sharma, D.K.; Andersen, S.B.; Ottosen, C.O.; Rosenqvist, E. Wheat cultivars selected for high Fv/Fm under heat stress maintain high photosynthesis, total chlorophyll, stomatal conductance, transpiration and dry matter. Physiol. Plant. 2015, 153, 284–298. [Google Scholar] [CrossRef]
  7. Zhao, R.; An, L.; Song, D.; Li, M.; Qiao, L.; Liu, N.; Sun, H. Detection of chlorophyll fluorescence parameters of potato leaves based on continuous wavelet transform and spectral analysis. Spectrochim. Acta A 2021, 259, 119768. [Google Scholar] [CrossRef]
  8. Yi, P.; Aoli, Z.; Tinge, Z.; Sheng, H.F.; Yan, G.; Yan, Q.T.; Ying, Z.; Kan, L. Using remotely sensed spectral reflectance to indicate leaf photosynthetic efficiency derived from active fluorescence measurements. J. Appl. Rem. Sens. 2017, 11, 026034. [Google Scholar] [CrossRef] [Green Version]
  9. Jia, M.; Li, D.; Colombo, R.; Wanlg, Y.; Wang, X.; Cheng, T.; Zhu, Y.; Yao, X.; Xu, C.; Ouer, G.; et al. Quantifying Chlorophyll Fluorescence Parameters from Hyperspectral Reflectance at the Leaf Scale under Various Nitrogen Treatment Regimes in Winter Wheat. Remote Sens. 2019, 11, 2838. [Google Scholar] [CrossRef] [Green Version]
  10. Liu, J.; Zhu, Y.; Tao, X.; Chen, X.; Li, X. Rapid prediction of winter wheat yield and nitrogen use efficiency using consumer-grade unmanned aerial vehicles multispectral imagery. Front. Plant Sci. 2022, 13, 1032170. [Google Scholar] [CrossRef] [PubMed]
  11. Torres-Tello, J.W.; Ko, S. A novel approach to identify the spectral bands that predict moisture content in canola and wheat. Biosyst. Eng. 2021, 210, 91–103. [Google Scholar] [CrossRef]
  12. Wu, Q.; Zhang, Y.; Zhao, Z.; Xie, M.; Hou, D. Estimation of Relative Chlorophyll Content in Spring Wheat Based on Multi-Temporal UAV Remote Sensing. Agronomy 2023, 13, 211. [Google Scholar] [CrossRef]
  13. Wilke, N.; Siegmann, B.; Postma, J.A.; Muller, O.; Krieger, V.; Pude, R.; Rascher, U. Assessment of plant density for barley and wheat using UAV multispectral imagery for high-throughput field phenotyping. Comput. Electron. Agric. 2021, 189, 106380. [Google Scholar] [CrossRef]
  14. Du, X.; Wan, L.; Cen, H.; Chen, S.; Zhu, J.; Wang, H.; He, Y. Multi-temporal monitoring of leaf area index of rice under different nitrogen treatments using UAV images. Int. J. Precis. Agric. Aviat. 2020, 1, 11–18. [Google Scholar] [CrossRef]
  15. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Daloye, A.M.; Erkbol, H.; Fritschi, F.B. Crop monitoring using satellite/UAV data fusion and machine learning. Remote Sens. 2020, 12, 1357. [Google Scholar] [CrossRef]
  16. Meyer, G.E.; Neto, J.C. Verification of color vegetation indices for automated crop imaging applications. Comput. Electron. Agric. 2008, 63, 282–293. [Google Scholar] [CrossRef]
  17. Woebbecke, D.M.; Meyer, G.E.; Von Bargen, K.; Mortensen, D.A. Color indices for weed identification under various soil, residue, and lighting conditions. Trans. ASAE 1995, 38, 259–269. [Google Scholar] [CrossRef]
  18. Vincent, L. Morphological grayscale reconstruction in image analysis: Applications and efficient algorithms. IEEE Trans. Image Process 1993, 2, 176–201. [Google Scholar] [CrossRef] [Green Version]
  19. Sellers, P.J.; Hall, F.G.; Nielsen, G.A. Vegetation/atmosphere transfer models. Adv. Space Res. 1987, 7, 149–159. [Google Scholar] [CrossRef]
  20. Hague, T.; Tillett, N.D.; Wheeler, H. Automated crop and weed monitoring in widely spaced cereals. Precis. Agric. 2006, 7, 21–32. [Google Scholar] [CrossRef]
  21. Guijarro, M.; Pajares, G.; Riomoros, I.; Herrera, P.J.; Burgos-Artizzu, X.P.; Ribeiro, A. Automatic segmentation of relevant textures in agricultural images. Comput. Electron. Agric. 2011, 75, 75–83. [Google Scholar] [CrossRef] [Green Version]
  22. Daughtry, C.S.T.; Walthall, C.L.; Kim, M.S.; De Colstoun, E.B.; McMurtrey, J.E. Estimating corn leaf chlorophyll concentration from leaf and canopy reflectance. Remote Sens. Environ. 2000, 74, 229–239. [Google Scholar] [CrossRef]
  23. Tucker, C.J. Red and photographic infrared linear combinations for monitoring vegetation. Remote Sens. Environ. 1979, 8, 127–150. [Google Scholar] [CrossRef] [Green Version]
  24. Kawashima, S. A new vegetation index for the monitoring of vegetation phenology and thermal stress. Int. J. Remote Sens. 2002, 23, 2003–2017. [Google Scholar] [CrossRef]
  25. Gitelson, A.A.; Kaufman, Y.J.; Stark, R.; Rundquist, D. Novel algorithms for remote estimation of vegetation fraction. Remote Sens. Environ. 2002, 80, 76–87. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, X.; Wang, M.; Wang, S.; Wu, Y. Extraction of vegetation information from visible unmanned aerial vehicle images. Trans. Chin. Soc. Agric. Eng. 2015, 31, 152–159. [Google Scholar] [CrossRef]
  27. Saberioon, M.M.; Amin, M.S.M.; Anuar, A.R.; Gholizadeh, A. Assessment of rice leaf chlorophyll content using visible bands at different growth stages at both the leaf and canopy scale. Int. J. Appl. Earth Obs. Geoinf. 2014, 32, 35–45. [Google Scholar] [CrossRef]
  28. Bendig, J.; Yu, K.; Aasen, H.; Bolten, A.; Bennertz, S.; Broscheit, J.; Gnyp, M.L.; Bareth, G. Combining UAV-based plant height from crop surface models, visible, and near infrared vegetation indices for biomass monitoring in barley. Int. J. Appl. Earth Obs. Geoinf. 2015, 39, 79–87. [Google Scholar] [CrossRef]
  29. Huete, A.; Justice, C.; Liu, H. Development of vegetation and soil indices for MODIS-EOS. Remote Sens. Environ. 1994, 49, 224–234. [Google Scholar] [CrossRef]
  30. Gitelson, A.A.; Merzlyak, M.N. Remote sensing of chlorophyll concentration in higher plant leaves. Adv. Space Res. 1998, 22, 689–692. [Google Scholar] [CrossRef]
  31. Gamon, J.A.; Surfus, J.S. Assessing leaf pigment content and activity with a reflectometer. New Phytol. 1999, 143, 105–117. [Google Scholar] [CrossRef]
  32. Birth, G.S.; McVey, G.R. Measuring the color of growing turf with a reflectance spectrophotometer. Agron. J. 1968, 60, 640–643. [Google Scholar] [CrossRef]
  33. Jasper, J.; Reusch, S.; Link, A. Active sensing of the N status of wheat using optimized wavelength combination: Impact of seed rate, variety and growth stage. Precis. Agric. 2009, 9, 23–30. [Google Scholar] [CrossRef]
  34. Filella, I.; Penuelas, J. The red edge position and shape as indicators of plant chlorophyll content, biomass and hydric status. Int. J. Remote Sens. 1994, 15, 1459–1470. [Google Scholar] [CrossRef]
  35. Qi, J.; Chehbouni, A.; Huete, A.R.; Kerr, Y.H.; Sorooshian, S. A modified soil adjusted vegetation index. Remote Sens. Environ. 1994, 48, 119–126. [Google Scholar] [CrossRef]
  36. Rondeaux, G.; Steven, M.; Baret, F. Optimization of soil-adjusted vegetation indices. Remote Sens. Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  37. Xing, N.C.; Huang, W.J.; Xie, Q.Y.; Shi, Y.; Ye, H.C.; Dong, Y.Y.; Wu, M.Q.; Sun, G.; Jiao, Q.J. A Transformed Triangular Vegetation Index for Estimating Winter Wheat Leaf Area Index. Remote Sens. 2019, 12, 16. [Google Scholar] [CrossRef] [Green Version]
  38. Sims, D.A.; Gamon, J.A. Relationships between leaf pigment content and spectral reflectance across a wide range of species, leaf structures and developmental stages. Remote Sens. Environ. 2002, 81, 337–354. [Google Scholar] [CrossRef]
  39. Rouse, J.W.; Haas, R.H.; Schell, J.A.; Deering, D.W. Monitoring vegetation systems in the Great Plains with ERTS. NASA Spec. Publ. 1973, 351, 309–317. [Google Scholar]
  40. Chen, J.M. Evaluation of vegetation indices and a modified simple ratio for boreal applications. Can. J. Remote Sens. 1996, 22, 229–242. [Google Scholar] [CrossRef]
  41. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  42. Fitzgerald, G.; Rodriguez, D.; O’Leary, G. Measuring and predicting canopy nitrogen nutrition in wheat using a spectral index—The canopy chlorophyll content index (CCCI). Field Crops Res. 2010, 116, 318–324. [Google Scholar] [CrossRef]
  43. Kimura, R.; Okada, S.; Miura, H.; Kamichika, M. Relationships among the leaf area index, moisture availability, and spectral reflectance in an upland rice field. Agric. Water Manag. 2004, 69, 83–100. [Google Scholar] [CrossRef]
  44. Verrelst, J.; Schaepman, M.E.; Koetz, B.; Kneubühler, M. Angular sensitivity analysis of vegetation indices derived from CHRIS/PROBA data. Remote Sens. Environ. 2008, 112, 2341–2353. [Google Scholar] [CrossRef]
  45. Devadas, R.; Lamb, D.W.; Simpfendorfer, S.; Backhouse, D. Evaluating ten spectral vegetation indices for identifying rust infection in individual wheat leaves. Precis. Agric. 2009, 10, 459–470. [Google Scholar] [CrossRef]
  46. Zha, Y.; Gao, J.; Ni, S. Use of normalized difference built-up index in automatically mapping urban areas from TM imagery. Int. J. Remote Sens. 2003, 24, 583–594. [Google Scholar] [CrossRef]
  47. Ju, C.H.; Tian, Y.C.; Yao, X.; Cao, W.X.; Zhu, Y.; Hannaway, D. Estimating leaf chlorophyll content using red edge parameters. Pedosphere 2010, 20, 633–644. [Google Scholar] [CrossRef]
  48. Clevers, J.G.P.W.; Gitelson, A.A. Remote estimation of crop and grass chlorophyll and nitrogen content using red-edge bands on Sentinel-2 and -3. Int. J. Appl. Earth Obs. Geoinf. 2013, 23, 344–351. [Google Scholar] [CrossRef]
  49. Wu, C.Y.; Niu, Z.; Tang, Q.; Huang, W.J. Estimating chlorophyll content from hyperspectral vegetation indices: Modeling and validation. Agric. For. Meteorol. 2008, 148, 1230–1241. [Google Scholar] [CrossRef]
  50. Drucker, H. Improving Regressors Using Boosting Techniques. In Proceedings of the Icml; Citeseer: Princeton, NJ, USA, 1997; Volume 97, pp. 107–115. [Google Scholar]
  51. Mørup, M.; Hansen, L.K. Automatic relevance determination for multi-way models. J. Chemom. A J. Chemom. Soc. 2009, 23, 352–363. [Google Scholar] [CrossRef]
  52. MacKay, D.J.C. Bayesian interpolation. Neural Comput. 1992, 4, 415–447. [Google Scholar] [CrossRef]
  53. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased Boosting with Categorical Features. Adv. Neural Inf. Process. Syst. 2018, 31, 6638–6648. [Google Scholar]
  54. Myles, A.J.; Feudale, R.N.; Liu, Y.; Woody, N.A.; Brown, S.D. An Introduction to Decision Tree Modeling. J. Chemom. A J. Chemom. Soc. 2004, 18, 275–285. [Google Scholar] [CrossRef]
  55. Wilde, J. Identification of multiple equation probit models with endogenous dummy regressors. Econ. Lett. 2000, 69, 309–312. [Google Scholar] [CrossRef]
  56. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B Stat. Methodol. 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  57. Geurts, P.; Ernst, D.; Wehenkel, L. Extremely randomized trees. Mach. Learn. 2006, 63, 3–42. [Google Scholar] [CrossRef] [Green Version]
  58. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K.; Mitchell, R.; Cano, I.; Zhou, T.; et al. Xgboost: Extreme Gradient Boosting. R Package Version 0.4-2. 2015, 1, 1–4. [Google Scholar]
  59. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  60. Huber, P.J. Robust estimation of a location parameter. Ann. Math. Stat. 1964, 35, 73–101. [Google Scholar] [CrossRef]
  61. Zhang, M.-L.; Zhou, Z.-H. ML-KNN: A Lazy Learning Approach to Multi-Label Learning. Pattern Recognit. 2007, 40, 2038–2048. [Google Scholar] [CrossRef] [Green Version]
  62. Durrant, S.D.; Bissell, J. Performance of kernel-based regression methods in the presence of outliers. J. Process Control 2010, 20, 959–967. [Google Scholar] [CrossRef]
  63. Efron, B.; Hastie, T.; Johnstone, I.; Tibshirani, R. Least angle regression. Ann. Stat. 2004, 32, 407–499. [Google Scholar] [CrossRef] [Green Version]
  64. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  65. GuolinKe, Q.M.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. Lightgbm: A Highly Efficient Gradient Boosting Decision Tree. Adv. Neural Inf. Process. Syst. 2017, 30, 52. [Google Scholar]
  66. Su, X.; Yan, X.; Tsai, C.-L. Linear Regression. Wiley Interdiscip. Rev. Comput. Stat. 2012, 4, 275–294. [Google Scholar] [CrossRef]
  67. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  68. Tropp, J.A.; Gilbert, A.C. Signal Recovery from Random Measurements via Orthogonal Matching Pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef] [Green Version]
  69. Crammer, K.; Dekel, O.; Keshet, J.; Shalev-Shwartz, S.; Singer, Y. Online passive-aggressive algorithms. J. Mach. Learn. Res. 2006, 7, 551–585. [Google Scholar]
  70. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  71. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  72. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  73. Drucker, H.; Burges, C.J.; Kaufman, L.; Smola, A.; Vapnik, V. Support Vector Regression Machines. Adv. Neural Inf. Process. Syst. 1996, 9, 155–161. [Google Scholar]
  74. Theil, H. A Rank-Invariant Method of Linear and Polynomial Regression Analysis. Indag. Math. 1950, 12, 173. [Google Scholar]
  75. Bolhar-Nordenkampf, H.R.; Long, S.P.; Baker, N.R.; Oquist, G.; Schreiber, U.L.E.G.; Lechner, E.G. Chlorophyll fluorescence as a probe of the photosynthetic competence of leaves in the field: A review of current instrumentation. Funct. Ecol. 1989, 3, 497–514. [Google Scholar] [CrossRef]
  76. Kalaji, H.M.; Schansker, G.; Brestic, M.; Bussotti, F.; Calatayud, A.; Ferroni, L. Frequently asked questions about chlorophyll fluorescence, the sequel. Photosynth. Res. 2017, 132, 13–66. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  77. Acevedo, E.; Silva, P.; Silva, H. Wheat Growth and Physiology. Bread Wheat Improv. Prod. 2002, 30, 39–70. [Google Scholar]
  78. Shewry, P.R. Wheat. J. Exp. Bot. 2009, 60, 1537–1553. [Google Scholar] [CrossRef]
  79. Tilling, A.K.; O’Leary, G.J.; Ferwerda, J.G.; Jones, S.D.; Fitzgerald, G.J.; Rodriguez, D.; Belford, R. Remote Sensing of Nitrogen and Water Stress in Wheat. Field Crops Res. 2007, 104, 77–85. [Google Scholar] [CrossRef]
  80. Guo, Y.; Chen, S.; Li, X.; Cunha, M.; Jayavelu, S.; Cammarano, D.; Fu, Y. Machine learning-based approaches for predicting SPAD values of maize using multi-spectral images. Remote Sens. 2022, 14, 1337. [Google Scholar] [CrossRef]
  81. Zarco-Tejada, P.J.; Berni, J.A.; Suárez, L.; Sepulcre-Cantó, G.; Morales, F.; Miller, J.R. Imaging chlorophyll fluorescence with an airborne narrow-band multispectral camera for vegetation stress detection. Remote Sens. Environ. 2009, 113, 1262–1275. [Google Scholar] [CrossRef]
  82. Liu, Y.; Hatou, K.; Aihara, T.; Kurose, S.; Omasa, K. A robust vegetation index based on different UAV RGB images to estimate SPAD values of naked barley leaves. Remote Sens. 2021, 13, 686. [Google Scholar] [CrossRef]
  83. Ahmad, S.; Kalra, A.; Stephen, H. Estimating soil moisture using remote sensing data: A machine learning approach. Adv. Water Resour. 2010, 33, 69–80. [Google Scholar] [CrossRef]
  84. Korotcov, A.; Tkachenko, V.; Russo, D.P.; Ekins, S. Comparison of deep learning with multiple machine learning methods and metrics using diverse drug discovery data sets. Mol. Pharmaceut. 2017, 14, 4462–4475. [Google Scholar] [CrossRef] [PubMed]
  85. Yuan, Y.; Wang, X.; Shi, M.; Wang, P. Performance comparison of RGB and multispectral vegetation indices based on machine learning for estimating Hopea hainanensis SPAD values under different shade conditions. Front. Plant Sci. 2022, 13, 928953. [Google Scholar] [CrossRef] [PubMed]
  86. Zhang, J. Multi-source remote sensing data fusion: Status and trends. Int. J. Image Data Fus. 2010, 1, 5–24. [Google Scholar] [CrossRef] [Green Version]
  87. Yang, L.; Shami, A. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing 2020, 415, 295–316. [Google Scholar] [CrossRef]
  88. Mittlböck, M. Calculating adjusted R2 measures for Poisson regression models. Comput. Meth. Prog. Biomed. 2002, 68, 205–214. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Geographical location of the experimental site.
Figure 1. Geographical location of the experimental site.
Agronomy 13 01003 g001
Figure 2. A photo for Phantom 4 multispectral.
Figure 2. A photo for Phantom 4 multispectral.
Agronomy 13 01003 g002
Figure 3. Calibration and validation results of the highest accuracy model in MS, RGB, RGB + MS data after hyperparameter optimization.
Figure 3. Calibration and validation results of the highest accuracy model in MS, RGB, RGB + MS data after hyperparameter optimization.
Agronomy 13 01003 g003
Figure 4. Feature importance assessment based on random forest model.
Figure 4. Feature importance assessment based on random forest model.
Agronomy 13 01003 g004
Table 1. Soil base fertility information of the experimental site.
Table 1. Soil base fertility information of the experimental site.
YearOrganic Matter
(g/kg)
Alkaline-N
(mg/kg)
Available-P
(mg/kg)
Available-K
(mg/kg)
PH
202014.3159.6321.32117.717.62
202113.9456.2720.83110.477.59
Table 2. Spectral parameters of multispectral sensors.
Table 2. Spectral parameters of multispectral sensors.
BandCenter Wavelength/nmBandwidth/nm
Blue (B)45016
Green (G)56016
Red (R)65016
Red Edge (RE)73016
Infrared (NIR)84026
Table 3. Vegetation indices based on RGB digital number value and calculation method.
Table 3. Vegetation indices based on RGB digital number value and calculation method.
Index NameCalculation FormulaReferences
Red (R), Green (G), Blue (B)Raw digital number value of each band/
Normalized Red r = R / ( R + G + B ) /
Normalized Green g = G / ( R + G + B ) /
Normalized Blue b = B / ( R + G + B ) /
Green Red Ratio Index G R R I = G / R /
Green Blue Ratio Index G B R I = G / B /
Red Blue Ratio Index R B R I = R / B /
Excess Red Vegetation Index E x R = 1.4 × r g [16]
Excess Green Vegetation Index E x G = 2 × g r b [16]
Excess Blue Vegetation Index E x B = 1.4 × b g [16]
Excess Green Minus Excess Red Index E x G R = E x G E x R [16]
Woebbecke Index W I = ( G B ) / ( G + R ) [17]
Normalized Difference Index N D I = ( r g ) / ( r + g + 0.01 ) [17]
Color Intensity I N T = ( R + B + G ) / 3 [18]
Green Leaf Index 1 G L I 1 = ( 2 × G R B ) / ( 2 × G + R + B ) [19]
Green Leaf Index 2 G L I 2 = ( 2 × G R + B ) / ( 2 × G + R + B ) [19]
Vegetative Index V E G = G / R ( 2 / 3 ) × b ( 1 / 3 ) [20]
Combination C O M = 0.25 × E x G + 0.3 × E x G R + 0.33 × C I V E + 0.12 × V E G [21]
Color Index of Vegetation C I V E = 0.441 × r 0.811 × g + 0.3856 × b + 18.79 [22]
Normalized Green–Red Vegetation Index N G R V I = ( G R ) / ( G + R ) [23]
Kawashima Index I K A W = ( R B ) / ( R + B ) [24]
Visible-band difference vegetation Index V D V I = ( 2 × g r b ) / ( 2 × g + r + b ) [25]
Visible Atmospherically Resistance Index V A R I = ( g r ) / ( g + r b ) [26]
Principal Component Analysis Index I P C A = 0.994 × | R B | + 0.961 × | G B | + 0.914 × | G R | [27]
Modified Green Red Vegetation Index M G R V I = G 2 R 2 / G 2 + R 2 [28]
Red Green Blue Vegetation Index R G B V I = G 2 B × R / G 2 + B × R [28]
Table 4. Vegetation indices based on MS sing-band reflectance and calculation method.
Table 4. Vegetation indices based on MS sing-band reflectance and calculation method.
Index NameCalculation FormulaReferences
Difference vegetation index D V I = R n i r R r e d [29]
Enhanced Vegetation Index E V I = 2.5 × R n i r R r e d / R n i r + 6 × R r e d 7.5 × R b l u e + 1 [29]
Leaf chlorophyll index L C I = ( R n i r R r e d e d g e ) / ( R n i r + R r e d ) [29]
Green Normalized Difference Vegetation GNDVI   = R nir   R green   / R nir   + R green   [30]
Ratio Between NIR and Green Bands V I ( nir / green )   = R nir   / R green   [31]
Ratio Between NIR and Red Bands V I ( nir / red ) = R nir   / R r e d [32]
Ratio Between NIR and Red Edge Bands V I ( nir / rededge ) = R nir   / R r e d e d g e [33]
Napierian Logarithm of The Red Edge l n R E = 100 × l n n i r l n r e d [34]
Modified Soil-Adjusted Vegetation Index 1 M S A V I 1 = ( 1 + L ) R n i r R r e d R n i r + R r e d + L ( L = 0.1 ) [35]
Modified Soil-Adjusted Vegetation Index 2 M S A V I 2 = R n i r + 0.5 2 × R n i r + 1 2 8 × R n i r R r e d / 2 [35]
Optimized Soil-Adjusted Vegetation Index O S A V I = ( 1 + 0.16 ) × R n i r R r e d R n i r + R r e d + 0.16 [36]
Modified Triangular Vegetation Index 2 M T V I 2 = 1.5 × 1.2 × R nir   R green   2.5 × R red   R green   2 × R n i r + 1 2 6 × R n i r 5 × R red   0.5 [37]
Normalized Difference Red Edge Index N D R E = R nir   R rededge   R nir   + R rededge   [38]
Normalized Difference Vegetation Index N D V I = R nir   R red   R nir   + R red   [39]
Modified Simple Radio M S R = R n i r R r e d 1 / R nir   + R red   + 1 [40]
Soil-Adjusted Vegetation Index S A V I = R n i r R r e d R n i r + R r e d + 0.5 × ( 1 + 0.5 ) [41]
Simplified Canopy Chlorophyll Content Index S C C C I = N D R E N D V I [42]
Modified Chlorophyll Absorption Reflectance Index MCARI   = R rededge   R red   0.2 × R rededge   R green   × R rededge   R red   [43]
Structure-Insensitive Pigment Index S I P I = R nir   R blue   R n i r + R r e d [44]
Transformed Chlorophyll Absorption Reflectance Index TCARI   = 3 × R rededge   R red   0.2 × R rededge   R green   × R rededge   R red   [45]
Normalized Difference Index N D I = R nir   R rededge   R nir   + R red   [46]
Red-Edge Chlorophyll Index 1 C l 1 = R nir   R rededge   1 [47]
Red-Edge Chlorophyll Index 2 C l 2 = R rededge   R green   1 [48]
Modified Chlorophyll Absorption Reflectance Index 2 M C A R I 2 = 1.5 × 2.5 × R nir   R rededge   1.3 × R nir   R g r e e n 2 × R n i r + 1 2 6 × R n i r 5 × R r e d 2 0.5 [49]
TCARI/OSAVI T C A R I O S A V I [49]
MCARI/OSAVI M C A R I O S A V I [49]
Table 5. Machine learning models.
Table 5. Machine learning models.
ModelAbbreviationReferences
AdaBoost RegressorADA[50]
Automatic Relevance Determination ARD[51]
Bayesian Ridge BR[52]
CatBoost Regressor CatBoost[53]
Decision Tree Regressor DT[54]
Dummy Regressor Dummy[55]
Elastic Net EN[56]
Extra Trees Regressor ET[57]
Extreme Gradient Boosting EGB[58]
Gradient Boosting Regressor GBR[59]
Huber Regressor Huber[60]
K Neighbors Regressor KNN[61]
Kernel RidgeKR[62]
Lasso Least Angle Regression LLAR[63]
Lasso Regression Lasso [64]
Least Angle Regression LAR [63]
Light Gradient Boosting Machine LGBM [65]
Linear Regression LR[66]
Multilayer Perceptron RegressorMLP[67]
Orthogonal Matching Pursuit OMP[68]
Passive Aggressive Regressor PAR[69]
Random Forest Regressor RF[70]
Random Sample Consensus RANSC[71]
Ridge Regression Ridge[72]
Support Vector Machine Regression SVM[73]
TheilSen Regressor TR[74]
Table 6. Basic statistics of the measured Fv/Fm.
Table 6. Basic statistics of the measured Fv/Fm.
DatasetMinimumMaximumMeanSTDEVCV (%)
Total Dataset0.5500.8480.7730.08110.4
Training set0.5510.8460.7750.0779.900
Test set0.5500.8480.7680.09011.700
Table 7. Correlation coefficients between MS, RGB vegetation indices, and Fv/Fm.
Table 7. Correlation coefficients between MS, RGB vegetation indices, and Fv/Fm.
MultispectralCorrelationRGBCorrelation
Vegetation IndicesCoefficientVegetation IndicesCoefficient
DVI0.857b−0.679
EVI0.869g0.827
NDVI0.899r0.283
GNDVI0.888GRRI0.309
NDRE0.850GBRI0.737
LCI0.797RBRI0.502
OSAVI0.892INT−0.816
VI(NIR/G)0.784GRVI0.319
VI(NIR/R)0.724NDI−0.324
VI(NIR/RE)0.807WI0.769
lnRE0.86IKAW0.501
MSAVI10.895GLI0.832
MSAVI20.896GLI2−0.126
MTVI20.863VARI−0.55
MSR0.849ExR−0.435
SAVI0.88ExG0.827
SCCCI0.756ExB−0.743
MCARI−0.793ExGR0.829
MCARI20.754VEG0.767
TCARI−0.71IPCA−0.821
NDI0.865CIVE−0.831
CL10.807COM0.827
CL20.835RGBVI0.843
SIPI0.898MGRVI0.324
TCARI/OSAVI−0.784VDVI0.832
MCARI/OSAVI−0.893
Table 8. Important features selected after data pre-processing.
Table 8. Important features selected after data pre-processing.
VIs TypeImportant Features
RGBRGBVI, ExR
MSMSAVI2
RGB + MSSIPI, ExR, VEG
Table 9. Accuracy assessment of different estimation models based on RGB vegetation indices.
Table 9. Accuracy assessment of different estimation models based on RGB vegetation indices.
ModelMAEMSERMSER2RMLSEMAPETT (s)
GBR0.0230.0010.0330.8000.0190.0320.010
RF0.0240.0010.0330.7950.0190.0320.058
XGB0.0260.0010.0340.7890.0200.0360.106
Catboost0.0240.0010.0340.7850.0200.0330.192
ET0.0240.0010.0340.7820.0200.0330.048
KNN0.0260.0010.0350.7710.0200.0350.008
ADA0.0280.0010.0350.7670.0210.0380.018
Huber0.0300.0020.0390.7110.0220.0400.006
Ridge0.0310.0020.0390.7070.0230.0410.006
LR0.0300.0020.0390.7070.0230.0400.006
LAR0.0300.0020.0390.7070.0230.0400.006
BR0.0300.0020.0390.7070.0230.0410.006
OMP0.0310.0020.0390.7070.0220.0410.004
ARD0.0310.0020.0390.7070.0220.0410.006
DT0.0310.0020.0400.6880.0230.0420.004
TR0.0360.0030.0490.5720.0290.0510.178
PAR0.0420.0030.0500.5280.0290.0560.006
LGBM0.0490.0040.0630.2830.0360.0670.018
RANSC0.0430.0050.0650.1950.0370.0620.008
MLP0.0550.0050.0670.1580.0380.0720.014
lasso0.0580.0060.076−0.0590.0450.0820.006
EN0.0580.0060.076−0.0590.0450.0820.006
LLAR0.0580.0060.076−0.0590.0450.0820.008
Dummy0.0580.0060.076−0.0590.0450.0820.004
SVM0.0720.0060.077−0.1040.0440.0930.006
KR0.7920.6320.795−116.0500.5231.0300.006
Table 10. Accuracy assessment of different estimation models based on MS vegetation indices.
Table 10. Accuracy assessment of different estimation models based on MS vegetation indices.
ModelMAEMSERMSER2RMLSEMAPETT (s)
Huber0.0210.0010.0270.8600.0150.0280.006
LR0.0220.0010.0270.8600.0150.0290.316
Ridge0.0220.0010.0270.8600.0150.0290.264
LAR0.0220.0010.0270.8600.0150.0290.004
OMP0.0220.0010.0270.8600.0150.0290.004
BR0.0220.0010.0270.8600.0150.0290.004
ARD0.0220.0010.0270.8600.0150.0290.006
TR0.0220.0010.0280.8490.0160.0290.020
KNN0.0260.0010.0330.7940.0190.0350.006
RF0.0320.0010.0380.7370.0220.0430.054
Catboost0.0330.0020.0390.7230.0220.0450.190
ADA0.0310.0020.0390.7160.0230.0420.012
GBR0.0340.0020.0400.7040.0230.0460.010
ET0.0340.0020.0410.6950.0230.0460.044
PAR0.0350.0020.0410.6840.0230.0460.006
XGB0.0360.0020.0420.6680.0240.0490.132
DT0.0370.0020.0430.6520.0250.0500.006
RANSC0.0350.0030.0470.5460.0270.0490.006
LGMB0.0450.0040.0620.2990.0360.0640.016
SVM0.0640.0050.0690.0970.0390.0820.006
Lasso0.0580.0060.076−0.0590.0450.0820.298
EN0.0580.0060.076−0.0590.0450.0820.006
LLAR0.0580.0060.076−0.0590.0450.0820.004
Dummy0.0580.0060.076−0.0590.0450.0820.004
MLP0.0760.0080.090−0.5670.0520.1010.016
KR0.7840.6190.787−113.5520.5281.0240.006
Table 11. Accuracy assessment of different estimation models based on RGB and MS indices.
Table 11. Accuracy assessment of different estimation models based on RGB and MS indices.
ModelMAEMSERMSER2RMLSEMAPETT (s)
ARD0.0210.0010.0260.8680.0150.0280.006
OMP0.0210.0010.0260.8680.0150.0280.007
Ridge0.0220.0010.0270.8580.0150.0290.006
LR0.0220.0010.0270.8580.0150.0290.514
LAR0.0220.0010.0270.8580.0150.0290.006
BR0.0220.0010.0270.8580.0150.0290.006
Huber0.0220.0010.0270.8570.0160.0290.006
TR0.0220.0010.0280.8490.0160.0300.178
RANSC0.0240.0010.0300.8300.0170.0320.010
KNN0.0230.0010.0300.8260.0170.0310.006
RF0.0250.0010.0330.8000.0190.0330.052
ET0.0250.0010.0340.7850.0200.0330.050
ADA0.0270.0010.0360.7560.0210.0370.020
Catboost0.0260.0010.0360.7530.0210.0360.214
XGB0.0270.0020.0380.7230.0220.0370.092
GBR0.0280.0020.0390.7210.0220.0370.012
DT0.0320.0020.0410.6900.0240.0430.006
PAR0.0390.0020.0470.5930.0270.0510.006
LGBM0.0430.0030.0550.4290.0320.0600.018
Lasso0.0580.0060.076−0.0590.0450.0820.006
EN0.0580.0060.076−0.0590.0450.0820.008
LLAR0.0580.0060.076−0.0590.0450.0820.006
Dummy0.0580.0060.076−0.0590.0450.0820.004
SVM0.0710.0060.076−0.0730.0430.0920.006
MLP0.0630.0080.082−0.2920.0470.0850.020
KR0.7990.6460.803−118.1590.5121.0420.006
Table 12. Hyperparameter optimization of the highest accuracy model in three image types.
Table 12. Hyperparameter optimization of the highest accuracy model in three image types.
Image TypeModelDatasetMAEMSERMSER2RMLSEMAPE
RGBGBRTraining set0.0140.0000.0210.9250.0120.019
Test set0.0270.0010.0360.8380.0200.037
MSHuberTraining set0.0210.0010.0270.8780.0150.028
Test set0.0180.0010.0250.9200.0140.024
MS + RGBARDTraining set0.0210.0010.0260.8820.0150.028
Test set0.0190.0010.0240.9250.0140.026
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Q.; Zhang, Y.; Xie, M.; Zhao, Z.; Yang, L.; Liu, J.; Hou, D. Estimation of Fv/Fm in Spring Wheat Using UAV-Based Multispectral and RGB Imagery with Multiple Machine Learning Methods. Agronomy 2023, 13, 1003. https://doi.org/10.3390/agronomy13041003

AMA Style

Wu Q, Zhang Y, Xie M, Zhao Z, Yang L, Liu J, Hou D. Estimation of Fv/Fm in Spring Wheat Using UAV-Based Multispectral and RGB Imagery with Multiple Machine Learning Methods. Agronomy. 2023; 13(4):1003. https://doi.org/10.3390/agronomy13041003

Chicago/Turabian Style

Wu, Qiang, Yongping Zhang, Min Xie, Zhiwei Zhao, Lei Yang, Jie Liu, and Dingyi Hou. 2023. "Estimation of Fv/Fm in Spring Wheat Using UAV-Based Multispectral and RGB Imagery with Multiple Machine Learning Methods" Agronomy 13, no. 4: 1003. https://doi.org/10.3390/agronomy13041003

APA Style

Wu, Q., Zhang, Y., Xie, M., Zhao, Z., Yang, L., Liu, J., & Hou, D. (2023). Estimation of Fv/Fm in Spring Wheat Using UAV-Based Multispectral and RGB Imagery with Multiple Machine Learning Methods. Agronomy, 13(4), 1003. https://doi.org/10.3390/agronomy13041003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop