Next Article in Journal
Daytime Sea Fog Identification Based on Multi-Satellite Information and the ECA-TransUnet Model
Previous Article in Journal
SaTSeaD: Satellite Triangulated Sea Depth Open-Source Bathymetry Module for NASA Ames Stereo Pipeline
Previous Article in Special Issue
Exploring Multisource Feature Fusion and Stacking Ensemble Learning for Accurate Estimation of Maize Chlorophyll Content Using Unmanned Aerial Vehicle Remote Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Seedling Oilseed Rape Crop Phenotype by Drone-Derived Multimodal Data

1
School of Engineering, Huazhong Agricultural University, Wuhan 430070, China
2
College of Plant Science and Technology, Huazhong Agricultural University, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(16), 3951; https://doi.org/10.3390/rs15163951
Submission received: 7 July 2023 / Revised: 30 July 2023 / Accepted: 31 July 2023 / Published: 9 August 2023

Abstract

:
In recent years, unmanned aerial vehicle (UAV) remote sensing systems have advanced rapidly, enabling the effective assessment of crop growth through the processing and integration of multimodal data from diverse sensors mounted on UAVs. UAV-derived multimodal data encompass both multi-source remote sensing data and multi-source non-remote sensing data. This study employs Image Guided Filtering Fusion (GFF) to obtain high-resolution multispectral images (HR-MSs) and selects three vegetation indices (VIs) based on correlation analysis and feature reduction in HR-MS for multi-source sensing data. As a supplement to remote sensing data, multi-source non-remote sensing data incorporate two meteorological conditions: temperature and precipitation. This research aims to establish remote sensing quantitative monitoring models for four crucial growth-physiological indicators during rapeseed (Brassica napus L.) seedling stages, namely, leaf area index (LAI), above ground biomass (AGB), leaf nitrogen content (LNC), and chlorophyll content (SPAD). To validate the monitoring effectiveness of multimodal data, the study constructs four model frameworks based on multimodal data input and employs Support Vector Regression (SVR), Partial Least Squares (PLS), Backpropagation Neural Network (BPNN), and Nonlinear Model Regression (NMR) machine learning models to create winter rapeseed quantitative monitoring models. The findings reveal that the model framework, which integrates multi-source remote sensing data and non-remote sensing data, exhibits the highest average precision (R2 = 0.7454), which is 28%, 14.6%, and 3.7% higher than that of the other three model frameworks, enhancing the model’s robustness by incorporating meteorological data. Furthermore, SVR consistently performs well across various multimodal model frameworks, effectively evaluating the vigor of rapeseed seedlings and providing a valuable reference for rapid, non-destructive monitoring of winter rapeseed.

1. Introduction

Field crop phenotypic information refers to the physical, physiological, and biochemical characteristics of crop growth and development [1], such as the leaf area index (LAI), above-ground biomass (AGB), leaf nitrogen content (LNC), and chlorophyll content, which are influenced by internal and external environmental factors [2,3,4]. These growth physiological parameters are important plant indicators for dynamic monitoring of vegetation growth and reflect the growth of crops [5]. The seedling stage has a decisive influence on growth, development, and yield formation, and growth monitoring of winter rape seedlings plays an important role in decision making for field production and crop regulation [6]. Traditional measurement methods rely on manual collection in the field, which is not only time-consuming and laborious, but may also cause some damage to the plant, while lacking real-time and spatial distribution accuracy [7]. Therefore, rapid, accurate, and nondestructive measurement of plant biomass is of great value in all aspects of precision agriculture.
To address this problem, research on the use of unmanned aerial vehicle (UAV) remote sensing to estimate crop growth and physiological parameters has been emerging. Remote sensing, as a cutting-edge information technology for terrain observation, can be used to quickly and accurately obtain real-time information on crop growth and physiology over large areas, and has been widely used in agriculture in recent years [8,9,10]. Multispectral images have the advantages of high spatial resolution and ease of operation [11]. Monitoring of field crops can be achieved quickly and nondestructively using the broadband extracted by UAVs carrying multispectral cameras in combination with existing spectral indices [12,13]. Many studies have shown that, by combining UAV multispectral images with high spatial resolution vegetation indices (VIs) and using existing mature machine learning algorithms, reliable models can be built to effectively and nondestructively monitor plant growth [14,15,16]. The high-spectral, multispectral (MS), and visible images (RGB) captured through UAV-based remote sensing exhibit distinct characteristics. Integrating these images with machine learning algorithms enables robust monitoring of crop growth. Nevertheless, the simultaneous leveraging of the advantages offered by multiple sensors remains underexplored.
With the advancements in remote sensing and agricultural technology, it is now possible to acquire a variety of remote sensing image data and non-remote sensing data, such as meteorological and soil data, from multiple sensors and time periods within the same geographical area. These datasets collectively form the multimodal data within the region [17,18]. Multimodal data refers to the fusion of diverse data sources, synthesizing the image information of multiple imaging sensors for the same target. Effective integration of the complementary information from different data sources mitigates the limitations of single-source data, including incomplete interpretation, uncertainty, and errors associated with monitoring the target, thereby enhancing the efficiency and depth of utilizing multi-source data [19,20]. Multimodal data can be categorized into two parts: (1) fusion between multi-source remote sensing data; (2) fusion between multi-source remote sensing data and multi-source non-remote sensing data.
In the field of multi-source remote sensing data fusion, fusion of multi-source data can make up for the limitations of a single image and increase the quality of experience (QoE), while enhancing the quality of remote sensing images. In addition, high-resolution remote sensing images have a positive impact on the accuracy of the subsequently constructed models [21]. There is a growing interest in using multi-source data to estimate crop growth in the field because fusing multi-source data can compensate for the limitations of a single image, and many studies have established high-precision nondestructive estimation models based on UAV remote sensing [22,23,24]. This research has shown that multimodal features are more advantageous than single-modal features, and can improve the feasibility and accuracy of the model in many ways, such as by fusing audio–visual information to improve some unimodal visual analysis systems [25,26,27]. However, using all VIs and texture features (Texs) obtained from different remote sensing sensors only as inputs to the monitoring models may lead to data redundancy and problems such as multicollinearity, which instead reduce the robustness of the monitoring models [28]. In addition, multi-source non-remote sensing data are underutilized in these models. Therefore, many scholars have proposed methods with deeper levels of fusion, such as remote image fusion methods in the multi-scale morphological gradient (MSMG) structural domain [29], and the hybrid fusion method of IR and visual images combining discrete smooth wavelet transform (DSWT), discrete cosine transform (DCT), and local spatial frequency (LSF) [30], to fuse multi-source remote sensing data using image fusion algorithms to enhance the spatial resolution of remote sensing images while maintaining the original information of the spectrum.
The fusion between multi-source non-remote sensing data and multi-source remote sensing data mainly involves the participation of non-remote sensing information, such as meteorological data, soil data, and geographic information, as auxiliary variables in remote sensing monitoring and classification applications [31]. Existing studies have demonstrated that local changes in the crop canopy as a response to environmental and field management changes affect crop yield, suggesting that non-remote sensing information as a feature may have a powerful role in crop yield prediction by incorporating non-remote sensing data in model training, and that visible or multispectral images can yield better prediction results [32,33,34]. However, there is a dearth of research on image fusion algorithms combined with multi-source non-remote sensing features to estimate the phenotypic information of oil-seed rape, which poses challenges in ensuring the stability of models in predicting crop growth information in different fields.
Based on this situation, an oil-seed rape test field in Shashi District, Jingzhou City, Hubei Province, was used as the study area in this study. Employing an enhanced image fusion algorithm, the study merges UAV visible light images with multispectral images. Subsequently, four widely used machine learning prediction methods—PLSR, NLR, SVR, and BP-NN—are harnessed to effectively integrate diverse non-remote sensing data sources for monitoring the growth and physiological parameters of rapeseed in the field. The main objectives were to (1) apply four regression methods (PLSR, NLR, SVR, and BP-NN) to establish a model for monitoring growth and physiological parameters of field rapeseed; (2) compare the performance of estimating physiological parameters of oil-seed rape growth using traditional VIs, Texs, and four input model frameworks based on multimodal data composition; and (3) compare and analyze the estimation results of the models to determine the best model for each growth index.

2. Materials and Methods

2.1. Field Experiment and Biomass Sampling

An experiment was conducted at Jingzhou Agricultural Science Academy, Shashi District, Jingzhou City, Hubei Province (112°20′35″E, 30°14′17″N). The test field covered an area of about 3800 m2, as shown in Figure 1, which is a schematic diagram of the field test area and the UAV remote sensing photography. The oil-seed rape cultivar was mainly Huayouza 50, jointly developed by Huazhong Agricultural University and Wuhan Liannong Seed Technology Co., Ltd. (Wuhan, China) with registration number GPD Oil-seed rape (2017) 420204. This trial was conducted from September 2022 to March 2023, depending on the developmental progress of winter oil-seed rape. A single-factor experiment was set up as follows: three N application levels: 8 kg/667 m2 (N8), 12 kg/667 m2 (N12), 16 kg/667 m2 (N16); three density treatments: 10,000 plants/667 m2 (D1), 30,000 plants/667 m2 (D3), 50,000 plants/667 m2 (D5); three sowing periods: September 25 (S925), October 10 (S1010), and October 25 (S1025). The distribution of the rape trial area and treatments are shown in Figure 2. The trial was set up with multiple replications; each plot area was 2 m × 2 m, row spacing was 0.5 m, the whole trial field was shaped like a trapezoid, and multiple protection rows were set up with a total of 546 plots. Except for the above treatment differences, other management measures were the same as the local high-yielding cultivation measures.

2.2. Collection and Processing of Rapeseed Phenotype Information

During the experiment, seedling oil-seed rape data, including UAV-based multispectral imaging data, LAI, SPAD, LNC, and AGB, were collected from November 2022 to March 2023, using an AccuPAR LP80 racing Radiation and Architecture of Canopies meter, a SPAD-502 chlorophyll meter, an NKY-6120 Nitrogen Analyzer, and electronic scales (Figure 3), in four consecutive collections during the critical fertility period of winter oil-seed rape in 21 November 2022, 8 December 2022, 10 January 2023, and 30 January 2023.

2.3. UAV Systems and Flight Missions

The UAV used a DJI Mavic 3M (Figure 4a) in this test to simultaneously collect RGB and multispectral imagery. The total weight of the UAV was 1.05 kg, the maximum pitch angle was 35°, the maximum horizontal flight speed was 21 m/s, and the flight endurance was about 43 min. Image acquisition was performed using the visible light camera and multispectral camera (Figure 4b) equipped with the UAV, and the details of the sensors are shown in Table 1. The choice of the wavelength is very important for the calculation of the VIs but, due to equipment limitations, the bandwidth and wavelength shift issues are not taken into account in the calculations, and multiple central band data from multispectral cameras are directly used [35]. The UAV was also equipped with a multispectral light intensity sensor on top of the UAV, which can monitor the incident light intensity in real time and compensate for multispectral imaging. The DJI RC Enterprise (Figure 4a) was used to automatically generate the flight routes, and Figure 4c shows the two-dimensional (2D) routes based on satellite map data as well as the three-dimensional (3D) routes based on the UAV elevation data.
Prior to the UAV mission, the heading/bypass overlap was set to 75% and 70%, respectively, and the maximum flight speed was 5 m/s, and the shooting time was selected between 10 a.m. and 2 p.m. when the weather was clear and there was direct sunlight. Four flights were conducted to simultaneously acquire UAV imagery at a flight altitude of 40 m and collect phenotypic data of oil-seed rape. Table 2 shows the acquisition of remote sensing data and field trial data.

2.4. Image Processing and Feature Extraction

2.4.1. Image Processing

The detector relative spectral response (RSR) shift effect affects the uniformity of the collected radiation signals. In this study, the camera calibration function in Metashape software was used to align and correct the UAV images, and operations such as image alignment, dense point cloud creation, grid generation, texture generation, and elevation image generation on the UAV images were carried out sequentially to obtain a high-resolution digital orthophoto map (DOM) [36]. Hazy scenes reduce the feasibility of image analysis; however, due to the suitable environment and proper conditions of the shooting process, the images collected by the UAV in this study can be regarded as fog-free images, and a de-fogging process was not needed [37]. The RGB images and the four-band MS images are stitched together to cover the whole test area, where MS_G, MS_R, MS_B, and MS_NIR represent the corresponding single-band spectral images. The stitched RGB images were processed using MATLAB 2020b, and the multispectral band images were independent (i.e., MS_G, MS_R, etc.). The band fusion of the MS image set was implemented using ArcGIS software, and the fusion mode of standard pseudo-color was selected to assign the three bands of NIR MS_NIR, red MS_R, and green MS_G to red, green, and blue colors, respectively, and the obtained images of vegetation or crops in red color. The bit depth of MS images after band fusion is 32 bits, and the remote sensing image is rendered to a depth of 8 bits, so that the rendered image after band fusion is processed in ArcGIS and can be read and processed by software such as MATLAB and Python. Finally, the imported band-fused standard dummy color large-field images are cropped using ENVI software to obtain images of individual fields.

2.4.2. VIs

In remote sensing applications, vegetation indices have been widely used to qualitatively and quantitatively evaluate vegetation cover and its growth vigor [38]. The vegetation indices used in this study are shown in Table 3.

2.4.3. GCFs

In order to improve the accuracy of the model, many scholars further extract texture features from each multispectral-based band to construct the model. In this study, a texture parameter based on gradual change features (GCFs) calculated from NDVI was chosen to construct a new texture index by obtaining the structural distribution characteristics of the spectral indices. NDVI, as one of the most commonly used spectral indices in agricultural remote sensing monitoring applications, is highly sensitive to changes in crop growth and physiological parameter indices, and can well distinguish between crop groups with different measures of growth potential [46]. In a 2 m × 2 m plot, a 1 m × 1 m area in the center was selected, and the grayscale images of NDVI were classified into five categories according to the size of the image elements using the K-means clustering algorithm: NDVI minimum area (A), NDVI small area (B), NDVI medium area (C), NDVI large area (D), and NDVI maximum area (E); the number of image elements occupied by each category was counted to characterize the area occupied by each category. The flow chart of this process is shown in Figure 5.
The NDVI values of the five types of areas were averaged as VA, VB, VC, VD, and VE, and the image elements were recorded as the area of the five areas as SA, SB, SC, SD, and SE. Generally, the areas with NDVI values less than 0.2 can be classified as non-vegetation covered areas. The VA and SA corresponding to this part can be discarded to improve the accuracy of the gradient feature data. Four gradient feature indicators were designed for four types of NDVI and area share, which are the vegetation index coefficient of variation (Vcv), area coefficient of variation (Scv), vegetation compactness (Rc), and vegetation density coefficient of variation (Ra), as expressed in Equations (1)–(6):
V m = V B + V C + V D + V E 4
S m = S B + S C + S D + S E 4
V c v = V B V m 2 + V C V m 2 + V D V m 2 + V E V m 2 4
S c v = S B S m 2 + S C S m 2 + S D S m 2 + S E S m 2 4
R c = S B 4 × S m
R a = V E V B 4 × V m

2.5. Non-Remote Sensing Auxiliary Data

Non-remote sensing data can be used as a supplement to remote sensing data and help to improve the scientific nature of the study. Daily maximum temperature (Tmax), minimum temperature (Tmin), and average rainfall (Aar) were selected as non-remote sensing data, and the data were obtained from the National Weather Science Data Center (http://data.cma.cn, accessed on 16 March 2023). The experiment collected meteorological data of the area where Daejeon is located, from 25 September 2022 to 30 January 2023. The daily average rainfall can be calculated by monthly average rainfall (monthly average rainfall divided by the number of days in the month), and the daily average rainfall was summed according to the time interval of remote sensing monitoring in this study as the rainfall data for the monitoring model.
The maximum temperature, Tmax, and minimum temperature, Tmin, in the non-remote sensing data are time series and dynamic, which are difficult to combine effectively with the remote sensing data collected in multiple time intervals. Therefore, this study introduced the effective cumulative temperature, Ae, to convert the time series continuous data into non-temporal discrete data, which was the sum of the effective temperature of the crop at a certain reproductive period and, numerically, the sum of the difference between the average temperature at a certain time period and the biological zero of the crops, as in Equation (7) [47]:
A e = i = 1 n T i B
where Ti is the average temperature during the ith period; B is the biological zero, which is the minimum temperature required to meet the crop’s continued growth and development, and is related to the crop’s species and development time. The biological zero of the oil-seed rape species in this study is generally between 4 °C and 5 °C, and here B = 4 °C was chosen.

3. Multimodal Data Fusion

Multi-source data fusion can be divided into two parts: (1) the fusion between multi-source remote sensing data; (2) the fusion between multi-source remote sensing data and multi-source non-remote sensing data. The former fully fuses UAV multispectral images and visible images using image fusion algorithms to realize the enhancement of image information, and the latter realizes the organic combination of multiple variables through machine learning algorithms.

3.1. Image Fusion

In this study, an edge-preserving filtering algorithm, guided filtering (GF), which was developed by Li et al. [48], was chosen. The algorithm is based on a local linear model to guide the information of the digital image to calculate the filtering output, which can be used in applications such as upsampling, local cropping, color space conversion, and multi-scale decomposition in image processing. In the multi-scale transformation session, GF takes another image as the guide, and this guide image can be the input or even the same image of this decomposition layer. By analyzing the distribution of pixel neighborhoods, a linearly invariant output image is generated, consisting primarily of an approximation image and a structure image. This approach effectively preserves the structural characteristics of the source image, facilitating efficient multi-scale decomposition.
The study performs multi-scale decomposition using GF, and the decomposed multi-scale representation can be reconstructed as a source image using Inspiratory Muscle Strength Training (IMST), as shown in Figure 6:
Suppose the source image used for multi-scale decomposition is IR, and the operator performing the decomposition operation is named G , as shown in Equation (8):
  operator   G   ,   I o u t = G β , I i n w h e r e I o u t = arg I o u t min I i n I o u t 2 2 + β α x G x I o u t 2 2 + α y G y I i n 2 2      α x = I i n x α , α y = I i n y α
where Gx and Gy are the difference operators for horizontal and vertical directions, respectively; the parameter β is a regularization constant to control the balance between horizontal and vertical targets; Iin is the image used for multi-scale decomposition; and Iout is calculated from the equations constructed by the operator G.
Let the images after GF decomposition be divided into two categories: an approximate image I R C and a set of detailed images I R D i ,   i = 1 , 2 , , n 1 ; then, I R C and I R D i can be shown as in Equation (9):
I R E ( i ) = G β i , I R E ( i 1 ) , i = 1 , 2 , , n 1 I R E ( 0 ) = I R , I R C = I R E ( n 1 ) I R D i = I R E ( i 1 ) I R E ( i )
where I R E ( n 1 ) is the image that has undergone n − 1 decomposition, which has the lowest resolution; the last decomposed image is generally taken as the approximate image of GF decomposition I R C , and the difference in the approximate images between different decomposition layers is considered as the detailed image of multi-scale decomposition.
Taking a UAV remote sensing image acquired from a large field as an example, one approximate image and two detailed images obtained after three GF decompositions are shown in Figure 7. The approximate image has the same color distribution as the source image, and the detailed image shows the rape field and the leaf texture of the crop within the field, so the GF decomposition can store the color or spectral information of the source image in the approximate image I R C and the structural and spatial information in the detailed image I R D i .
The decomposed image can construct the input source image by approximating the image and detailed image, which achieves the inverse multi-scale transformation and nearly lossless restoration of the source image. This is the theoretical basis of the multi-scale decomposition-based image fusion algorithm, and the transformation process is shown in Equation (10):
I R = I R C + i = 1 n 1 I R D ( i ) , i = 1 , 2 , , n 1
In the GF image fusion algorithm, if the source image is I 1 , I 2 , , I M for a total of M images, the approximate image I 1 C , I 2 C , , I M C and the detailed image I 1 D ( i ) , I 2 D ( i ) , , I M D ( i ) can be obtained after the calculation of Equations (9) and (10).
The GF decomposed image needs to be fused according to the fusion rules, given a set number of source images, where the nth image is f. Laplace filtering is performed on image f to obtain a high-pass image H n , as shown in Equation (11):
H n = h n f L ( f ) h n f = 1 m 2 1 1 1 1 1 1 1 1 1
where L f represents the logarithmic spectrum of the image and h n f is an m × m matrix. Usually, m = 3 and the change in m has a small effect on the calculation of the significance map. The local average of the absolute values of H n is used to construct the significance map S n , as shown in Equation (12):
S n = H n g r g , δ g
where g is the low-pass filter, the size of rg and δg is m − 1, and the size of the low-pass filter is [2(m − 1) + 1]2. The same remote sensing image of a large field of oil-seed rape exemplified in Section 3.1 is used to calculate the log spectrum and significance map of the image. Finally, the significance map extracted in the previous step is used to calculate the initial weight map, Pn, of the source image, which is calculated as shown in Equation (13):
P n ( p , q ) = 1    i f S n ( p , q ) = max S 1 ( p , q ) , S 2 ( p , q ) , , S n ( p , q ) 0    o t h e r w i s e
where S n ( p , q ) denotes the significant value at the pixel ( p , q ) in the nth image. The weighted average of the weights of the approximation and detailed images for the best estimate is calculated as the weight maps of P C and P D ( i ) , where i is the number of layers of the decomposition. Then, the new fusion rule is shown in Equation (14):
C ¯ = P C I 1 + ( 1 P C ) I 2 D ¯ = i = 1 n 1 P D ( i ) I 1 D ( i ) + i = 1 n 1 ( 1 P D ( i ) ) I 1 D ( i ) F = C ¯ + D ¯
where P C ( k ) and P D ( i ) are the weight values at the kth pixel of the weight map, which is in the range [0, 1]. C ¯ is the fused approximate image, D ¯ is the fused detailed image, and F is the fused image.

3.2. Machine Learning

Machine learning algorithms possess exceptional capabilities in nonlinear regression prediction. They find increasingly widespread application in precision agriculture and remote sensing monitoring, demonstrating particularly remarkable performance in scenarios such as remote sensing image segmentation, land cover classification, and phenotypic indicator monitoring. By leveraging these algorithms, it becomes possible to effectively and accurately monitor the growth of oil-seed crops by seamlessly integrating multi-source remote sensing and non-remote sensing data. In this study, we selected four commonly used machine learning regression prediction algorithms: PLSR, NLR, SVR, and BP-NN.
PLSR is an extension of the least squares method that effectively addresses the issue of multicollinearity among variables. It offers simplicity in computation and high predictive accuracy. Assuming the input data of the model, denoted X, are in the form of an N × M dimensional matrix, and the corresponding model output, denoted Y, is an N × 1 dimensional matrix, performing matrix decomposition on the input and output yields the result shown in Equation (15):
X = T P + B 1 Y = U Q + B 2
where T and U are the component score matrices of X and Y, P and Q are the factor loading matrices, and B1 and B2 are the residuals fitted by the PLS algorithm. A regression relation U = TE (E is the regression coefficient matrix) is established for the component score matrices T and U of model input X and output Y, and by substitution, Equation (16) can be obtained:
Y   =   E T Q   +   B 2
A linear regression prediction model can be built when the response value Y corresponds to the growth and physiological parameter indices of rapeseed seedlings and the input X corresponds to multiple sources of data with different modalities.
NMR is an extension of multiple linear regression. Multiple linear regression establishes the relationship between two or more input variables X and output variable Y. By combining the optimal combination of multiple interrelated factors of input variables, the output variable is jointly predicted or estimated. However, in practical applications, many of the input variables do not exhibit purely linear relationships with the output variables. By introducing interaction terms, non-linear variables can be combined to fit the non-linear part of the output variables. Assuming that y is the output variable and x 1 , x 2 , , x k is the input variable, the model of NMR can be represented as shown in Equation (17):
y = β 0 + i = 1 k β i x i + i = 1 k j = 1 k β i j x i x j + ε
where β 0 is the constant term, β i is the linear regression coefficient, β i j is the nonlinear regression coefficient, and ε is the fitting error.
SVR is a nonlinear regression method suitable for solving small-sample, high-dimensional problems. If the sample data used for regression training are xi and yi, where i = 1, 2, …, n, xi is the sample value of the input vector x consisting of n training patterns, and yi is the corresponding value of the desired model output. Then, the output of the regression model yi can be expressed as shown in Equation (18):
y = w T ϕ ( x ) + b
where the coefficients w and b are adjustable model parameters, w is a one-dimensional array, and ϕ ( x ) is a nonlinear transformation function that maps the input space to a high-dimensional feature space. The parameters w and b in the equation are then estimated by minimizing the cost function J w , ξ , ξ i * , which is defined by Equation (19):
min J w , ξ , ξ i * = 1 2 w 2 + C i = 1 N ξ + ξ i * s . t .     y i y i = ε + ξ i , i = 1 , 2 , , N      y i y i = ε + ξ i * , i = 1 , 2 , , N      ξ i 0 , ξ i * 0 , i = i = 1 , 2 , , N
where ξ i and ξ i * are positive relaxation variables, ε is the distance between y i and y i , and C is a positive real constant. The values of w and b in the equation and their modal outputs are obtained by calculating them in MATLAB software.
BPNN is an algorithm that uses error back propagation, which mainly consists of an implicit layer, an input layer, and an output layer. By back propagating the mean square error to the input layer, the connection weights between each neural layer are continuously modified until the actual output value has the minimum error with the predicted value.

4. Results and Discussion

4.1. Correlation Analysis

To avoid possible overfitting problems in monitoring models, researchers have explored the application of the Spearman correlation coefficient to perform multivariate correlation analysis on the independent variables (multi-source remote sensing data and multi-source non-remote sensing data) and dependent variables (physiological parameters of oil-seed rape growth) used in model construction [49]. The Spearman coefficient enables the analysis of the correlation between variables that do not conform to normality assumptions in the data. Figure 8 shows the test results of data normality, where the diagonal line of the matrix plot is a univariate density plot, which can show the type of data distribution of the variables. The other parts of the scatter matrix plots represent the linear correlation between different variables. In this study, the correlation analysis of Spearman was chosen because the linear correlation between the multi-source data and the oil-seed rape growth physiological data was poor and only a small portion of the data satisfied the normal distribution.
Figure 9 shows the Spearman correlation multivariate analysis; the correlation coefficients between different independent variables, and between independent variables and dependent variables, are shown in the figure, and the correlation coefficients are visualized in a heat map. Among these, the correlation between the vegetation closeness, Rc, of the multi-source remote sensing data and all four growth physiological indicators of oil-seed rape was poor, with correlation coefficients below 0.1. This may be related to the period of the crop, and the parameter is more suitable for modeling during the period from sowing to seedling emergence of oil-seed rape, rather than the period when the crop canopy covers a large area of soil. In addition, the correlation of the variables of the six spectral indices based on multi-source remote sensing data is high, which are prone to dimensional disasters when involving high-dimensional problems, leading to model overfitting; thus, it is necessary to reduce the dimensionality of the six spectral indices.

4.2. PCA Data Dimensionality Reduction

This study uses principal component analysis (PCA) to reduce the dimensionality of spectral indices. PCA, as an algebraic theory-based data dimensionality reduction method, can transform multiple variables into several linearly uncorrelated orthogonal vectors to completely represent the decision space [50,51]. There exists a certain degree of linear correlation among the original six spectral indices, allowing for the synthesis of information and features among multiple variables using a reduced set of composite variables. The raw data were processed using PCA to obtain the contributions of the variables, and variables were arranged in order from the largest to the smallest contribution. The combination of variables with a cumulative contribution of 95% was selected as the result of data dimensionality reduction. Figure 10 illustrates the distribution of variable contributions and the three-dimensional representation of the spectral data after dimensionality reduction. Among these, the cumulative contribution of NDVI, NRI, and MSR reached 95%; thus, the original spectral indices can be effectively reduced to this combination of three variables.
In summary, after a series of data analyses and processing, the multi-source data and model outputs used for the monitoring model of oil-seed rape growth parameters in the field are: (1) multi-source remote sensing data after dimensionality reduction: spectral indices NDVI, NRI and MSR; texture features Vcv, Scv and Ra; (2) multi-source non-remote sensing data: meteorological data Ae and Aar; and (3) growth physiological parameters: LAI, AGB, SPAD, and LNC.

4.3. Phenotypic Prediction of Rapeseed Crops during Seedling Stage

4.3.1. Four Multimodal Data Model Frameworks

To ascertain the efficacy of utilizing diverse data sources for monitoring the growth of rapeseed during the early stages in agricultural fields, this study constructed a four-input Model Framework for Multimodal data (MFM) based on distinct multimodal data inputs: (1) a UAV remote sensing monitoring model based on a single data source; (2) a UAV remote sensing monitoring model based on a single data source (adding texture features); (3) a UAV remote sensing monitoring model based on the fusion of multi-source remote sensing data; and (4) a UAV remote sensing monitoring model based on the fusion of multi-source remote sensing data and multi-source non-remote sensing data.
Table 4 shows the differences between the four model frameworks. MFM1 used the field remote sensing monitoring method chosen by many studies, which only acquired MS images through a single data source (multispectral sensor) from UAV remote sensing, and, after processing, obtained spectral indices (VIs) to establish a growth monitoring model of oil-seed rape in fields [52,53,54]. MFM2 added spatial texture feature information, which is a common method used to enhance the accuracy of monitoring models [55,56]. MFM3 is a remote sensing monitoring method based on the fusion of multi-source remote sensing data, in which MS images and RGB images were obtained from multiple data sources (multispectral sensors and visible sensors) of UAV remote sensing. Following the application of an image fusion algorithm, a high-resolution HR_MS image was obtained, from which spectral indices (VIs_F) and texture features (GCFs_F) were derived [57]. The final model, MFM4, is a large field monitoring model framework that integrated multi-source remote sensing data and multi-source non-remote sensing data. On the basis of MFM3, effective cumulative temperature Ae and average rainfall Aar were added as model correction variables to enhance the interpretability and robustness of the model, and to improve the model’s performance in different application scenarios.

4.3.2. Forecast Results

The study used the root mean square error (MSE) and the coefficient of determination, R2, to evaluate the predictive effect of the model.
For machine learning modeling, 80% of the data were allocated as the training set, while the remaining 20% served as the test set. The holdout cross-validation method was used to divide the training and test sets, with the training set used to train the model parameters and the test set used to evaluate the accuracy and error of the model. Figures S1–S4 show the estimation results of the four modal inputs for the four growth physiological parameters of oil-seed rape, respectively. The plots show the true and predicted values of the test set in each model, as well as the coefficient of determination and mean square error of the evaluated models. Each plot represents the four model frameworks, MFM1, MFM2, MFM3, and MFM4, from the first row to the fourth row, and the first column to the fourth column represents the four machine learning models.
From Figures S1–S4, it can be seen that models with different frameworks have different accuracy in estimating the growth physiological indicators of rapeseed seedlings, indicating that the use of drone multispectral images combined with machine learning algorithms can indeed effectively estimate field crop phenotype information [14,15,16]. The monitoring model based on MFM4 has the highest accuracy and MFM1 has the lowest accuracy using only modal inputs from a single data source. The accuracy of the model frameworks for each type of modal input was ranked from largest to smallest as MFM4, MFM3, MFM2, and MFM1. There are several factors that lead to this situation. First, considering the addition of texture features and meteorological data, the increase in data type, such as the addition of meteorological data in MFM4 compared to MFM3, improves the practicability and generalization ability of the model. Second, due to multi-source remote sensing image fusion, the quality of remote sensing images is enhanced while the experience quality (QoE) is increased. The accuracy of the model established using high-resolution remote sensing images is improved and the root mean square deviation is reduced [21]. The results show that multi-source data can effectively enrich the complementary information from different sensors. The superposition model, MFM3, which belongs to the multi-modal data based on the image fusion algorithm, shows an advantage over MFM2 in terms of model accuracy and adaptability. This is in line with the conclusions of previous studies, and is important for establishing a reliable physiological monitoring model for oilseed rape growth [18,24,25].
In this study, a model for quantitative remote sensing monitoring of winter oilseed rape seedling growth status based on multimodal data is proposed, and four modeling frameworks with different inputs are compared. Quantitative analysis of the differences between the different modal model frameworks is the focus of this study. The results of the machine learning model evaluation based on the four multimodal model frameworks are shown in Table 5; the best values of mean square error and coefficient of determination for the monitoring models built with the four sets of modal inputs are highlighted to represent the best model for that modal input [58,59]. The best model accuracy and mean square error built with MFM1 were R2 = 0.5730 and MSE = 7.7398, while those with MFM2 were R2 = 0.6350 and MSE = 7.5148. The difference between the inputs of the two modalities was that the latter had an additional set of texture features, which is an improved method used by many agricultural remote sensing researchers, and the addition of texture features had a facilitative effect on building a more accurate model. The best model accuracy and mean square error of MFM3 was R2 = 0.7183 and MSE = 7.3309, which was a 13.1% improvement in prediction model accuracy and a 2.4% reduction in mean square error compared to MFM2. MFM3 was based on MFM2 to make changes to the source MS images used to extract spectral indices and texture features. Based on the RMGF image fusion algorithm used in this study, the high-resolution HR_MS images were obtained by making full use of the spectral features of MS and the spatial structure features of RGB; the complementary information of multi-source remote sensing data from two sensors was combined to reduce the inhibiting effect of a single information source. This formed a complete and consistent information description of the target. It is thus possible to draw the conclusion, consistent with the previous study, that feature fusion can solve problems such as data redundancy and multicollinearity, thus improving the accuracy of the model [60]. The best model accuracy and mean square error of MFM4 was R2 = 0.7454 and MSE = 6.6630, which improved the model accuracy by 3.7% and reduced the mean square error by 9.1% compared to MFM3. The modal input added multi-source non-remote sensing data, i.e., meteorological data obtained from different sensors, and the improvement in accuracy was smaller, but the mean square error of the model was significantly reduced. This proved that the addition of multi-source non-remote sensing data can improve the robustness of the model to enhance its ability to be generalized to different application scenarios and ensure its modeling effectiveness in other scenarios or scales [31,61].
The modeling effects of different machine learning models varied under different modal inputs, as well as for different physiological indicators of oil-seed rape growth. In Table 5, the BPNN algorithm obtained the highest coefficient of determination and the NMR algorithm obtained the lowest mean square error for the modal input of MFM1; the BPNN obtained the best model parameter values for the modal input of MFM2; and the SVR obtained the best model parameter values for both MFM3 and MFM4. When the data sources were small, NMR showed the ability to provide a solution in low-dimensional space; however, when the data sources were large, NMR had too many linear and nonlinear terms, which may generate singular matrices during the resolution process and cause overfitting of the model [62]. The PLSR did not obtain the best model parameters, mainly because PLSR is a fitting process based on the component contributions, which may be better in solving the problem of multiple covariance [63]. However, in this study, the data with high correlation were dimensionally reduced before modeling, and the data had weak multicollinearity and the modeling effect of PLSR was poor. The performance of the models also varied for different physiological indicators of seedling rape growth. For instance, the SVR algorithm consistently achieved the highest accuracy in the prediction of LAI, while NMR obtained the majority of optimal estimates in the prediction of AGB.

5. Conclusions

This study focused on the early-stage growth of oil-seed rape and utilized data from unmanned aerial vehicle (UAV) visible-light images, multispectral images, and four crucial growth physiological indicators measured in real time. Four MFMs with different modal inputs were proposed for physiological monitoring of rapeseed growth during the early stage, and four machine learning models, SVR, PLS, BPNN, and NMR, were used to compare the differences between multi-source data and single-source data in monitoring of oil-seed rape using remote sensing. The results demonstrate that the models, which incorporate VIs and GCFs extracted from an image fusion algorithm, as well as effective accumulated temperature (Ae) and average rainfall (Aar) as correction variables, effectively leveraged complementary information from various UAV remote sensing data sources. This mitigated the inhibitory effect of single-source data on the models, which were able to accurately detect rapeseed growth during the early stage. Among these, the SVR model based on multi-source remote sensing data and multi-source non-remote sensing data exhibited high accuracy with minimal error, showcasing robustness and generalizability in various scenarios. This research provides a theoretical basis for precise field management and agricultural production.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/rs15163951/s1, Figure S1: Prediction results of LAI for four multimodal models; Figure S2: Prediction results of AGB for four multimodal models; Figure S3: Prediction results of LNC for four multimodal models; Figure S4: Prediction results of SPAD for four multimodal models.

Author Contributions

Conceptualization, G.Z., Y.Y. and Y.R.; methodology, J.Z.; software, J.W. (Jiang Wang); validation, J.W. (Jiang Wang); formal analysis, X.W.; investigation, Y.Y.; resources, J.W. (Jian Wang) and Z.J.; data curation, J.Z.; writing—original draft preparation, X.W.; writing—review and editing, X.W.; visualization, J.W. (Jiang Wang); supervision, G.Z. and Y.Y.; project administration, G.Z. and Y.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jin, X.; Yang, W.; Doonan, J.H.; Atzberger, C. Crop phenotyping studies with application to crop monitoring. Crop J. 2022, 10, 1221–1223. [Google Scholar] [CrossRef]
  2. Huang, J.; Sedano, F.; Huang, Y.; Ma, H.; Li, X.; Liang, S.; Tian, L.; Zhang, X.; Fan, J.; Wu, W. Assimilating a synthetic Kalman filter leaf area index series into the WOFOST model to improve regional winter wheat yield estimation. Agric. Forest Meteorol. 2016, 216, 188–202. [Google Scholar] [CrossRef]
  3. Dobermann, A.; Pampolino, M.F. Indirect leaf area index measurement as a tool for characterizing rice growth at the field scale. Commun. Soil Sci. Plant Anal. 1995, 26, 1507–1523. [Google Scholar] [CrossRef]
  4. Wang, J.-J.; Li, Z.; Jin, X.; Liang, G.; Struik, P.C.; Gu, J.; Zhou, Y. Phenotyping flag leaf nitrogen content in rice using a three-band spectral index. Comput. Electron. Agric. 2019, 162, 475–481. [Google Scholar] [CrossRef]
  5. Zhao, C.; Zhang, Y.; Du, J.; Guo, X.; Wen, W.; Gu, S.; Wang, J.; Fan, J. Crop Phenomics: Current Status and Perspectives. Front. Plant Sci. 2019, 10, 714. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Hussain, S.; Gao, K.; Din, M.; Gao, Y.; Shi, Z.; Wang, S. Assessment of UAV-Onboard Multispectral Sensor for non-destructive site-specific rapeseed crop phenotype variable at different phenological stages and resolutions. Remote Sens. 2020, 12, 397. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, T.; Liu, Y.; Wang, M.; Fan, Q.; Tian, H.; Qiao, X.; Li, Y. Applications of UAS in crop biomass monitoring: A review. Front. Plant Sci. 2021, 12, 616689. [Google Scholar] [CrossRef]
  8. Aasen, H.; Bolten, A. Multi-temporal high-resolution imaging spectroscopy with hyperspectral 2D imagers–From theory to application. Remote Sens. Environ. 2018, 205, 374–389. [Google Scholar] [CrossRef]
  9. Bhadra, S.; Sagan, V.; Maimaitijiang, M.; Maimaitiyiming, M.; Newcomb, M.; Shakoor, N.; Mockler, T.C. Quantifying leaf chlorophyll concentration of sorghum from hyperspectral data using derivative calculus and machine learning. Remote Sens. 2020, 12, 2082. [Google Scholar] [CrossRef]
  10. Padalia, H.; Sinha, S.K.; Bhave, V.; Trivedi, N.K.; Kumar, A.S. Estimating canopy LAI and chlorophyll of tropical forest plantation (North India) using Sentinel-2 data. Adv. Space Res. 2020, 65, 458–469. [Google Scholar] [CrossRef]
  11. Tanabe, R.; Matsui, T.; Tanaka, T.S. Winter wheat yield prediction using convolutional neural networks and UAV-based multispectral imagery. Field Crops Res. 2023, 291, 108786. [Google Scholar] [CrossRef]
  12. Yang, G.; Liu, J.; Zhao, C.; Li, Z.; Huang, Y.; Yu, H.; Xu, B.; Yang, X.; Zhu, D.; Zhang, X. Unmanned aerial vehicle remote sensing for field-based crop phenotyping: Current status and perspectives. Front. Plant Sci. 2017, 8, 1111. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Xie, C.; Yang, C. A review on plant high-throughput phenotyping traits using UAV-based sensors. Comput. Electron. Agric. 2020, 178, 105731. [Google Scholar] [CrossRef]
  14. Deng, L.; Mao, Z.; Li, X.; Hu, Z.; Duan, F.; Yan, Y. UAV-based multispectral remote sensing for precision agriculture: A comparison between different cameras. ISPRS J. Photogramm. Remote Sens. 2018, 146, 124–136. [Google Scholar] [CrossRef]
  15. Johansen, K.; Morton, M.J.; Malbeteau, Y.; Aragon, B.; Al-Mashharawi, S.; Ziliani, M.G.; Angel, Y.; Fiene, G.; Negrão, S.; Mousa, M.A. Predicting biomass and yield in a tomato phenotyping experiment using UAV imagery and random forest. Front. Artif. Intel. 2020, 3, 28. [Google Scholar] [CrossRef] [PubMed]
  16. Lee, H.; Wang, J.; Leblon, B. Intra-field canopy nitrogen retrieval from unmanned aerial vehicle imagery for wheat and corn fields. Can. J. Remote Sens. 2020, 46, 454–472. [Google Scholar] [CrossRef]
  17. Gilabert, M.; Moreno, A.; Maselli, F.; Martínez, B.; Chiesi, M.; Sánchez-Ruiz, S.; García-Haro, F.; Pérez-Hoyos, A.; Campos-Taberner, M.; Pérez-Priego, O. Daily GPP estimates in Mediterranean ecosystems by combining remote sensing and meteorological data. ISPRS J. Photogramm. Remote Sens. 2015, 102, 184–197. [Google Scholar] [CrossRef]
  18. Sun, C.; Bian, Y.; Zhou, T.; Pan, J. Using of multi-source and multi-temporal remote sensing data improves crop-type mapping in the subtropical agriculture region. Sensors 2019, 19, 2401. [Google Scholar] [CrossRef] [Green Version]
  19. Zhang, Z.; An, Y. Ocean application conception of sky, earth, and sea multi base collaborative multi source fusion. Satell. Appl. 2019, 2, 24–29. [Google Scholar]
  20. Pawłowski, M.; Wróblewska, A.; Sysko-Romańczuk, S. Effective Techniques for Multimodal Data Fusion: A Comparative Analysis. Sensors 2023, 23, 2381. [Google Scholar] [CrossRef]
  21. Zhai, G.; Min, X. Perceptual image quality assessment: A survey. Sci. China Inf. Sci. 2020, 63, 211301. [Google Scholar] [CrossRef]
  22. Liu, Y.S.; Jiang, M.Y.; Liao, C.Z. In Multifocus Image Fusion Based on Multiresolution Transform and Particle Swarm Optimization. Adv. Mater. Res. 2013, 756, 3281–3285. [Google Scholar] [CrossRef] [Green Version]
  23. Lu, J.; Eitel, J.U.; Engels, M.; Zhu, J.; Ma, Y.; Liao, F.; Zheng, H.; Wang, X.; Yao, X.; Cheng, T. Improving Unmanned Aerial Vehicle (UAV) remote sensing of rice plant potassium accumulation by fusing spectral and textural information. Int. J. Appl. Earth OBS 2021, 104, 102592. [Google Scholar] [CrossRef]
  24. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Hartling, S.; Esposito, F.; Fritschi, F.B. Soybean yield prediction from UAV using multimodal data fusion and deep learning. Remote Sens. Environ. 2020, 237, 111599. [Google Scholar] [CrossRef]
  25. Fei, S.; Hassan, M.A.; Xiao, Y.; Su, X.; Chen, Z.; Cheng, Q.; Duan, F.; Chen, R.; Ma, Y. UAV-based multi-sensor data fusion and machine learning algorithm for yield prediction in wheat. Precis. Agric. 2023, 24, 187–212. [Google Scholar] [CrossRef]
  26. Min, X.; Zhai, G.; Zhou, J.; Farias, M.C.; Bovik, A.C. Study of Subjective and Objective Quality Assessment of Audio-Visual Signals. IEEE Trans. Image Process. 2020, 29, 6054–6068. [Google Scholar] [CrossRef]
  27. Min, X.; Zhai, G.; Zhou, J.; Zhang, X.P.; Yang, X.; Guan, X. A Multimodal Saliency Model for Videos with High Audio-Visual Correspondence. IEEE Trans. Image Process. 2020, 29, 3805–3819. [Google Scholar] [CrossRef]
  28. Zheng, H.; Cheng, T.; Zhou, M.; Li, D.; Yao, X.; Tian, Y.; Cao, W.; Zhu, Y. Improved estimation of rice aboveground biomass combining textural and spectral analysis of UAV imagery. Precis. Agric. 2019, 20, 611–629. [Google Scholar] [CrossRef]
  29. Jin, X.; Jiang, Q.; Yao, S.; Zhou, D.; Nie, R.; Lee, S.-J.; He, K. Infrared and visual image fusion method based on discrete cosine transform and local spatial frequency in discrete stationary wavelet transform domain. Infrared Phys. Technol. 2018, 88, 1–12. [Google Scholar] [CrossRef]
  30. Tan, W.; Xiang, P.; Zhang, J.; Zhou, H.; Qin, H. Remote sensing image fusion via boundary measured dual-channel PCNN in multi-scale morphological gradient domain. IEEE Access 2020, 8, 42540–42549. [Google Scholar] [CrossRef]
  31. Torgbor, B.A.; Rahman, M.M.; Brinkhoff, J.; Sinha, P.; Robson, A. Integrating Remote Sensing and Weather Variables for Mango Yield Prediction Using a Machine Learning Approach. Remote Sens. 2023, 15, 3075. [Google Scholar] [CrossRef]
  32. Thenkabail, P.S.; Biradar, C.M.; Noojipady, P.; Dheeravath, V.; Li, Y.; Velpuri, M.; Gumma, M.; Gangalakunta, O.R.P.; Turral, H.; Cai, X. Global irrigated area map (GIAM), derived from remote sensing, for the end of the last millennium. Int. J. Remote Sens. 2009, 30, 3679–3733. [Google Scholar] [CrossRef]
  33. Zhou, J.; Zhou, J.; Ye, H.; Ali, M.L.; Chen, P.; Nguyen, H.T. Yield estimation of soybean breeding lines under drought stress using unmanned aerial vehicle-based imagery and convolutional neural network. Biosyst. Eng. 2021, 204, 90–103. [Google Scholar] [CrossRef]
  34. Yang, Q.; Shi, L.; Han, J.; Zha, Y.; Zhu, P. Deep convolutional neural networks for rice grain yield estimation at the ripening stage using UAV-based remotely sensed images. Field Crops Res. 2019, 235, 142–153. [Google Scholar] [CrossRef]
  35. Cui, Z.; Kerekes, J.P. Potential of Red Edge Spectral Bands in Future Landsat Satellites on Agroecosystem Canopy Green Leaf Area Index Retrieval. Remote Sens. 2018, 10, 1458. [Google Scholar] [CrossRef] [Green Version]
  36. Cui, Z.; Kerekes, J.P. Impact of Wavelength Shift in Relative Spectral Response at High Angles of Incidence in Landsat-8 Operational Land Imager and Future Landsat Design Concepts. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5873–5883. [Google Scholar] [CrossRef]
  37. Min, X.; Zhai, G.; Gu, K.; Yang, X.; Guan, X. Objective Quality Evaluation of Dehazed Images. IEEE Trans. Intell. Transp. Syst. 2019, 20, 2879–2892. [Google Scholar] [CrossRef]
  38. Lukas, V.; Huňady, I.; Kintl, A.; Mezera, J.; Hammerschmiedt, T.; Sobotková, J.; Brtnický, M.; Elbl, J. Using UAV to Identify the Optimal Vegetation Index for Yield Prediction of Oil Seed Rape (Brassica napus L.) at the Flowering Stage. Remote Sens. 2022, 14, 4953. [Google Scholar] [CrossRef]
  39. Rouse, J.; Haas, R.; Schell, J.; Deeng, R.; Harlan, J. Monitoring the vernal advancement of retrogradation (greenwave effect) of natural vegetation. In Type III Final Report RSC 1978-4; Remote Sensing Center, Texas A&M University: College Station, TX, USA, 1974; pp. 1–93. [Google Scholar]
  40. Schleicher, T.D.; Bausch, W.C.; Delgado, J.A.; Ayers, P.D. Evaluation and Refinement of the Nitrogen Reflectance Index (NRI) for Site-Specific Fertilizer Management; 2001 ASAE Annual Meeting; American Society of Agricultural and Biological Engineers: St. Joseph, MI, USA, 1998; Volume 1. [Google Scholar]
  41. Gitelson, A.A.; Zur, Y.; Chivkunova, O.B.; Merzlyak, M.N. Assessing carotenoid content in plant leaves with reflectance spectroscopy. Photochem. Photobiol. 2002, 75, 272–281. [Google Scholar] [CrossRef]
  42. Gitelson, A.A.; Viña, A.; Arkebauer, T.J.; Rundquist, D.C.; Keydan, G.; Leavitt, B. Remote estimation of leaf area index and green leaf biomass in maize canopies. Geophys. Res. Lett. 2003, 30, 52. [Google Scholar] [CrossRef] [Green Version]
  43. Baret, F.; Guyot, G. Potentials and limits of vegetation indices for LAI and APAR assessment. Remote Sens. Env. 1991, 35, 161–173. [Google Scholar] [CrossRef]
  44. Goel, N.S.; Qin, W. Influences of canopy architecture on relationships between various vegetation indices and LAI and FPAR: A computer simulation. Remote Sens. Rev. 1994, 10, 309–347. [Google Scholar] [CrossRef]
  45. Chen, J.M. Evaluation of vegetation indices and a modified simple ratio for boreal applications. Can. J. Remote Sens. 1996, 22, 229–242. [Google Scholar] [CrossRef]
  46. de la Iglesia Martinez, A.; Labib, S. Demystifying normalized difference vegetation index (NDVI) for greenness exposure assessments and policy interventions in urban greening. Env. Res. 2023, 220, 115155. [Google Scholar] [CrossRef]
  47. Zhao, F.; Yang, G.; Yang, H.; Long, H.; Xu, W.; Zhu, Y.; Meng, Y.; Han, S.; Liu, M. A Method for Prediction of Winter Wheat Maturity Date Based on MODIS Time Series and Accumulated Temperature. Agriculture 2022, 12, 945. [Google Scholar] [CrossRef]
  48. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  49. May, J.O.; Looney, S.W. Sample size charts for Spearman and Kendall coefficients. J. Biom. Biostat. 2020, 11, 1–7. [Google Scholar]
  50. Fırat, H.; Asker, M.E.; Hanbay, D. Classification of hyperspectral remote sensing images using different dimension reduction methods with 3D/2D CNN. Remote Sens. Appl. 2022, 25, 100694. [Google Scholar] [CrossRef]
  51. Lapajne, J.; Knapič, M.; Žibrat, U. Comparison of Selected Dimensionality Reduction Methods for Detection of Root-Knot Nematode Infestations in Potato Tubers Using Hyperspectral Imaging. Sensors 2022, 22, 367. [Google Scholar] [CrossRef]
  52. Jiang, Y.; Wei, H.; Hou, S.; Yin, X.; Wei, S.; Jiang, D. Estimation of Maize Yield and Protein Content under Different Density and N Rate Conditions Based on UAV Multi-Spectral Images. Agronomy 2023, 13, 421. [Google Scholar] [CrossRef]
  53. de Oliveira, R.P.; Rodrigues, B.J.M.; Alves, P.A.; Pereira, O.J.L.; Cristiano, Z.; Angeli, F.C.E. Predicting Sugarcane Biometric Parameters by UAV Multispectral Images and Machine Learning. Agronomy 2022, 12, 1992. [Google Scholar] [CrossRef]
  54. Mohidem, N.A.; Jaafar, S.; Rosle, R.; Che’Ya, N.N.; Arif, S.J.; Fazlil, I.W.F.; Ismail, M.R. Application of multispectral UAV for paddy growth monitoring in Jitra, Kedah, Malaysia. IOP Conf. Ser. Earth Environ. Sci. 2022, 1038, 012053. [Google Scholar] [CrossRef]
  55. Zhang, X.; Zhang, K.; Sun, Y.; Zhao, Y.; Zhuang, H.; Ban, W.; Chen, Y.; Fu, E.; Chen, S.; Liu, J.; et al. Combining Spectral and Texture Features of UAS-Based Multispectral Images for Maize Leaf Area Index Estimation. Remote Sens. 2022, 14, 331. [Google Scholar] [CrossRef]
  56. Zheng, H.; Ma, J.; Zhou, M.; Li, D.; Yao, X.; Cao, W.; Zhu, Y.; Cheng, T. Enhancing the Nitrogen Signals of Rice Canopies across Critical Growth Stages through the Integration of Textural and Spectral Information from Unmanned Aerial Vehicle (UAV) Multispectral Imagery. Remote Sens. 2020, 12, 957. [Google Scholar] [CrossRef] [Green Version]
  57. Lu, L.; Fengqiao, W.; Cheolkon, J. LRINet: Long-range imaging using multispectral fusion of RGB and NIR images. Inf. Fusion 2023, 92, 177–189. [Google Scholar] [CrossRef]
  58. Zhou, C.; Gong, Y.; Fang, S.; Yang, K.; Peng, Y.; Wu, X.; Zhu, R. Combining spectral and wavelet texture features for unmanned aerial vehicles remote estimation of rice leaf area index. Front. Plant Sci. 2022, 13, 957870. [Google Scholar] [CrossRef]
  59. Usha, S.G.A.; Vasuki, S. Significance of texture features in the segmentation of remotely sensed images. Optik 2022, 249, 168241. [Google Scholar] [CrossRef]
  60. Saini, P.; Kumar, A. Effect of Fusion of Statistical and Texture Features on HSI based Leaf Images with Both Dorsal and Ventral Sides. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 305–312. [Google Scholar] [CrossRef] [Green Version]
  61. Islam, M.D.; Di, L.; Qamer, F.M.; Shrestha, S.; Guo, L.; Lin, L.; Mayer, T.J.; Phalke, A.R. Rapid Rice Yield Estimation Using Integrated Remote Sensing and Meteorological Data and Machine Learning. Remote Sens. 2023, 15, 2374. [Google Scholar] [CrossRef]
  62. Aswed, G.K.; Ahmed, M.N.; Mohammed, H.A. Predicting initial duration of project using linear and nonlinear regression models. Int. J. Adv. Technol. Eng. Explor. 2022, 9, 1730. [Google Scholar]
  63. Fu, T.T.; Sieng, Y.W. A comparative study between PCR, PLSR, and LW-PLS on the predictive performance at different data splitting ratios. Chem. Eng. Commun. 2022, 209, 1439–1456. [Google Scholar]
Figure 1. Field diagram of the test site and UAV: (a) location of the research field in Hubei Province; (b) aerial view of UAV; (c) location of sampling plots.
Figure 1. Field diagram of the test site and UAV: (a) location of the research field in Hubei Province; (b) aerial view of UAV; (c) location of sampling plots.
Remotesensing 15 03951 g001
Figure 2. Distribution of rape experimental area and treatment. N8, N12, N16 represent three N application levels; D1, D3, D5 represent three density treatments; S925, S1010, S1025 represent three sowing periods.
Figure 2. Distribution of rape experimental area and treatment. N8, N12, N16 represent three N application levels; D1, D3, D5 represent three density treatments; S925, S1010, S1025 represent three sowing periods.
Remotesensing 15 03951 g002
Figure 3. LAI, SPAD, PNC, and AGB and weather data collection at the rapeseed seedling stage.
Figure 3. LAI, SPAD, PNC, and AGB and weather data collection at the rapeseed seedling stage.
Remotesensing 15 03951 g003
Figure 4. UAV remote sensing device and flight control system: (a) DJI RC Enterprise, UAV, and flight controller; (b) multispectral and visible light sensor of UAV; (c) route planning (2D and 3D schematic).
Figure 4. UAV remote sensing device and flight control system: (a) DJI RC Enterprise, UAV, and flight controller; (b) multispectral and visible light sensor of UAV; (c) route planning (2D and 3D schematic).
Remotesensing 15 03951 g004
Figure 5. Acquisition process of a texture parameter based on GCFs.
Figure 5. Acquisition process of a texture parameter based on GCFs.
Remotesensing 15 03951 g005
Figure 6. Multi-scale decomposition of GF.
Figure 6. Multi-scale decomposition of GF.
Remotesensing 15 03951 g006
Figure 7. Images based on GF decomposition.
Figure 7. Images based on GF decomposition.
Remotesensing 15 03951 g007
Figure 8. Scatter matrix of multi-source data variables of the model.
Figure 8. Scatter matrix of multi-source data variables of the model.
Remotesensing 15 03951 g008
Figure 9. Spearman correlation analysis.
Figure 9. Spearman correlation analysis.
Remotesensing 15 03951 g009
Figure 10. PCA dimensionality reduction result: (a) cumulative contribution of variables; (b) data distribution after dimensionality reduction.
Figure 10. PCA dimensionality reduction result: (a) cumulative contribution of variables; (b) data distribution after dimensionality reduction.
Remotesensing 15 03951 g010
Table 1. Sensor parameters of the UAV.
Table 1. Sensor parameters of the UAV.
Sensor CategorySpectral Area (μm)ResolutionField of View (H° × V°)
Visible lightN/A1600 × 120056° × 84°
MultispectralGreen: 0.560; Red: 0.650; Red edge: 0.730; NIR: 0.860800 × 60047.2° × 73.9°
Table 2. Acquisition of remote sensing data.
Table 2. Acquisition of remote sensing data.
Test TimeDate of Remote Sensing Image AcquisitionPrecise TimeHeight (m)Heading/Sideways Overlap
19 November 2022–21 November 202221 November 202211:30–12:004075/70%
8 December 2022–10 December 20228 December 202212:00–12:304075/70%
9 January 2023–11 January 202310 January 202312:00–12:304075/70%
29 January 2023–31 January 202330 January 202313:30–14:004075/70%
Note: The acquisition of remote sensing data and field trial data was completed simultaneously during the trial time.
Table 3. Calculation of spectral index.
Table 3. Calculation of spectral index.
Spectral IndexAbbreviationsCalculation FormulaSource
Normalized vegetation indexNDVINDVI = (NIR − R)/(NIR + R)[39]
Nitrogen reflection indexNRINRI = (G − R)/(G + R)[40]
Greenness vegetation indexGNDVIGNDVI = (NIR − G)/(NIR + G)[41,42]
Ratio vegetation indexRVIRVI = NIR/R[43]
Non-linear vegetation indexNLI(NIR × NIR − R)/(R × R − NIR)[44]
Modified simple ratio indexMSR(NIR/R − 1)/[(NIR/R + 1)^(1/2)][45]
Table 4. Four multimodal data model frameworks.
Table 4. Four multimodal data model frameworks.
Model FrameworkData SourceModel Input
MFM1Original MS imagesSpectral index VIs (NDVI; NRI; MSR)
MFM2Original MS imagesSpectral index VIs (NDVI; NRI; MSR)
Texture feature GCFs (Vcv; Scv; Ra)
MFM3HR_MS image after image fusionSpectral indices VIs_F (NDVI_F; NRI_F; MSR_F)
Texture Features GCFs_F (Vcv_F; Scv_F; Ra_F)
MFM4HR_MS image after image fusion.
Meteorological Data
Spectral indices VIs_F (NDVI_F; NRI_F; MSR_F)
Texture Features GCFs_F (Vcv_F; Scv_F; Ra_F)
Meteorological data MDs (Ae; Aar)
Table 5. Machine learning models for four multimodal model frameworks to evaluate physiological indicators of rapeseed growth.
Table 5. Machine learning models for four multimodal model frameworks to evaluate physiological indicators of rapeseed growth.
Modal InputMachine Learning ModelsEvaluation IndicatorsPhysiological Indicators of Oil-Seed Rape GrowthAverage
LAIAGBLNCSPAD
MFM1SVRR-square0.67730.38640.47290.72370.5651
MSE0.206832.16816.025615.286213.4217
PLSRR-square0.45280.33910.19710.21830.3018
MSE0.404228.787411.683141.941920.7042
BPNNR-square0.42330.32150.53040.82770.5257
MSE0.364915.27484.16549.15417.7398
NMRR-square0.36540.44820.70430.77410.5730
MSE0.426216.50123.842812.37048.2852
MFM2SVRR-square0.70290.46970.48290.77840.6085
MSE0.224920.84867.267612.085710.1067
PLSRR-square0.52120.34060.28460.45860.4013
MSE0.360139.90557.896129.394419.3890
BPNNR-square0.62530.44770.578750.88810.6350
MSE0.264117.92294.9966.87627.5148
NMRR-square0.42810.50190.62390.80790.5905
MSE0.373138.68855.72528.005813.1982
MFM3SVRR-square0.78020.59090.63710.86510.7183
MSE0.147117.83313.72567.61787.3309
PLSRR-square0.62980.45140.41450.59750.5233
MSE0.291929.21225.824819.795913.7812
BPNNR-square0.65050.47030.59910.89350.6534
MSE0.296227.19464.45565.20769.2885
NMRR-square0.44660.59180.70270.82020.6403
MSE0.385317.63654.1299.41117.8155
MFM4SVRR-square0.80710.63560.66460.87420.7454 *
MSE0.141117.43723.37155.57186.6630 *
PLSRR-square0.59730.49030.44940.65260.5474
MSE0.325118.72335.718115.775610.1355
BPNNR-square0.77020.44380.63510.88520.6836
MSE0.122233.06066.15425.617711.2387
NMRR-square0.60450.56020.69150.82660.6707
MSE0.253924.18863.57738.51089.1327
Note: * represents the best results of this model framework for estimating physiological parameters of seedling oil-seed rape growth. The bold represents the best value of mean squared error and coefficient of determination, and also represents the best model of the modal input.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Wei, X.; Wang, J.; Zhou, G.; Wang, J.; Jiang, Z.; Zhao, J.; Ren, Y. Prediction of Seedling Oilseed Rape Crop Phenotype by Drone-Derived Multimodal Data. Remote Sens. 2023, 15, 3951. https://doi.org/10.3390/rs15163951

AMA Style

Yang Y, Wei X, Wang J, Zhou G, Wang J, Jiang Z, Zhao J, Ren Y. Prediction of Seedling Oilseed Rape Crop Phenotype by Drone-Derived Multimodal Data. Remote Sensing. 2023; 15(16):3951. https://doi.org/10.3390/rs15163951

Chicago/Turabian Style

Yang, Yang, Xinbei Wei, Jiang Wang, Guangsheng Zhou, Jian Wang, Zitong Jiang, Jie Zhao, and Yilin Ren. 2023. "Prediction of Seedling Oilseed Rape Crop Phenotype by Drone-Derived Multimodal Data" Remote Sensing 15, no. 16: 3951. https://doi.org/10.3390/rs15163951

APA Style

Yang, Y., Wei, X., Wang, J., Zhou, G., Wang, J., Jiang, Z., Zhao, J., & Ren, Y. (2023). Prediction of Seedling Oilseed Rape Crop Phenotype by Drone-Derived Multimodal Data. Remote Sensing, 15(16), 3951. https://doi.org/10.3390/rs15163951

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop