1. Introduction
Farmland plays an important role in the material basis of human survival and development. With the development of precision agriculture, the determination of farmland types and the types of crops grown has become a basic task [
1]. The Large Area Crop Inventory Experiment (LACIE) of the United States and the Monitoring Agriculture with Remote Sensing (MARS) program of the European Union both include the task of remote sensing surveying of farmland [
2]. In 2003, the Russian Ministry of Agriculture also established a national agricultural monitoring system [
3]. This system acquires information on farmland area, maps of farmland use status, and crop growth status. These systems are based on the intra-annual variation process of moderate resolution imaging spectroradiometer (MODIS) vegetation indices for the estimation and analysis of crop and farmland areas. In the African region, Egypt used multi-temporal MODIS data and time series analysis methods to analyze the satellite images of irrigated regions at different stages [
4]. This program has achieved the goals of farmland area surveying and monitoring of dynamic changes throughout Egypt. The rice–wheat rotation field is one of the main farming methods in China, with 4.8 million hectares of land under cultivation [
5]. However, Chinese farmland areas are relatively small and fragmented [
6]. In the study of the relationship between farmland fragmentation and agricultural production cost, many scholars use the average plot area, the number of plots, and the average plot distance to quantitatively reflect farmland fragmentation [
7,
8,
9]. According to statistics, in China, the average arable land area of each household is only 0.58 hm
2, the number of plots in each household is as high as 5.34, the average plot area is only 0.11 hm
2, and the average distance from home is 1.06 km [
10]. Thus, monitoring farming progress in rice–wheat rotation fields is an essential part of ensuring grain production. At present, this monitoring is generally conducted by manual field surveys that are time-consuming and inefficient. At the same time, inadequate monitoring of the farming conditions during the wheat harvesting stage could directly affect wheat harvesting and rice planting, resulting in a decrease in farmland utilization.
Remote sensing technology has been widely used for crop growth monitoring due to its advantages of high efficiency and accuracy [
11,
12,
13]. Based on the sensing distance, the most commonly used remote sensing platforms can be classified into satellite remote sensing platforms, unmanned aerial vehicle (UAV) remote sensing platforms, and ground platforms [
14]. Compared with the latter two, satellite remote sensing platforms can cover a larger area and have been used to assess various crop growth parameters such as leaf area index [
15], cover [
16], biomass [
17], chlorophyll, nitrogen [
18,
19], and yield [
20]. However, satellite remote sensing platforms produce only low-resolution images, have a long access cycle, and are affected by the weather [
21]. These drawbacks make them unsuitable for farming progress monitoring of small-scale rice–wheat rotation farmlands. In the ground platforms, various mobile vehicles or portable instruments are exploited to obtain information about farmland or crops. Although the ground platforms can provide images with high resolution, their relatively low working efficiency cannot meet the requirements of fast and efficient monitoring of farming progress.
In recent years, UAV remote sensing technology has developed rapidly, which is making up, gradually, for the shortcomings of satellite remote sensing platforms and ground remote sensing platforms. UAV remote sensing platforms include fixed-wing and multirotor configurations. Fixed-wing UAVs have the advantages of fast flight, high flight efficiency, long endurance time, large payload capacity, and a high flight altitude [
22,
23]. However, they have certain requirements for taking off and landing, they cannot hover, and they cause blurred images because of their high-speed shooting [
24]. Multirotor UAVs have a simple structure, the ability to hover, and modest requirements for taking off and landing, but they possess a slow flight speed, short endurance time, a low flight altitude, and small payload volume [
24]. UAV remote sensing platforms can acquire high-resolution crop canopy images. Moreover, they can carry different image sensors for assessing various crop growth parameters, including biomass [
25], yield [
26], LAI [
27], chlorophyll [
28], and nitrogen [
29]. Makanza et al. [
30] obtained visible light images by UAV and successfully achieved the estimation of canopy cover in maize fields. Roth [
31] et al. realized the estimation of soybean LAI based on UAV RGB images, and the R
2 reached 0.92, laying a foundation for automatic LAI evaluation. Their results provide a theoretical basis for crop growth monitoring.
The above-mentioned studies demonstrated the potential of UAV remote sensing platforms for crop growth monitoring in farmland. It has been shown that crop classification using the UAV images analysis technique is an effective tool [
32]. Wilke et al. [
33] classified wheat with different lodging degrees by using a UAV-based canopy height model, the classification accuracy of R
2 reached 0.96. Wang et al. [
34] extracted vegetation information by calculating various vegetation indices. The most accurate index was found to be VDVI (the visible-band difference vegetation index) that could reach an accuracy of more than 90%; however, they didn’t classify different field types. The above-mentioned studies mostly used single extraction methods such as color, texture features, or color indices; however, there are very few studies in the literature using multi-feature fusion and different extraction methods for classification and comparison.
Furthermore, UAV remote sensing images contain high-throughput information that could be used to characterize changes in farming progress, but most investigations only used the high-throughput information to monitor crop growth, and few studies were oriented toward farmland research to monitor farming progress during harvest. Roth et al. [
35] successfully predicted yield and protein content through the dynamics of soybean vegetative growth based on hyperspectral images. They provided strong evidence for UAV remote sensing in crop physiological analysis. However, their research focused on the crop itself, with little mention of monitoring soybean fields to guide farm operations. Hassan et al. [
36] realized the rapid monitoring of NDVI in the wheat growth cycle by using the multi-spectral UAV platform, which can realize the prediction of grain yield. Their research focuses on the flowering stage and filling stage of wheat growth stage, but the growth and field monitoring of wheat harvest stage are seldom mentioned. Field monitoring during the wheat harvest stage is important for guiding farmland operation of this stage. This indicates the importance of conducting researches on farming progress monitoring in rice–wheat rotation fields based on UAV remote sensing platforms during the wheat harvest stage.
Therefore, the objective of this study was to explore the optimal classification method for crop progress classification of rice–wheat rotation fields at the harvest stage. This classification method can effectively improve rice–wheat rotation farmland monitoring practices for farmers and provide a new idea for the visible image classification based on UAV. In this paper, we present a method to effectively extract information on farming progress using UAV RGB images and the RM model. The remainder of the paper is organized as follows.
Section 2 describes the proposed methods and the corresponding materials being used.
Section 3 describes the performance of the proposed method as well as the comparison with traditional machine learning methods.
Section 4 concludes this paper and provides further discussion.
2. Materials and Methods
2.1. Image Acquisition and Preprocessing
We used DJI Inspire 2 UAV (
Figure 1) as the platform to collect RGB images. The main technical parameters of the platform are as follows: hover accuracy: vertical: ±0.5 m, horizontal: ±1.5 m. The UAV can fly for 23 min with a pair of batteries and can hover and fly stably indoors and outdoors. The UAV was equipped with a Zenmuse X7 lens. We collected the UAV RGB images from the experimental field of Yangzhou University on 28 May 2020. The weather on the measurement day was sunny, and the UAV flew for a short time from 11:00 to 13:00 to ensure sufficient and stable sunlight. The UAV is battery powered with autonomy for about two hours. The shooting angle was perpendicular to the ground, and the flight height was about 20 m, as shown in
Figure 2. The test field is located in Wenhui Road Campus of Yangzhou University (32.39° N and 119.42° E, in 2020), with a total area of 3.76 hectares. The type of soil in this area is loamy soil. The main crops in the tested fields in May and June were wheat (cultivars: Yangmai-16 and Yangmai-23) and rice (cultivars: Suxiu-867 and Naneng-9108).
We used the image stitching method and obtained RGB images to synthesize the whole test field orthophoto. Following the processing procedure of Lu et al. [
37]; image alignment, georeferencing, mosaicking, dense point cloud construction, orthoimage generation, and calibration were conducted using PhotoScan Professional software (Agisoft LLC, St. Petersburg, Russia) and ENVI/IDL software (Harris Geospatial Solutions, Inc., Broomfield, CO, USA) based on the GPS location and camera internal parameters from each aerial image. Due to a large amount of data, to avoid the influence of blocking factors such as buildings on the classification of farming progress monitoring, only a part of the area located in the mid-western part of the flight area was selected as the region of interest in this study. As shown in
Figure 2, the region of interest should include the type of farming progressing from the wheat harvest stage to the rice planting stage. The region of interest we selected here contained 16611 × 5962 pixels, and the corresponding grayscale values (red, green, blue) were stored in TIFF format, where each value consists of 8 bytes.
2.2. Color Index Construction and Selection
We calculated and analyzed the information from the RGB images and features of the color index to identify the key feature parameters that can be used to distinguish different farming progress types. Through visual interpretation and field investigation, samples were randomly selected from four agricultural progress types, as shown in
Figure 3. As seen in the images, unharvested wheat fields (
Figure 3a) are golden yellow, harvested wheat fields (
Figure 3b) are light yellow due to wheat stubble, tilled wheat fields (
Figure 3c) are grayish brown, and irrigated fields (
Figure 3d) are gray. The color differences in these areas result from the differences in the phenotypic materials and the structures of the fields.
Different types of farming progress in rice–wheat rotation fields clearly appear in the visible light band. After all essential image processing, digital number (DN) values of
R,
G, and
B channels for each plot were extracted from each image using the region of interest (ROI) tool in ENVI/IDL. The DN values of RGB channels in remotely sensed RGB images can quantitatively reflect the radiance and reflectance characteristics in the visible spectrum of the crop canopy [
38]. Normalized DNs (
r,
g, and
b, Equations (1)–(3)) were reported to have superior capacity for vegetation estimation compared to original RGB DN values [
39]. In this study, we will construct a further method based on the data in three visible light bands extracted by UAV; we will further analyze the regular changing in each band to create and select indices. The extraction of color features is realized by ENVI5.3.
where
R,
G, and
B are the DNs of the red, green, and blue bands of the ROIs of each plot in the UAV-based images, respectively.
The color feature extraction is illustrated in
Figure 4. The RGB values of different farming progress types were different. The RGB values of four different types of farming progress all decreased step by step with the R, G, and B channels, while the degree of changing for different farming progress types varied. The unharvested wheat fields showed the greatest change in the RGB channels, and the irrigated fields showed the least change in the RGB channels. Therefore, we constructed color indices based on this changing pattern to classify the four types of farming progress fields. The constructed color indices were NRBDI (normalized red-blue–difference index) and MRBDI (modified red–blue difference index). We also tried to use the color indices constructed by previous authors such as NGBDI and NGRDI as shown in
Table 1.
By recombining red, green, and blue images, a grayscale image resulted with perceptual difference in brightness for different land feature types [
40]. We screened the indices, which were selected or constructed by visual interpretation to achieve accurate classification. The selected samples of different farming progress types were also analyzed to obtain the characteristic values of different farming progress types to analyze the feasibility of the index classification.
2.3. Model Construction
By means of visual observation combined with field investigation, samples were randomly selected from each agricultural progress type to cover most features of the same agricultural progress type. Ten samples of unharvested wheat, harvested wheat, tilled, and irrigated fields were selected by visual inspection. We analyzed these samples and the corresponding grayscale images for these samples. Then, we constructed the model based on the color indices after analyzing the histograms of the indices.
The histograms of the grayscale values of the three color indices on the farmland with different types of farming progress are shown in
Figure 5. NRBDI is mainly distributed around 0.2 on unharvested wheat fields, around 0.15 on harvested wheat fields, around 0.1 on tilled fields, and below 0.07 on flooded fields. The NGBDI is mainly distributed around 0.17, 0.13, 0.075, and below 0.053, respectively. The MRBDI is distributed around 0.4, 0.3, 0.2, and below 0.14, respectively. The distributions are all convex arcs, with lower proportions of values at both ends and more concentrated in the middle.
The distribution of the constructed indices was calculated from the above analysis. Although there is less distribution at both ends, there is still a certain amount of overlap. To be able to classify fields in their entirety to meet the needs of agricultural operations, we considered reducing or even eliminating the influence of the values at both ends of the distribution on the classification. Therefore, we created a new method called regional mean model. The rice–wheat rotation fields are rectangular, and according to the size of the fields, we choose a sample size of 100 × 100 pixels. The color index value at the (
i,
j) point in the image is denoted as A =
f(
x_((
i,
j))), and the index range is set to 0–1. The average value of the color index is calculated for each sample, as given in Equation (4).
The analysis using Equation (5) shows that the algorithm can reduce the influence of a few extremes at both ends of the distribution. Equation (5) is used as the basis of this study for the classification of agricultural progress in farmland.
The characteristic values based on the color index and the model were analyzed by calculation of the distribution, the mean, and the standard deviation. In this study, the visual interpretation results were used as the actual values to verify the accuracy of the classification method.
The visual interpretation information was extracted by vectorizing unharvested wheat fields, harvested wheat fields, tilled fields, and irrigated fields in the UAV images based on ground survey data.
To verify the model developed in this paper, the classification accuracy was compared with the classification accuracies of other existing models. Three classification models, that is, random forest (RF) [
41,
42], support vector machines (SVM) [
43], and maximum likelihood classification (MLC) [
44], were used for this comparison.
These three classification methods were implemented using ENVI5.3. The classification results of the four methods were compared to verify the advantages of the method developed in this study.
2.4. Accuracy Assessment
In this study, we used the region of interest (ROI) (32.39° N and 119.42° E, in 2020) as a reference. This reference forms a comparison array with the number of pixels of each class in the classified image. The confusion matrix is then computed. The data of ROI were derived from the visual interpretation of high-resolution images of the study area. We considered the distribution area and complexity of different farming progress distributions in the rice–wheat rotation fields, and finally selected the following: eight unharvested ROIs with a total area of 2593.8 m
2; seven harvested ROIs with a total area of 2493.85 m
2; four tilled ROIs with a total area of 419.85 m
2; one irrigated ROI with a total area of 885.06 m
2. The validation samples of different farming progress types were obtained from the ROIs. We randomly selected the validation samples from the validation ROIs, and the number of samples for each farming progress type was 150. The samples of ground truth ROIs are the total ROI samples, including 4346 unharvested wheat samples, 2243 harvested fields samples, 228 tilled fields samples, and 1518 irrigated fields samples. The confusion matrix is a core component of accuracy evaluation [
45]. It can not only describe the classification accuracy but also indicate the confusion among categories. The basic statistics for the error matrix were the product accuracy (Prod. Acc.), omission, and commission (
Table 2). The Kappa coefficient indicates the degree of matching between the classification result and the real feature category [
46]. The coefficient is used to evaluate the quality of classification. The formula is
In the formula, is the total number of columns in the error matrix; is the number of pixels in row and column of the error matrix (i.e., the number of correctly classified pixels); and are the total numbers of pixels in row i and column i, respectively; N is the total number of pixels used for accuracy assessment.
4. Conclusions and Discussion
A classification method combining color indices and the regional mean model was proposed for analyzing the types of agricultural progress types in rice–wheat rotation farmland. The extraction and classification of the farm progress information of rice–wheat rotation in the study area were successfully realized. We verified the model accuracy, and the Kappa coefficients obtained by combining the NRBDI, NGBDI, and MRBDI with the RM model were 0.86, 0.82, and 0.88, respectively. Furthermore, by comparing the results of cropland information from different classification methods, determining the optimal combination of features, and comparing the results with traditional machine learning classification methods, we demonstrated that the method of color indices combined with the RM model was the best for farmland classification, with 96.9% extraction accuracy for unharvested wheat fields, 89.0% extraction accuracy for harvested wheat fields, 78.2% extraction accuracy for post-harvest tilled fields, and 100% extraction accuracy for irrigated fields. The method employed in this study was more accurate than the traditional machine learning classification methods in identifying the different farming progress types in the field, which should help to target the whole-field operation for different types of farming progress.
We proposed a new approach for the UAV-based classification of agricultural progress in rice–wheat rotation fields. Overall, the method was successful in achieving accurate classification of unharvested, harvested, tilled, and irrigated fields. We found that the MRBDI classification method combined with the RM model achieves the classification of the whole-field area of the rice–wheat rotation fields, thereby eliminating the “pretzel” effect occurring in traditional machine learning classification, which was favorable to the machine for the next farming operation of the whole-field. The method also overcame the difficulty of distinguishing between unharvested fields mixed with rice-stubble and post-harvested rice-stubble fields. We believe that this classification method can be a new technique for monitoring the agricultural progress of rice–wheat rotation fields.
In addition, the method provided new insights and enhancement to farmland and crop classification. Its distinctive feature was that it can be applied to a wide range of rice–wheat rotation fields, overcoming the difficulty of distinguishing between the soil in rice-stubble interstices and tilled fields, thus achieving an accurate classification without removing the interference of soil background. This method also provided ideas for the extraction of rice–wheat rotation fields and UAV image preprocessing. Since the method can be applied to the classification of rice–wheat rotation fields, we believe it can avoid artificial field surveys and save human and material resources. Furthermore, the method avoided shelving farmland, thereby improving the efficiency of farm production.
While the results obtained in this study were highly accurate, there is still room for improvement. First, this study presented a classification method combining color indices and the RM model. The results obtained may be further improved if parameters such as texture and luminance are incorporated. Second, when collecting farmland information, the UAV images were collected under clear skies and low wind speed conditions between 11:00 am and 13:00 pm of local time. It remains to be further verified whether this method applies to other weather conditions, such as cloudy days. Third, this study was aimed at the classification of farming progress of farmland in the rice-stubble wheat region, concentrating on the wheat harvest and rice planting stages. Other stages were not classified, and thus the application of this method in monitoring the whole growing stages of wheat and rice was not verified. Fourth, this study area, therefore, the classification accuracy for a large area of farmland, remains to be verified.