Next Article in Journal
Modification of Gene Expression, DNA Methylation and Small RNAs Expression in Rice Plants under In Vitro Culture
Next Article in Special Issue
Mapping of Phenological Traits in Northeast China Maize (Zea mays L.)
Previous Article in Journal
Characterization of a Wheat-Dasypyrum breviaristatum Chromosome Addition and Its Derived Progenies Carrying Novel Dasypyrum-Specific Gliadin Genes
Previous Article in Special Issue
Remote Sensing of Maize Plant Height at Different Growth Stages Using UAV-Based Digital Surface Models (DSM)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Protecting Steppe Birds by Monitoring with Sentinel Data and Machine Learning under the Common Agricultural Policy

by
Francisco Javier López-Andreu
*,
Zaida Hernández-Guillen
,
Jose Antonio Domínguez-Gómez
,
Marta Sánchez-Alcaraz
,
Juan Antonio Carrero-Rodrigo
,
Joaquin Francisco Atenza-Juárez
,
Juan Antonio López-Morales
and
Manuel Erena
Institute of Agricultural and Food Research and Development of Murcia-IMIDA, Mayor Street, La Alberca, 30150 Murcia, Spain
*
Author to whom correspondence should be addressed.
Agronomy 2022, 12(7), 1674; https://doi.org/10.3390/agronomy12071674
Submission received: 7 June 2022 / Revised: 8 July 2022 / Accepted: 11 July 2022 / Published: 14 July 2022
(This article belongs to the Special Issue Geoinformatics Application in Agriculture)

Abstract

:
This paper shows the work carried out to obtain a methodology capable of monitoring the Common Agricultural Policy (CAP) aid line for the protection of steppe birds, which aims to improve the feeding and breeding conditions of these species and contribute to the improvement of their overall biodiversity population. Two methodologies were initially defined, one based on remote sensing (BirdsEO) and the other on Machine Learning (BirdsML). Both use Sentinel-1 and Sentinel-2 data as a basis. BirdsEO encountered certain impediments caused by the land’s slope and the crop’s height. Finally, the methodology based on Machine Learning offered the best results. It evaluated the performance of up to 7 different Machine Learning classifiers, the most optimal being RandomForest. Fourteen different datasets were generated, and the results they offered were evaluated, the most optimal being the one with more than 150 features, including a time series of 8 elements with Sentinel-1, Sentinel-2 data and derived products, among others. The generated model provided values higher than 97% in metrics such as accuracy, recall and Area under the ROC Curve, and 95% in precision and recall. The methodology is transformed into a tool that continuously monitors 100% of the area requesting aid, continuously over time, which contributes positively to optimizing the use of administrative resources and a fairer distribution of CAP funds.

1. Introduction

According to data from the Pan-European Common Bird Monitoring Scheme (PECBMS) [1], farmland birds have suffered a significant population decline of over 50% on average over the last 40 years. One of the main drivers of the sharp decline in bird populations has been agricultural intensification, which has led to global biodiversity loss [1]. One of the most affected groups is steppe birds that depend on the habitats where agricultural activities occur, being very sensitive to the degradation of these environments.
The loss of fallow land has been linked to the decline of steppe bird populations in Spain, [2], because fallow land (agricultural land left unsown to allow the recovery of soil nutrients) is vital, among other things, for the feeding and maintenance of the species. For this reason, traditional extensive fallow management is considered beneficial for improving biodiversity conservation, especially steppe birds, in agricultural environments.
Since 1960, the European Commission (EC) has aimed to provide a harmonized framework for agriculture through the CAP by defining European cooperation policies and managing support to enforce these rules. In addition, the EC has built a shared context for all its members by using the numerous regulations to achieve a coherent framework for the CAP development [3] with the applicability of the latest technologies. Tarjuelo et al. [4] propose a strategy to reconcile agriculture and biodiversity conservation, based on the recovery of fallow land through CAP, to guarantee and conserve the future of steppe birds.
Farmers were not allowed support until the 2014-2020 CAP period, provided they met three measures: maintaining permanent pasture, growing a minimum of three different crops and establishing landscape features considered essential for biodiversity on 5% of the farmland [5].
The new CAP, therefore, aims to promote the recovery of fallow land and thus also promote the recovery of steppe birds, which have their feeding and breeding grounds in cereal crops. Modifying certain traditional cultivation practices can lead to improved feeding and breeding conditions for these species and thus contribute to improving their population and biodiversity. The Autonomous Community of the Region of Murcia (CARM) aid line for the protection of steppe birds aims to achieve this and sets out obligations associated with the aid to use medium or long-cycle cereal varieties (wheat, spelt, barley, rye, oats and triticale) for sowing and not to harvest before 15 July.
Regulation 1306/2013 [6] is the most recent CAP financing, management, and monitoring law. It formed the basis for the beginning of the regulation of the uses of new technologies, especially remote sensing: Sentinels satellites, unmanned aerial vehicles (UAV), and georeferenced photographs (Regulation (EU) 2018/746 [7]).
Remote sensing has been incorporated into the agricultural sector as a helpful tool for tracking and monitoring activities such as crop production, land use monitoring or environmental protection [8,9]. The last challenge is its incorporation as an essential tool in the CAP. The CAP wants to take advantage of the new Earth Observation (EO) philosophy the European Space Agency (ESA) is developing in the 21st Century. Copernicus Programme (CP), devised by ESA, aims to provide accurate, up-to-date and easily accessible information to improve environmental management and understand and mitigate the effects of climate change.
Some studies are concerned with analyzing the potential of Sentinel-1 (S1) data for assessing different soil, and cereal parameters, such as the one proposed by Bousbih et al. [10]. Other works conclude that the synergistic use of radar and optical data for crop classification generates better information and thus increases classification accuracy compared to optical-only classification [11].
The first remote sensing methodology used to study vegetation has been the opportunity to perform various operations with the bands of satellite images, where the result can be converted into a spectral index. The first and most used index is the Normalized Difference Vegetation Index (NDVI) [12,13]. The NDVI is an indicator of the greenness of the biomass [14]. The NDVI has been used as: Tool for wheat yield estimation [15], time series [16], and prediction model [17].
The Data Cube (DC) Concept is a data cube that describes a multi-dimensional array of values of n-dimensions. This concept was begun used with hyperspectral images in the 90s. The irregular time series cube (ITSC) is a data cube in which data are generated with varying time intervals per spectral index. The ITSC are created from optical images: Landsat [18], MODIS [19], Sentinel-2 (S2) and Sentinel-3 [20] due to cloud cover. However, the use of S1 images has allowed to generation time series cube (TSC), a series of values of a quantity obtained at successive times, often with equal intervals between them [21]. The use of the ITSC and TSC are ways to visualize and understand a cartographic time series quickly.
The use of the DC has increased due to the development of computing, the cloud infrastructures philosophy, and the Analysis-Ready Data (ARD) products. In compliance with operational standards established by the International Organization for Standardization (ISO) and the Open Geospatial Consortium (OGC), data cubes belong to the category of coverages. The OGC Community Practice has defined such a data cube as “established based on a Coverage with a specific grid system for geospatial data with at least one dimension of the spatial or temporal definition” [22,23]. The ARD is then fundamental to the EO data cube, which integrates this data into one logical array multi-temporal thematic analysis. The DC input on machine learning [24,25] allows for generating crop modeling [26]. In this work, the inputs to the generated DCs were multitemporal S1 and S2 images and associated third-level products, such as polarization or vegetation indices, as well as digital elevation model or land use data.
This work aimed to develop a methodology to establish whether applicants for aid for the protection of steppe birds complied with the commitments established in this aid. To this end, following the principles of CAP monitoring, use was to be made of the enormous amount of data provided by the Copernicus Programme through the S1 and S2 ground monitoring missions. In addition, monitoring had to be carried out for all the plots assigned to the aid continuously over time. For each plot, a light traffic color had to be established, whereby red indicated that the commitment was not being fulfilled, green that it was, and yellow that it was impossible to make a clear judgment. Two methodologies were defined and developed in parallel, the results of which were compared to assess which offered the better solution. The first methodology presented a more traditional approach based on remote sensing and finding a biophysical index to obtain an optimal solution. The second methodology was based on Machine Learning and aimed to generate a model that would offer a correct solution with good metrics.

2. Study Area

The study area corresponds to the areas where farmers have applied for aid in the CARM. The different plots are located in the zone of the southwest and plateau of Murcia Region. Cereal land cover (wheat, barley, oats) and fallow and plowed soils. The Murcia Region is characterized by warm temperatures almost all year round. According to the distribution of average temperatures in the region, they exceed 17 C in the Murcia city area, the Abanilla-Fortuna basin, and the coast reaching 20 C around Aguilas town. In the interior, it descends to 16 C, 14–16 C in the Altiplano, and 12 C in the Northwest.
Altitude is another factor that increases temperature variations on a local scale, reaching even lower values in the highest areas of Northwest Murcia. The annual thermal amplitudes are 12–13 C on the coast, 14–16 C in depressions and interior valleys to rise to 17 C in the Altiplano. The average rainfall of the basin can be estimated at 375 mm per year [27], 472 mm at the head, and 317 mm at the bottom.
Figure 1 shows a map indicating the study area location of this work.
Finally, the height of the cereals cultivated in the study area does not exceed 50 cm at the point of greatest vegetative development, which is a crucial characteristic to take into account.

3. Materials

The materials used in this work come from various sources, such as vector and alphanumeric information from help applications, satellite images or data on agricultural land use, together with information on the terrain orography of the study area.
For the development of this work, a proprietary tool based on free software was implemented. This tool was implemented in the Python programming language. It used libraries such as rasterio or geopandas for GIS data processing, and for the functionality linked to Machine Learning, it used scikit-learn. A PostgreSQL database with the PostGIS spatial extension for data storage. The requests library and the Copernicus Open Access API Hub were combined to access the satellite image repository.
Finally, the tool was deployed by an in-house server with Windows 2016 Server as the operating system, 64 GB RAM, 4TB hard drive, and an NVIDIA RXT2080Ti graphics card.

3.1. Aid Declarations Data

The information related to the aid applications is included in the Aid Management application of the Regional Ministry of Water, Agriculture, Livestock, Fisheries, and Environment of the Region of Murcia.
The information we can find in this dataset is identifying the agricultural parcel requesting the aid, the code of the aid requested, and the enclosure’s geometry. The reference system of the vector data is ETRS89 (EPSG:25830).
For the campaign under study, 2019, included in the study area, 1466 applications were submitted, with 9690 hectares. Out of the total number of applications, a reduced number is selected to elaborate the ground-truth matrix, which will be used later for learning the model.

3.2. Remote Sensing Data

The satellite images used to carry out this work belong to ESA’s Copernicus Programme [28] and, more specifically, to the missions focused on terrestrial monitoring, such as S1 and S2. While S1 provides radar images, S2 provides optical images.
The main instrument of S1 is a synthetic aperture radar (SAR) in C-band. The main advantage of this type of satellite is that it can provide images regardless of whether it is day or night and in any weather conditions. The reference system in which these images are provided initially is EPSG:4326. Relative orbits 8 (slices 24 and 25) and 103 (slices 2 and 3) provide coverage of the study area. The product used from the S1 was the Ground Range Detected (GRD).
We did not use the native images provided in the ESA repository for this work, but we carried out a processing [29] for further analysis. This processing applies a series of standard corrections and consists of applying the following seven steps:
  • Apply Orbit File
  • Thermal Noise Removal
  • Border Noise Removal
  • Radiometric Calibration
  • Speckle filtering
  • Range Doppler Terrain Correction
  • Conversion to dB
From the last item in the above list, we get the VV and VH polarized backscatter σ 0_VV and σ 0_VH, in decibels (dB) [30,31,32,33,34] with a damping value of 1 and a kernel size of 7 × 7 to minimize radar speckle in the pictures.
The VV and VH channels are used for the characteristic matrix and the VV+VH and VV-VH dual polarizations.
S2 carries a Multispectral Instrument (MSI), defined as a high resolution, wide swath multispectral imaging system operating in 13 spectral bands at multiple spatial resolutions. The product type used is S2MSI2A, whose main features are level 2A processing, orthorectified and UTM geocoding, BOA, and multispectral reflectance [35]. All elements of this time series had 0% cloud cover over the study area and were provided in the EPSG:32630 reference system. The S2 scene providing coverage for the study area is T30SWH.
All the bands offered in the S2 images were used to carry out this work. The Table 1 shows detailed technical information about these bands:
Third-level products derived from S2 imagery were also used for this work. These products are the vegetation indices NDVI [13], GNDVI [36], SAVI [37], and OSAVI [31].
In addition, we also used images from a unmanned aerial vehicle (UAV).

3.3. External Data

The Geographic Information System of Agricultural Parcels, SIGPAC [38], allows the geographic identification of the parcels declared by farmers and livestock farmers in any aid scheme related to the area cultivated or used by livestock. In addition to the above information, SIGPAC provides the declared land use for each declared agricultural enclosure. At a general level, the SIGPAC distinguishes 30 different agricultural land uses. For this work, we did not have information on the specific crop of each of the studies included in the study area but rather a typology in which they could be included.
The data on the orography of the terrain included in the study area is obtained through the digital elevation model (DEM) provided by the National Center for Geographic Information (CNIG) [39]. The digital model used in this work has a grid pitch of 5 m and is presented in the ETRS89 geodetic reference system.

3.4. Field Data Campaign

The field campaigns were designed based on the information provided by the CARM. The plots were documented by survey in Survey123, filled with a Samsung Galaxy Ta2 Tablet, and photographed with Canon EOS 70D camera installed on a tripod. In a random way and based on the plot with the aid applications provided by the CARM, it was decided in which plots the field radiometry would be carried out with an Ocean Optics radiometer. Through these campaigns, a ground-truth matrix was created.

4. Methods

The work we are concerned with in this paper was approached through two different methodologies, the results of which will be confronted at the end of the paper. These methodologies were developed in parallel and aimed at determining whether aid declarants complied with their obligations.
The first methodology, BirdsEO, had a more traditional approach as it used earth observation techniques based on satellite imagery or derived products. The objective was to obtain an index derived from these satellite images to discern whether respondents complied with their obligations.
The second methodology, called BirdsML, presented a somewhat more contemporary approach, as it used Machine Learning techniques in combination with satellite imagery from Sentinel missions 1 and 2. The objective of this methodology was to obtain a Machine Learning-based model that would predict, based on probabilities, whether or not aid applicants were compliant.
The project has a duration of three years. It began on 1 July 2019. The planning was dynamic, and it changed with the evolution of the project. The activities schedule can be consulted in Table A1.

4.1. BirdsEO

The project was planned as a decision tree, as shown in Figure 2, which inputs were the images and outputs were the plots harvested before or after July 15. Decisions would be made based on thresholds of spectral and polarization indices. First, we work with the satellite images; for S2, we check the cloud cover and generate the spectral indices and, in the same way, for S1, we develop the derived products and their corresponding polarised indices. After this step, the previous results are incorporated into the time series. Then, a trend analysis of the time series is carried out to check that there is a drop in the indices as the plot is harvested [17]. Finally, because the images can only identify which plots are harvested or not, it is necessary to identify which farmers have applied for aid; for this reason, it is needed to compare with the CARM database to know which applicants do or do not meet the aid requirements.
The first step was to make those responsible in the Ministry of Agriculture of the CARM and the farmers understand the problems with protecting steppe birds in the Region of Murcia and what needs to be covered. The field data campaigns were selected from aid requested GIS, and they have three different aims: radiometry campaign, UAV campaign (Figure 3), and getting data about crop type and phenology using Survey123.
The radiometry campaigns aimed to know the surface’s spectral response through reflectance measures. The reflectance (R( λ )) was obtained according to:
R ( λ ) = L c ( λ ) L w ( λ ) C a l ( λ )
Surface radiance ( L c ( λ )), surface white reference radiance ( L w ( λ )), and correction factor white ( C a l ( λ )). The continuous spectra are transformed to discrete spectra according to UAV and S2 bands:
R B i = o f b ( λ ) R ( λ ) d λ o f b ( λ ) d λ
Spectral response functions of S2 and UAV bands ( f b ( λ )), cover reflectance (R( λ )), and cover reflectance of S2 and UAV bands ( R B i ).
The UAV flights were simultaneous to the radiometry campaigns above the same crop areas. Acquired data were processed using Pix4Dmapper (Pix4D SA, Cheseaux—Lausanne, Switzerland), where image calibration, point cloud densification, and ortho-photomosaics (in WGS 84 UTM Zone 30 coordinate system) were calculated from each of the datasets [40].
The following biophysical indices were generated using the red, green and blue bands of the S2 images: Normalized Green Red Difference Index (NGRDI) [40,41], Green Leaf Index (GLI) [42], Red Green Blue Vegetation Index (RGBVI) [43], Visible Atmospherically Resistant Index (VARI) [44] and Triangular Greenness Index (TGI) [45].
The Sentinels images were downloaded weekly from EarthExplorer ESA Hub Copernicus. The S2 images were converted to 10 m spatial resolution using ESA free software, the Sentinel Application Platform (SNAP) [46].
Each image’s bands were resampled to 10 m using SNAP [47] and calculated the following spectral indices. However, there is a wide variety of them [44]: Normalized Difference Vegetation Index (NDVI), Green Difference Vegetation Index (GDVI), Green Normalized Difference Normalized Vegetation Index (GNDVI), Normalized difference near-infrared/shortwave infrared Normalized Burn Ratio (NBR) and Normalized Difference Infrared Index (NDII) [48]. Other spectral indices are proposed based on inverse reflectance differences, such as Anthocyanin Reflectance Index (ARI) or Carotenoid Reflectance Index (CRI550) [49]. However, most indices are based on reflectance in more than three bands. Some indices, such as the Soil Adjusted Vegetation Index (SAVI) [50] and Chlorophyll Absorption Ratio Index (CARI) [51], are based on a transformation technique to minimize the influence of soil brightness or the effects of non-photosynthetic materials, and they have very complex forms. In addition, integrated indexes such as the Transformed Chlorophyll Absorption Ratio (TCARI) and Optimized Soil Adjusted Vegetation Index (OSAVI) were taken into account for this work [52]. The spectral index of each date was exported to ENVI format. In ENVI v4.8, irregular time series spectral indices cube (ITSIC) was generated.
The S1 images used in this project were Ground Range Detected (GRD) Interferometric Wide Swath (IW) products. The radar images were projected onto a standard 10-m grid in GRD output using SNAP-free software from the ESA.
The polarization indices were calculated [30]: Polarization Ratio Index (PRI), Normalize Difference Polarization Index (NDPI), and Polarization Index (PI). Finally, it is exported to ENVI v4.8 format, and a time series polarisation index cube (TSPIC) was generated.

4.2. BirdsML

Figure 4 shows a summary of the workflow of the methodology defined to carry out this work:
In short, the objective of this work was to develop a methodology that, through satellite data and other data from different sources and the use of Machine Learning techniques, would make it possible to check whether the aid applicants complied with the obligations undertaken. Following the CAP monitoring process guidelines, the final result was based on a traffic light system, with red indicating that the commitment was not fulfilled, green that it was fulfilled and yellow that it is impossible to give a clear verdict.

4.2.1. Sentinel Repository

Using the Copernicus Hub API, a process was implemented that established a connection with the Hub and obtained the S1 and S2 images for a specific date after performing the appropriate queries.
This same process was in charge of preprocessing the Sentinel images. On the one hand, on the S1 images, applying the transformations indicated in that section. On the other hand, it generated the third-level products derived from the S2 images, indicated in the section with the same name.
Once the preprocessed Sentinel images were available, a cropping operation was applied to adjust them to the study area. Other reprojection operations were applied so that all the input data had the same reference system. In this case, the reference system selected was EPSG:32630.

4.2.2. Input Data

In addition to the preprocessed Sentinel images, the implemented methodology used information related to the ground-truth plot, agricultural land use according to SIGPAC, and the digital elevation model.
In this part, the implemented process performed tasks related to the standardization of the input data in order to make their combination possible. In this way, rasterization was applied to the true-land parcel. The agricultural use information performed a spatial cross-referencing with the true-land parcel and subsequent rasterization.
Once all the input data were in raster format, proceed with the resampling. They all had a spatial resolution of 10 m and, finally, with a reprojection to the EPSG:32630 reference system.

4.2.3. Ground-Truth Data

Theground-truth data used contained elements of two different classes. Class 1 indicates that the obligation to help was fulfilled, and class 99 indicates that it was not. The samples were generated through information obtained from field visits.
As a result, a sample of 43,096 pixels was obtained, which was distributed as shown in the Table 2:
For the Fulfill class, utilizing field visits, a series of plots were selected that were going to be cultivated with medium or long-cycle cereals and that had not applied for the aid associated with the protection of steppe birds. In this way, it was possible to generate ground-truth information that would not subsequently have to be predicted by the model, thus avoiding, as far as possible, over-fitting and favoring better generalization by the model, which translates into more accurate predictions. They were incorporated into the ground-truth matrix as polygons.
For the Not Fulfill class, on S2 images, a series of polygons were digitized randomly over areas where no cereals of the type included in the aid obligations were grown. These polygons were located on areas with a similar appearance to the crops they intended to identify and on areas with a completely different appearance.
The spatial resolution of the ground truth matrix was the same as that of the input data (Section 4.2.2) so that each pixel was 10 × 10 m in size.
Concerning the Sentinel images, as the objective was to infer whether the aid declarations had complied with the requested aid obligations as of 15 July 2019 crop year, the implemented process generated a time series composed of 8 images between March and July 2019.
At this point, to discern how each input data contributed to the model’s performance, the implemented process generated different learning matrices by combining these input data. In addition to the input data indicated in the previous section, the developed process incorporated, in the ground-truth matrix, features related to date and location. In this way, the following datasets were generated:
Table 3 specifies the features and the number of bands that made up each of the datasets generated:
As a method of selecting training and test items, the traditional percentage division, whereby a percentage, e.g., 80%, of the original dataset, goes to training and the rest to testing, was discarded. Instead, K-Fold cross-validation was used to adjust and validate the model [53], with k = 5. This method, for K times, mixes the dataset randomly and then divides the dataset into K groups. One of the groups is taken as test data set, and the remaining as training data. With the division done, it fits a model on the training dataset and evaluates it with the test dataset. Once finished, the score of the selected metrics is recorded, and the model is discarded. The final performance of the proposed model will be the average of the K times the process has been run. Thus, we have a less optimistic estimate of the model’s performance.
Finally, data normalization operations and balancing of the ground-truth matrix [54] were applied to ensure that the evaluated models competed on an equal footing.

4.2.4. Model Selection

The implemented process evaluated the performance of several Machine Learning classification models, including the classic KNN, Decision Tree (CART), Extra Trees (EXT), Random Forest (RF), Bagging (BG), or more novel XGBoost [55] (XGB), and LightGBM [56] (LGBM).
The dataset used for this part of the process was the one with the most features, labeled S2-S1-COOR-OTHER.
The approach for developing this work was to have simple models initially and then sophisticate the model with the best initial performance. For this initial step, it was decided to use the baseline of all the selected models to evaluate their performance. These baseline models were created with the minimum possible definition and using the indispensable parameters for their creation. The default definition established by the scikit-learn library was used for all of them. This section aimed to contextualize and understand the datasets’ characteristics and the possible potential of the selected models with the created datasets. K-Fold type cross-validation was used for the validation of the models, with k = 5, and the selected metrics were Accuracy (ACC),Kappa, Precision, Recall, F1, and Area under the ROC Curve (ROC-AUC Curve) [57].

4.2.5. Model Optimization

Once the model with the best performance was selected, it was optimized. For this purpose, many hyperparameters with a wide range of possible values were selected. Again, cross-validation was used for the combination of values. The hyperparameters selected for optimization were the number of estimators, maximum number of features, maximum depth, number of elements per leaf, or class weight. The Grid Search method was used as an optimization method, which evaluates the model’s performance for every possible combination that can be given, and finally, the combination with the set of values that offers the best performance.

4.2.6. Model Validation

Again, cross-validation was the technique selected to validate the optimized model. The metrics generated were the same as those in the Model Selection section. In addition, the confusion matrix was generated, which, although not a metric in itself, is beneficial for understanding the model’s performance. Learning and model validation curves were also generated, along with a list of the importance of different features in making a prediction.

4.2.7. Prediction

In the final phase, the implemented process predicted the total surface of the study area at the pixel level. It generated a raster in which, for each pixel, the predicted class and the probabilities of belonging to the two existing classes were recorded.
On the other hand, to make a definitive judgment as to whether or not the declarant complied with the aid obligations, based on the traffic light system, it was necessary to move from a classification by pixel to a classification by an object, at the parcel level.
For this purpose, the algorithm obtained different zonal statistics between the raster resulting from the Machine Learning classification and the parcel with the aid declarations. At this point, generic statistics were stored, such as the mean, standard deviation, number of pixels or majority, and other custom-created ones, such as the number of pixels per class and the average probability of the prediction for each class.
Given the zonal statistics of the previous paragraph, to eliminate the influence of non-target items found at the boundaries of the plots, an adjustment was made to analyze the results obtained, excluding the pixels on the perimeter of the enclosure. Once these pixels were excluded, the same statistical information was generated as in the previous paragraph.
In short, this part of the methodology was used to generate many parameters to infer whether or not an applicant complied with the requested aid obligations, regardless of the metrics that can be extracted from the models.
Figure 5 shows, for a given part of the study area, the land use on a date close to 15 July 2019, and the plots applied for steppe bird protection aid. The orthoimage corresponds to a true-color composition from the S2 image of 10 July 2019.
Figure 6a shows the prediction performed by BirdsML for the above area. Figure 6b shows the results in the form of a traffic light once the classification at the object level is obtained from the classification per pixel for each of the plots. To arrive at this object-level classification, the pixels of the different classes are counted, excluding those on the perimeter of the plots. To minimize the aids awarded when the obligations are not met, a threshold of 70% is set for the parcel to be marked as fulfilled.

5. Results

5.1. BirdsEO

In the first stage, the field data allowed the construction of a spectral library. Thanks to using the ArcGIS Survey123 ESRI’s forms-based tool for creating, sharing and analyzing surveys [58]. In this way, it has been possible to generate a library where, for each point, all the information acquired during the different campaigns is visualized and thus obtain a cube of irregular time series field data (ITSFDC). The ITSFDCs are converted to Google Earth files to be visualized and located quickly and easily on any platform and computer.
The high spatial resolution of UAVs, in this case, is not a reasonable solution since it requires a high number of flights and the processing of large amounts of files due to the large extension of the study area.
A hyperspectral image is when there is an image of one wavelength in each layer for a specific date. The ITSC was generated per each spectral index in several files because its size was too large and could not be parsed because the pilot plot was scattered east of CARM. Figure 7 illustrates a 3D cube of a hyper-temporal image as each layer corresponds to a different date of the same variable.
Because the philosophy of the work is to study the trend and to see how the images evolve before and after 15 July, an image before this deadline and one immediately after it had to be available. In the trend analysis for a specific date, the information before and after this date cannot be missing. The image immediately before 15 July was discarded because it had cloud cover, as shown in Figure 8. It was impossible to analyze what type of crop existed in each plot and its state of harvest.
The time series cubes σ 0_VV(dB) and σ 0_VH(dB), as well as the corresponding polarisation index cube (TSPIC), were generated. A suitable pixel was selected in each plot and generated its time series. The radar images obtained by the Sentinel 1A and 1B satellites have a periodicity or revisit for six days. In 2020, there is an image on 12 July and another on July 18 of the study area. The last visit to the field was made on July 16, ending the field campaign and understanding that the situation observed on that day would be the same as that observed by the satellite on 18 July.
A first comparative analysis of the NDPVH-VV values of 12 and 18 July shows apparent differences between both dates. These differences are observed in a grey range map of the temporal difference map Δ (NDPVH-VV) (Figure 9), that is, the difference between NDPVH-VV (After 15 July = 18 July) and NDPVH-VV (Before 15 July). July = 12 July). In the difference map Δ (NDPVH-VV), we obtain positive values for the areas that do not change and negative values for the areas that change, as observed in the time series of the areas in which the field monitoring has been carried out:
  • Δ (NDPVH-VV) = NDPVH-VV(After 15 July)-NDPVH-VV (Before 15 July)
  • Δ (NDPVH-VV) > 0 → No change (red) θ Not harvested
  • Δ (NDPVH-VV) < 0 → Change (green) θ Harvested
Figure 9 shows that there are positive and negative values in the same plot. This is because the slope and plant height influence the radar response, so there are no negative pixels in plots with mostly positive pixels and vice versa.
Figure 10 shows the temporal evolution of σ 0_VV of a harvested plot (a) and a non-harvested plot (c), together with the temporal evolution of the NDPI, subfigures b and d, where it can be seen that on the dates when harvesting takes place (before July 15) the slope tends to be ascending and then descending.
If we focus the analysis within the plots, we observe a corresponding heterogeneity of values of Δ (NDPVH-VV). This heterogeneity is due to the radar signal and the Earth’s surface. The first factor is the terrain relief, and the second is the scattering mechanisms of the radar signal on vegetation.
The approach of developing a methodology that would allow an index to be obtained from traditional techniques using images in the radar and optical range presented several complications: in the optical range due to cloud cover and the radar range due to the slope of the terrain and the height of the crop. For these reasons, all results obtained by this methodology were considered inconclusive.

5.2. BirdsML

In this section, the metrics analyzed by each classifier and the results obtained by the selected model for each dataset generated will be discussed in detail.

5.2.1. Best Model Selection

Once the ground-truth matrix was generated, the first task of the implemented tool was to select the best performing Machine Learning model for the dataset with the most features (S2-S1-COOR-OTHER), with its normalized or standardized data, balanced and with the baseline definition of each of the models. The models selected for evaluation could be divided between the classics, such as KNN or Random Forest, and the newer ones, such as XGBoost or LightGBM.
The results and the standard deviation obtained for each of these selected models are shown in Table 4:
To select the model that would pass to the optimization phase, we decided to give a higher value to the metrics Kappa, Recall, F1 and AUC-ROC Curve. It was decided to give less weight to the ACC metric score because we had an imbalanced dataset (Table 2). When the distribution of the classes is highly skewed, it can become an unreliable measure. The same strategy was chosen concerning Precision since this metric focuses more on detecting the positive class. In this case, it penalized more the fact of not detecting a negative class since it translates into an economic allocation to an applicant who did not comply with the obligations associated with the aid.
The KNN, RF and LGBM models present very close results on these metrics. The standard deviation in Table 4 is obtained from each of the iterations carried out in the K-Fold cross-validation. Thus, the higher the standard deviation, the more significant the difference in the K results obtained. Paying attention to this standard deviation, we observe lower values in the RF model, implying that the k results generated were more stable, presenting more minor differences between them. It was finally decided to select Random Forest.

5.2.2. Data Analysis

Once the best-performing model was selected, the process adjusted the model and validated it for each of the datasets generated after optimizing this model. Table 5 shows the metrics obtained for each of the datasets generated:
After finding that the date-related features of the Sentinel images did not contribute positively to improved model performance, it was decided to remove them from successive datasets. From the table above, it can be seen that the best performing dataset was the one that incorporated all the features (S2-S1-COOR-OTHER).

5.2.3. Optimized Hyperparameters

The relationship of the values established for each of the hyperparameters that were taken into account in the optimization phase of the Random Forest model is shown below:
  • Number of trees in the forest (n_estimators): 800
  • The minimum number of samples required to split an internal node (min_samples_split): 35
  • The minimum number of samples required to be at a leaf node (min_samples_leaf): 60
  • The number of samples to draw to train each base estimator (max_samples): 0.9
  • The maximum depth of the tree (max_depth): 60
  • The function to measure the quality of a split (criterion): gini
  • Weights associated with classes (class_weight): balanced

5.2.4. Features Importance

Figure 11 shows the importance of the characteristics in the final performance of the model, grouped by type. It can be seen that the S1 features were of little importance and that the S2 bands were the most important.
In Figure 12, the relevance of each time series data in the generated model can be consulted. We can see how the initial dates had little impact on the final performance of the model.

5.2.5. Learning Curve

Figure 13 shows the learning curve of the model generated in this work, in which the relationship between the number of samples and accuracy can be seen. It shows how a plateau was reached towards the end of the number of available samples.

6. Discussion

The objective of satellite image monitoring is to carry out a periodic and continuous control that verifies the status of an activity observed through a time series of images. This work has aimed to propose a methodology to automatically monitor the aid line for the protection of steppe birds promoted by the CAP, using Sentinel 1 and 2 missions and derived products that improve the decision-making process on granting aid. As part of the methodology, a traffic light map was developed to indicate which applications comply with the aid requirements (green) and which do not (red) [24].
To achieve our objectives, we needed to use images from S1 and S2. Two methodologies with different approaches were proposed to choose the most suitable for the achievement of the proposed objective at the end of the process:
  • BirdsEO: proposes a strategy based on using more traditional remote sensing to combine different indices to see if the helpline conditions are met.
  • BirdsML: proposes combining ML algorithms with satellite imagery, mainly S1 and S2, and derived products (vegetation or polarised indices).
To achieve our goal, it was necessary to collect data for BirdsML from S2 without cloud cover; specifically, eight days were chosen. After selecting the images, different ML models were evaluated, RF was the one with the best metrics. Once the model was established, the model was optimized, finally obtaining values between 95–97% for all the monitored metrics. In the prediction phase, a pixel-level classification was obtained, then translated into a plot classification to get results like a traffic light. A threshold of 70% of predicted pixels was set with class 1 or Fulfill to establish which plot met the obligations. Furthermore, the BirdsEO methodology was inconclusive due to the crop typology, as the plant height did not allow results to be obtained using traditional remote sensing. The confusion created by satellite images to determine the plant’s height is due to the slope of the surface or the inclination of the plant caused by the wind. This problem is typical in radar images and S1 images too [59,60,61] . Therefore the methodology finally proposed as a solution to the problem was BirdsML.
Different proposals in the literature address the issue of detecting crops such as cereals using ML models. Paper [62] applies the use of time series on SAR S1 data, valid for the classification of several types of crops and determining that cereals have a different time signature than the rest of the crops. In paper [63] they demonstrated that estimation of cover crop biomass is feasible using only VIs, as well as integrating optical remote sensing and SAR. Finally, paper [64] presented a novel technique based on deep learning for crop type mapping, using NDVI time series for crop type mapping and compared with other advanced supervised learning techniques.
The BirdsML methodology can establish which applicants meet the conditions for this type of aid. These mechanisms represent progress in the digital transformation of public administrations, as it reduces the time spent on administrative tasks, improve efficiency in the management and granting of aid and generates savings in field trips or on-site controls. As weaknesses, we can highlight that the ground-truth dataset was unbalanced. For this reason, it would have been possible to develop further the application of techniques to balance it and check its repercussion on the model artificially. In addition, we could stop using traditional ML models and explore more carefully other more recent models such as LightGBM, XGBOOST, or even neural networks. The main idea is to improve the methodology during future campaigns by strengthening its weak points.

7. Conclusions

This work aimed to generate a methodology to discern whether applicants for aid for the protection of steppe birds complied with their commitments. These commitments include sowing medium or long-cycle cereal varieties and not harvesting before 15 July.
Two methodologies have been proposed to achieve the objective, one with a more traditional approach based on remote sensing (BirdsEO) and the other using Machine Learning (BirdsML). The first one used satellite images from the S1 and S2 missions to try to obtain an index as a solution. The second was based on combining data from different sources, such as time series of satellite images with their associated products, land uses and the aid applicant plots. This data was processed by applying Machine Learning techniques to judge whether an applicant was complying with the required obligations. In this way, we would end up with two results, and the final solution could be either of the two methodologies or a combination of both.
About the so-called BirdsEO methodology, certain impediments were encountered, mainly related to the land’s slope and the crop’s height. Therefore, the results were inconclusive, as they did not allow us to discern whether any medium or long-cycle cereal varieties were grown on the plots that applied for assistance.
In BirdsML, up to 14 different datasets were generated, and seven different Machine Learning classifiers were evaluated. The classifier that presented the most optimal results on the dataset with the highest number of features and the baseline definition was Random Forest, with 97% in metrics such as accuracy, recall and ROC-AUC Curve, and 95% in precision and recall. The LightGBM model also gave excellent results. As for the datasets generated, the one with the best metrics had the most significant number of features, called S2-S1-COOR-OTHER, which had a total of 156 independent variables, including a time series of satellite images S1 and S2, from 8 different dates.
At a more detailed level, it was found that the S1 data did not provide good metrics. On the other hand, contrary to what one might initially think, incorporating date-related features for each time series element into the Sentinel images did not help the model generalize better. Incorporating information related to the geographic location of the data offered a slight improvement in the model metrics.
Focusing on the importance of the features grouped by typology, it was observed that S2 data, grouping images and derived vegetation indices, summed about 70%, and S1 data (images + indices) only 4%. Characteristics related to land use and geographic location evenly distributed the remaining percentage.
Regarding the S2 time series dates, it was observed that the initial dates, when the crop is less developed, have little relevance to the model’s predictions.
Finally, we observed that a plateau phase was reached in the model learning curve towards the end of the maximum number of available samples, but it was not observed.
Among the improvements of this line of research is introducing the concept of feature engineering, whereby a compromise between dimensionality reduction of the ground-truth matrix and model performance would be sought, or whereby new relevant features such as agroclimatic information would be introduced [65]. Another option could be working more deeply with the LightGBM model or neural networks.

Author Contributions

Conceptualization, F.J.L.-A., Z.H.-G., J.A.D.-G. and M.E.; methodology, F.J.L.-A., Z.H.-G., J.A.D.-G., J.A.L.-M. and M.E.; software, F.J.L.-A. and J.A.L.-M.; validation, Z.H.-G., J.A.D.-G., M.S.-A., J.A.C.-R. and M.E.; formal analysis, F.J.L.-A., Z.H.-G., J.A.D.-G. and M.E.; investigation, F.J.L.-A., Z.H.-G., J.A.D.-G. and M.E.; resources, Z.H.-G., J.F.A.-J., M.S.-A. and J.A.C.-R.; writing–original draft preparation, F.J.L.-A., J.F.A.-J., J.A.L.-M. and J.A.D.-G.; writing–review and editing, F.J.L.-A., J.A.L.-M., J.A.D.-G. and M.E.; visualization, J.F.A.-J., M.S.-A. and J.A.C.-R.; supervision, F.J.L.-A., Z.H.-G., J.A.L.-M. and M.E.; project administration, Z.H.-G. and M.S.-A.; funding acquisition, M.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European Regional Development Fund (ERDF), through project FEDER 14-20-25 “Impulso a la economía circular en la agricultura y la gestión del agua mediante el uso avanzado de nuevas tecnologías-iagua”.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

The authors thank the European Space Agency, the European Commission, and the Copernicus Programme.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The workflow followed during this project is shown in Table A1.
Table A1. Activity schedule.
Table A1. Activity schedule.
ActivityMethodologyActivity Period
State of ArtBirdsEO, BirdsMLJuly-2019 & September-2019
Radiometry CampaignsBirdsEOJuly-2019 & September-2019
UAV CampaignsBirdsEOJuly-2019 & September-2019
Survey123 CampaignsBirdsEO, BirdsMLJuly-2019 & September-2019
S1&S2 DownloadBirdsEO, BirdsMLJuly-2019 & September-2019
S1&S2 Indices GenerationBirdsEO, BirdsMLJuly-2019 & September-2019
S1&S2 Indices AnalysisBirdsEO, BirdsMLJuly-2019 & September-2019
Machine LearningBirdsMLJanuary-2020 & December-2021
Paper and ReportBirdsEO, BirdsMLJanuary-2022 & June-2022

References

  1. PanEuropean Common Bird Monitoring Scheme. Available online: https://pecbms.info/ (accessed on 12 November 2021).
  2. Traba, J.; Morales, M.B. The decline of farmland birds in Spain is strongly associated to the loss of fallowland. Sci. Rep. 2019, 9, 9473. [Google Scholar] [CrossRef]
  3. Matthews, A. The European Union’s common agricultural policy and developing countries: The struggle for coherence. Eur. Integr. 2008, 30, 381–399. [Google Scholar] [CrossRef]
  4. Tarjuelo, R.; Margalida, A.; Mougeot, F. Changing the fallow paradigm: A win–win strategy for the post-2020 Common Agricultural Policy to halt farmland bird declines. J. Appl. Ecol. 2020, 57, 642–649. [Google Scholar] [CrossRef]
  5. Regulation 1307/2013. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:32013R1307 (accessed on 28 October 2021).
  6. Regulation 1306/2013. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:32013R1306&from=en (accessed on 21 October 2021).
  7. Regulation 2018/746. Available online: https://eur-lex.europa.eu/legal-content/EN/TXT/PDF/?uri=CELEX:32018R0746&from=EN (accessed on 2 April 2021).
  8. Wójtowicz, M.; Wójtowicz, A.; Piekarczyk, J. Application of remote sensing methods in agriculture. Commun. Biometry Crop. Sci. 2016, 11, 31–50. [Google Scholar]
  9. Shanmugapriya, P.; Rathika, S.; Ramesh, T.; Janaki, P. Applications of remote sensing in agriculture—A Review. Int. J. Curr. Microbiol. Appl. Sci. 2019, 8, 2270–2283. [Google Scholar] [CrossRef]
  10. Bousbih, S.; Zribi, M.; Lili-Chabaane, Z.; Baghdadi, N.; El Hajj, M.; Gao, Q.; Mougenot, B. Potential of Sentinel-1 radar data for the assessment of soil and cereal cover parameters. Sensors 2017, 17, 2617. [Google Scholar] [CrossRef] [Green Version]
  11. Van Tricht, K.; Gobin, A.; Gilliams, S.; Piccard, I. Synergistic use of radar Sentinel-1 and optical Sentinel-2 imagery for crop mapping: A case study for Belgium. Remote Sens. 2018, 10, 1642. [Google Scholar] [CrossRef] [Green Version]
  12. Rouse, J.; Haas, R.; Schell, J.; Deering, D. Monitoring vegetation systems in the great plains with ERTS proceeding. In Third Earth Reserves Technology Satellite Symposium; NASA: Greenbelt, MD, USA, 1974; SP-351; Volume 30103017. [Google Scholar]
  13. Tucker, C.J. Red and photographic infrared linear combinations for monitoring vegetation. Remote Sens. Environ. 1979, 8, 127–150. [Google Scholar] [CrossRef] [Green Version]
  14. Copernicus Global Land Service. Available online: https://land.copernicus.eu/global/products/ndvi (accessed on 4 May 2021).
  15. Sultana, S.R.; Ali, A.; Ahmad, A.; Mubeen, M.; Zia-Ul-Haq, M.; Ahmad, S.; Ercisli, S.; Jaafar, H.Z. Normalized difference vegetation index as a tool for wheat yield estimation: A case study from Faisalabad, Pakistan. Sci. World J. 2014, 2014, 725326. [Google Scholar] [CrossRef] [Green Version]
  16. Chao Rodríguez, Y.; El Anjoumi, A.; Domínguez Gómez, J.; Rodríguez Pérez, D.; Rico, E. Using Landsat image time series to study a small water body in Northern Spain. Environ. Monit. Assess. 2014, 186, 3511–3522. [Google Scholar] [CrossRef]
  17. Domínguez, J.A.; Kumhálová, J.; Novák, P. Winter oilseed rape and winter wheat growth prediction using remote sensing methods. Plant Soil Environ. 2015, 61, 410–416. [Google Scholar]
  18. Echavarría-Caballero, C.; Domínguez-Gómez, J.A.; González-García, C.; García-García, M.J. Water quality spatial-temporal analysis of gravel pit ponds in the southeast regional park Madrid (Spain) from 1984 to 2009. Geocarto Int. 2022. Available online: https://www.tandfonline.com/doi/epub/10.1080/10106049.2022.2037736?needAccess=true (accessed on 6 April 2022). [CrossRef]
  19. Wardlow, B.D.; Egbert, S.L.; Kastens, J.H. Analysis of time-series MODIS 250 m vegetation index data for crop classification in the US Central Great Plains. Remote Sens. Environ. 2007, 108, 290–310. [Google Scholar] [CrossRef] [Green Version]
  20. Shrestha, A.B. Enhancing Temporal Series of Sentinel-2 and Sentinel-3 Data Products: From Classical Regression to Deep Learning Approach. Ph.D. Thesis, Universidade NOVA de Lisboa, Lisbon, Portugal, 2021. [Google Scholar]
  21. Wagner, W.; Bauer-Marschallinger, B.; Navacchi, C.; Reuß, F.; Cao, S.; Reimer, C.; Schramm, M.; Briese, C. A Sentinel-1 backscatter datacube for global land monitoring applications. Remote Sens. 2021, 13, 4622. [Google Scholar] [CrossRef]
  22. Open Data Cube. Available online: https://www.opendatacube.org/overview (accessed on 21 May 2021).
  23. Data Application of the Month: Earth Observation Data Cubes. Available online: https://www.un-spider.org/links-and-resources/daotm-data-cubes (accessed on 15 January 2022).
  24. López-Andreu, F.J.; Erena, M.; Dominguez-Gómez, J.A.; López-Morales, J.A. Sentinel-2 images and machine learning as tool for monitoring of the common agricultural policy: Calasparra rice as a case study. Agronomy 2021, 11, 621. [Google Scholar] [CrossRef]
  25. López-Andreu, F.J.; López-Morales, J.A.; Erena, M.; Skarmeta, A.F.; Martínez, J.A. Monitoring System for the Management of the Common Agricultural Policy Using Machine Learning and Remote Sensing. Electronics 2022, 11, 325. [Google Scholar] [CrossRef]
  26. Tenreiro, T.R.; García-Vila, M.; Gómez, J.A.; Jiménez-Berni, J.A.; Fereres, E. Using NDVI for the assessment of canopy cover in agricultural crops within modelling research. Comput. Electron. Agric. 2021, 182, 106038. [Google Scholar] [CrossRef]
  27. García, M.J.V. Atlas Global de la Región de Murcia/Coordinación Científica, Asunción Romero Díaz; Coordinación Cartográfica, Francisco Alonso Sarría: Reseña Bibliográfica; Universidad de Murcia: Murcia, Spain, 2008; Volume 24, Available online: https://digitum.um.es/digitum/handle/10201/11948 (accessed on 5 May 2022).
  28. European Space Agency. Available online: https://www.esa.int (accessed on 5 May 2022).
  29. Filipponi, F. Sentinel-1 GRD preprocessing workflow. Multidiscip. Digit. Publ. Inst. Proc. 2019, 18, 11. [Google Scholar]
  30. Daughtry, C.S.; Walthall, C.; Kim, M.; De Colstoun, E.B.; McMurtrey Iii, J. Estimating corn leaf chlorophyll concentration from leaf and canopy reflectance. Remote Sens. Environ. 2000, 74, 229–239. [Google Scholar] [CrossRef]
  31. Rondeaux, G.; Steven, M.; Baret, F. Optimization of soil-adjusted vegetation indices. Remote Sens. Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  32. Lee, J.S. Speckle analysis and smoothing of synthetic aperture radar images. Comput. Graph. Image Process. 1981, 17, 24–32. [Google Scholar] [CrossRef]
  33. Lee, J.S.; Grunes, M.R.; De Grandi, G. Polarimetric SAR speckle filtering and its implication for classification. IEEE Trans. Geosci. Remote Sens. 1999, 37, 2363–2373. [Google Scholar]
  34. Lee, J.S.; Wen, J.H.; Ainsworth, T.L.; Chen, K.S.; Chen, A.J. Improved sigma filter for speckle filtering of SAR imagery. IEEE Trans. Geosci. Remote Sens. 2008, 47, 202–213. [Google Scholar]
  35. e Santos, D.A.; Martinez, J.; Harmel, T.; Borges, H.; Roig, H. Evaluation of Sentinel-2/Msi Imagery Products Level-2a Obtained by Three Different Atmospheric Corrections for Monitoring Suspended Sediments Concentration in Madeira River, Brazil. In Proceedings of the IEEE Latin American GRSS & ISPRS Remote Sensing Conference (LAGIRS), Santiago, Chile, 22–26 March 2020; pp. 207–212. [Google Scholar]
  36. Gitelson, A.A.; Kaufman, Y.J.; Merzlyak, M.N. Use of a green channel in remote sensing of global vegetation from EOS-MODIS. Remote Sens. Environ. 1996, 58, 289–298. [Google Scholar] [CrossRef]
  37. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  38. SIGPAC. Available online: https://sigpac.mapa.gob.es/fega/visor/ (accessed on 6 May 2022).
  39. CNIG. Available online: https://centrodedescargas.cnig.es/CentroDescargas/index.jsp (accessed on 6 May 2022).
  40. Li, Y.; Chen, J.; Ma, Q.; Zhang, H.K.; Liu, J. Evaluation of Sentinel-2A surface reflectance derived using Sen2Cor in North America. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1997–2021. [Google Scholar] [CrossRef]
  41. Tsouros, D.C.; Bibi, S.; Sarigiannidis, P.G. A review on UAV-based applications for precision agriculture. Information 2019, 10, 349. [Google Scholar] [CrossRef] [Green Version]
  42. Hunt, E.R.; Cavigelli, M.; Daughtry, C.S.; Mcmurtrey, J.E.; Walthall, C.L. Evaluation of digital photography from model aircraft for remote sensing of crop biomass and nitrogen status. Precis. Agric. 2005, 6, 359–378. [Google Scholar] [CrossRef]
  43. Gobron, N.; Pinty, B.; Verstraete, M.M.; Widlowski, J.L. Advanced vegetation indices optimized for up-coming sensors: Design, performance, and applications. IEEE Trans. Geosci. Remote Sens. 2000, 38, 2489–2505. [Google Scholar]
  44. Hunt, E.R., Jr.; Doraiswamy, P.C.; McMurtrey, J.E.; Daughtry, C.S.; Perry, E.M.; Akhmedov, B. A visible band index for remote sensing leaf chlorophyll content at the canopy scale. Int. J. Appl. Earth Obs. Geoinf. 2013, 21, 103–112. [Google Scholar] [CrossRef] [Green Version]
  45. Bendig, J.; Yu, K.; Aasen, H.; Bolten, A.; Bennertz, S.; Broscheit, J.; Gnyp, M.L.; Bareth, G. Combining UAV-based plant height from crop surface models, visible, and near infrared vegetation indices for biomass monitoring in barley. Int. J. Appl. Earth Obs. Geoinf. 2015, 39, 79–87. [Google Scholar] [CrossRef]
  46. Gitelson, A.A.; Kaufman, Y.J.; Stark, R.; Rundquist, D. Novel algorithms for remote estimation of vegetation fraction. Remote Sens. Environ. 2002, 80, 76–87. [Google Scholar] [CrossRef] [Green Version]
  47. SNAP. Available online: https://step.esa.int/main/download/snap-download/ (accessed on 15 October 2021).
  48. Kobayashi, N.; Tani, H.; Wang, X.; Sonobe, R. Crop classification using spectral indices derived from Sentinel-2A imagery. J. Inf. Telecommun. 2020, 4, 67–90. [Google Scholar] [CrossRef]
  49. Zarco-Tejada, P.J.; Miller, J.R.; Noland, T.L.; Mohammed, G.H.; Sampson, P.H. Scaling-up and model inversion methods with narrowband optical indices for chlorophyll content estimation in closed forest canopies with hyperspectral data. IEEE Trans. Geosci. Remote Sens. 2001, 39, 1491–1507. [Google Scholar] [CrossRef] [Green Version]
  50. Gitelson, A.A.; Chivkunova, O.B.; Merzlyak, M.N. Nondestructive estimation of anthocyanins and chlorophylls in anthocyanic leaves. Am. J. Bot. 2009, 96, 1861–1868. [Google Scholar] [CrossRef]
  51. Gitelson, A.A.; Merzlyak, M.N.; Chivkunova, O.B. Optical properties and nondestructive estimation of anthocyanin content in plant leaves. Photochem. Photobiol. 2001, 74, 38–45. [Google Scholar] [CrossRef]
  52. Wu, C.; Niu, Z.; Tang, Q.; Huang, W. Estimating chlorophyll content from hyperspectral vegetation indices: Modeling and validation. Agric. For. Meteorol. 2008, 148, 1230–1241. [Google Scholar] [CrossRef]
  53. Ramezan, C.A.; Warner, T.A.; Maxwell, A.E. Evaluation of sampling and cross-validation tuning strategies for regional-scale machine learning classification. Remote Sens. 2019, 11, 185. [Google Scholar] [CrossRef] [Green Version]
  54. Kaur, H.; Pannu, H.S.; Malhi, A.K. A systematic review on imbalanced data challenges in machine learning: Applications and solutions. ACM Comput. Surv. (CSUR) 2019, 52, 1–36. [Google Scholar] [CrossRef] [Green Version]
  55. Chen, T.; He, T.; Benesty, M.; Khotilovich, V.; Tang, Y.; Cho, H.; Chen, K. Xgboost: Extreme Gradient Boosting; R Package Version 0.4-2; 2015; Volume 1, pp. 1–4. Available online: https://cran.microsoft.com/snapshot/2017-12-11/web/packages/xgboost/vignettes/xgboost.pdf (accessed on 19 September 2021).
  56. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.Y. Lightgbm: A highly efficient gradient boosting decision tree. Adv. Neural Inf. Process. Syst. 2017, 30, 3147–3155. [Google Scholar]
  57. Carvalho, D.V.; Pereira, E.M.; Cardoso, J.S. Machine learning interpretability: A survey on methods and metrics. Electronics 2019, 8, 832. [Google Scholar] [CrossRef] [Green Version]
  58. ArcGIS Survey123. Available online: https://survey123.arcgis.com/ (accessed on 19 September 2021).
  59. Abdikan, S.; Sekertekin, A.; Ustunern, M.; Sanli, F.B.; Nasirzadehdizaji, R. Backscatter analysis using multi-temporal Sentinel-1 SAR data for crop growth of maize in Konya Basin, Turkey. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci 2018, 42, 9–13. [Google Scholar] [CrossRef] [Green Version]
  60. Dirgahayu, D.; Parsa, I.M. Detection Phase Growth of Paddy Crop Using SAR Sentinel-1 Data. In Proceedings of the IOP Conference Series: Earth and Environmental Science, Moscow, Russia, 27 May–6 June 2019; Volume 280, p. 012020. [Google Scholar]
  61. Schiavon, E.; Taramelli, A.; Tornato, A.; Pierangeli, F. Monitoring environmental and climate goals for European agriculture: User perspectives on the optimization of the Copernicus evolution offer. J. Environ. Manag. 2021, 296, 113121. [Google Scholar] [CrossRef]
  62. Arias, M.; Campo-Bescós, M.Á.; Álvarez-Mozos, J. Crop classification based on temporal signatures of Sentinel-1 observations over Navarre province, Spain. Remote Sens. 2020, 12, 278. [Google Scholar] [CrossRef] [Green Version]
  63. Jennewein, J.S.; Lamb, B.T.; Hively, W.D.; Thieme, A.; Thapa, R.; Goldsmith, A.; Mirsky, S.B. Integration of Satellite-Based Optical and Synthetic Aperture Radar Imagery to Estimate Winter Cover Crop Performance in Cereal Grasses. Remote Sens. 2022, 14, 2077. [Google Scholar] [CrossRef]
  64. Seydi, S.T.; Amani, M.; Ghorbanian, A. A Dual Attention Convolutional Neural Network for Crop Classification Using Time-Series Sentinel-2 Imagery. Remote Sens. 2022, 14, 498. [Google Scholar] [CrossRef]
  65. Erena, M.; López, J.; García, P.; Caro, M.; Belda, F.; Palenzuela, J.; Toledano, F.; Torralba, P.; González-Barbera, G.; García-Pintado, J. Estimación de precipitación combinada radar-pluviómetros y publicación mediante servicios OGC. XV CNTIG. CSIC 2012, 1, 412–419. [Google Scholar]
Figure 1. Study area in the Autonomous Community of the Region of Murcia.
Figure 1. Study area in the Autonomous Community of the Region of Murcia.
Agronomy 12 01674 g001
Figure 2. Planning according to decision tree.
Figure 2. Planning according to decision tree.
Agronomy 12 01674 g002
Figure 3. Field data campaigns. (a) Radiometry. (b) UAV.
Figure 3. Field data campaigns. (a) Radiometry. (b) UAV.
Agronomy 12 01674 g003
Figure 4. BirdsML methodology scheme.
Figure 4. BirdsML methodology scheme.
Agronomy 12 01674 g004
Figure 5. BirdsML methodology scheme.
Figure 5. BirdsML methodology scheme.
Agronomy 12 01674 g005
Figure 6. From pixel-level to object-level classification. (a) BirdsML prediction. (b) Results as traffic light system.
Figure 6. From pixel-level to object-level classification. (a) BirdsML prediction. (b) Results as traffic light system.
Agronomy 12 01674 g006
Figure 7. Example of the ITSSC file (in red the pilot areas).
Figure 7. Example of the ITSSC file (in red the pilot areas).
Agronomy 12 01674 g007
Figure 8. Example of the cloud cover with the pilot areas on 13 July 2019, the first image after the deadline (15 July).
Figure 8. Example of the cloud cover with the pilot areas on 13 July 2019, the first image after the deadline (15 July).
Agronomy 12 01674 g008
Figure 9. Map of Δ (NDPVH-VV) as a function of the change threshold.
Figure 9. Map of Δ (NDPVH-VV) as a function of the change threshold.
Agronomy 12 01674 g009
Figure 10. Time series of various covers. (a,b) Harvested Cereal. (c,d) Unharvested Cereal.
Figure 10. Time series of various covers. (a,b) Harvested Cereal. (c,d) Unharvested Cereal.
Agronomy 12 01674 g010
Figure 11. Importance of characteristics grouped by typology.
Figure 11. Importance of characteristics grouped by typology.
Agronomy 12 01674 g011
Figure 12. Relevance of time series dates.
Figure 12. Relevance of time series dates.
Agronomy 12 01674 g012
Figure 13. Learning curve of the model.
Figure 13. Learning curve of the model.
Agronomy 12 01674 g013
Table 1. Characteristics of S2 bands used.
Table 1. Characteristics of S2 bands used.
BandResolution (m)Wavelength (nm)Description
B1600.443Aerosol
B2100.490Blue
B3100.560Green
B4100.665Red
B5200.705Red edge 1
B6200.740Red edge 2
B7200.783Red edge 3
B8100.842Near infrared
B9600.945Water vapor
B10601.375Cirrus
B11201.610SWIR 1
B12202.190SWIR 2
Table 2. Distribution among classes of the ground-truth data.
Table 2. Distribution among classes of the ground-truth data.
ClassPixelsPercentage
Fulfill (1)34,30480
Not Fulfill (99)879220
Table 3. Features of the generated datasets.
Table 3. Features of the generated datasets.
DatasetS2 + IndicesS1 + IndicesDateCoords.SIGPACDEMFeatures
S2 120
S2-COOR 122
S2-DATE 125
S2-OTHER 122
S2-COOR-OTHER 124
S1 32
S1-COOR 34
S1-DATE 37
S1-OTHER 34
S1-COOR-OTHER 36
S2-S1 152
S2-S1-COOR 154
S2-S1-OTHER 154
S2-S1-COOR-OTHER 156
Table 4. Metrics obtained during the model selection phase.
Table 4. Metrics obtained during the model selection phase.
ModelMetrics (%)
ACCKappaPrecisionRecallF1AUC-ROC Curve
KNN96.75 ± 10.3993.37 ± 19.2693.31 ± 17.4393.89 ± 4.3695.23 ± 13.5495.55 ± 8.19
CART93.44 ± 15.1285.58 ± 25.6688.97 ± 20.6693.95 ± 16.7588.99 ± 19.1593.63 ± 12.43
EXT96.36 ± 11.3191.58 ± 22.4294.47 ± 17.6193.99 ± 12.8593.70 ± 16.0495.48 ± 11.48
BG92.75 ± 15.1383.87 ± 25.9587.12 ± 21.5193.60 ± 16.8587.70 ± 19.3693.08 ± 12.45
RF97.40 ± 5.5293.14 ± 11.1094.98 ± 9.2795.31 ± 6.5194.75 ± 6.4996.62 ± 5.70
XGB96.63 ± 7.3390.72 ± 19.4793.67 ± 16.1394.93 ± 15.5592.63 ± 15.6996.00 ± 8.98
LGBM96.75 ± 5.4592.64 ± 14.8193.71 ± 13.6595.06 ± 10.1194.06 ± 11.3596.12 ± 6.92
Table 5. Metrics obtained by Random Forest for each of the generated datasets.
Table 5. Metrics obtained by Random Forest for each of the generated datasets.
DatasetMetrics (%)
ACCKappaPrecisionRecallF1ROC-AUC Curve
S295.28 ± 8.9786.34 ± 24.4990.02 ± 20.3389.63 ± 19.1389.32 ± 18.7893.18 ± 12.47
S2-COOR96.92 ± 8.1391.49 ± 22.3393.67 ± 19.1793.91 ± 15.7693.44 ± 17.1795.80 ± 10.69
S2-DATE94.06 ± 8.9981.26 ± 26.0591.32 ± 20.3281.58 ± 23.0384.80 ± 20.6689.42 ± 13.75
S2-OTHER95.12 ± 11.1989.40 ± 21.0590.83 ± 20.1496.59 ± 6.0392.24 ± 14.8595.66 ± 8.42
S2-COOR-OTHER97.84 ± 6.1494.46 ± 14.4495.04 ± 13.7497.33 ± 6.4395.81 ± 10.6497.65 ± 6.05
S182.61 ± 8.4626.73 ± 34.5263.34 ± 35.3626.11 ± 30.8832.46 ± 32.4461.59 ± 16.37
S1-COOR92.85 ± 9.8980.67 ± 24.8288.14 ± 21.4489.36 ± 22.3084.62 ± 20.0891.5 ± 11.69
S1-DATE90.53 ± 9.9716.73 ± 35.2164.52 ± 36.1118.38 ± 33.6927.12 ± 37.2857.64 ± 18.73
S1-OTHER91.82 ± 12.5982.40 ± 24.5283.05 ± 24.5697.03 ± 5.3987.19 ± 17.4193.76 ± 8.43
S1-COOR-OTHER96.19 ± 7.4889.87 ± 19.3092.18 ± 17.1896.05 ± 15.0792.05 ± 15.4796.13 ± 8.45
S2-S194.92 ± 9.3384.81 ± 24.0492.43 ± 19.4885.38 ± 17.2387.87 ± 18.2491.37 ± 11.77
S2-S1-COOR96.43 ± 8.2590.38 ± 21.2193.70 ± 18.7192.88 ± 14.5992.60 ± 16.0595.11 ± 9.96
S2-S1-OTHER95.10 ± 11.2789.44 ± 21.1794.79 ± 20.1797.33 ± 5.7595.67 ± 14.9297.61 ± 8.44
S2-S1-COOR-OTHER97.77 ± 3.0994.29 ± 7.2694.79 ± 6.9097.33 ± 3.2595.67 ± 5.3597.61 ± 3.04
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

López-Andreu, F.J.; Hernández-Guillen, Z.; Domínguez-Gómez, J.A.; Sánchez-Alcaraz, M.; Carrero-Rodrigo, J.A.; Atenza-Juárez, J.F.; López-Morales, J.A.; Erena, M. Protecting Steppe Birds by Monitoring with Sentinel Data and Machine Learning under the Common Agricultural Policy. Agronomy 2022, 12, 1674. https://doi.org/10.3390/agronomy12071674

AMA Style

López-Andreu FJ, Hernández-Guillen Z, Domínguez-Gómez JA, Sánchez-Alcaraz M, Carrero-Rodrigo JA, Atenza-Juárez JF, López-Morales JA, Erena M. Protecting Steppe Birds by Monitoring with Sentinel Data and Machine Learning under the Common Agricultural Policy. Agronomy. 2022; 12(7):1674. https://doi.org/10.3390/agronomy12071674

Chicago/Turabian Style

López-Andreu, Francisco Javier, Zaida Hernández-Guillen, Jose Antonio Domínguez-Gómez, Marta Sánchez-Alcaraz, Juan Antonio Carrero-Rodrigo, Joaquin Francisco Atenza-Juárez, Juan Antonio López-Morales, and Manuel Erena. 2022. "Protecting Steppe Birds by Monitoring with Sentinel Data and Machine Learning under the Common Agricultural Policy" Agronomy 12, no. 7: 1674. https://doi.org/10.3390/agronomy12071674

APA Style

López-Andreu, F. J., Hernández-Guillen, Z., Domínguez-Gómez, J. A., Sánchez-Alcaraz, M., Carrero-Rodrigo, J. A., Atenza-Juárez, J. F., López-Morales, J. A., & Erena, M. (2022). Protecting Steppe Birds by Monitoring with Sentinel Data and Machine Learning under the Common Agricultural Policy. Agronomy, 12(7), 1674. https://doi.org/10.3390/agronomy12071674

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop