Next Article in Journal
Image and Video Forensics
Next Article in Special Issue
A Semi-Supervised Reduced-Space Method for Hyperspectral Imaging Segmentation
Previous Article in Journal
Analysis, Design and Realization of a Furnace for In Situ Wettability Experiments at High Temperatures under X-ray Microtomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms

1
Information Retrieval and Data Analytics Laboratory, ENSIAS, Mohammed V University in Rabat, Rabat 10100, Morocco
2
Embedded Systems and Artificial Intelligence Department, Moroccan Foundation for Advanced Science Innovation and Research (MAScIR), Rabat 10100, Morocco
*
Author to whom correspondence should be addressed.
J. Imaging 2021, 7(11), 241; https://doi.org/10.3390/jimaging7110241
Submission received: 8 September 2021 / Revised: 12 October 2021 / Accepted: 18 October 2021 / Published: 17 November 2021
(This article belongs to the Special Issue Advances in Multi/Hyperspectral Imaging)

Abstract

:
Smart agriculture is a new concept that combines agriculture and new technologies to improve the yield’s quality and quantity as well as facilitate many tasks for farmers in managing orchards. An essential factor in smart agriculture is tree crown segmentation, which helps farmers automatically monitor their orchards and get information about each tree. However, one of the main problems, in this case, is when the trees are close to each other, which means that it would be difficult for the algorithm to delineate the crowns correctly. This paper used satellite images and machine learning algorithms to segment and classify trees in overlapping orchards. The data used are images from the Moroccan Mohammed VI satellite, and the study region is the OUARGHA citrus orchard located in Morocco. Our approach starts by segmenting the rows inside the parcel and finding all the trees there, getting their canopies, and classifying them by size. In general, the model inputs the parcel’s image and other field measurements to classify the trees into three classes: missing/weak, normal, or big. Finally, the results are visualized in a map containing all the trees with their classes. For the results, we obtained a score of 0.93 F-measure in rows segmentation. Additionally, several field comparisons were performed to validate the classification, dozens of trees were compared and the results were very good. This paper aims to help farmers to quickly and automatically classify trees by crown size, even if there are overlapping orchards, in order to easily monitor each tree’s health and understand the tree’s distribution in the field.

1. Introduction

Agriculture is one of the oldest and most fundamental human major fields. In recent years, it has become more important because it provides the basic necessities for living. Thanks to new technologies, including satellite images and artificial intelligence, agriculture is becoming smart. Many tasks have been made easier for farmers, and the yield is increasing in quality and quantity. Among the crops that can be studied are tree crops, and one of the essential components is the canopy [1,2].
The canopy is the aerial part of a plant community or the top of a crop, formed by plant crowns with emergent trees and shade trees. For forests, the canopy refers to the top layer or habitat area created by the mature tree’s crown. Knowing the crown’s size is essential for orchard management, decision-making, irrigation, fertilization, spraying, pruning, and yield estimation [3]. To obtain the tree’s crown size information from an image, it is necessary to use segmentation techniques [4]. The latter presents many challenges, such as the types of images to be used, the appropriate algorithm to considered, the technique of validating the results, etc. For example, many projects use unmanned aerial vehicles (UAV) [5], while others use high-resolution satellites images [6]. However, each of them has some advantages and disadvantages. The UAV or satellite images selection depends mainly on resolution, historical data, cost, etc. In addition, the overlap of trees is a significant factor that influences segmentation because, in most of these cases, the crowns cannot be separated from each other. For example, some citrus varieties such as mandarin and citrange have a crown diameter between three and five meters [7]. So, if the distance between two trees is less than three meters between them, they will overlap.
In general, we can distinguish three types of orchards: the first (Figure 1a) are separate trees (not overlapping), the second (Figure 1b) are trees that form rows (semi-overlapping), and third (Figure 1c), trees that are very condensed (fully overlapping).
Technically, tree canopy classification and segmentation based on remote sensing images is one of the main challenges in smart agriculture [8,9,10,11,12]. The complexity of this challenge is different from one orchard to another. It depends on the trees distribution in the orchard, the size of the canopies, and the images’ resolution. In some cases, the trees do not overlap in the field, meaning they can be directly detected and segmented using object segmentation algorithms [13,14,15,16,17]. However, sometimes the trees overlap, making it difficult for the algorithms to segment each one individually. In this case, there are other solutions, such as segmenting rows or the entire parcel canopy. Thus, these segmentations require high-resolution images to obtain good results (a minimum of 0.7 m in spatial resolution) [10,18,19,20,21,22,23].
Currently, there are a lot of approaches and techniques to achieve object segmentation. For example, some classical algorithms follow the thresholding technique by taking the image in grayscale [24]. There are clustering algorithms that follow unsupervised learning by dividing data into several groups (clusters) with common characteristics between them (similarity), based on several mathematical metrics such as Euclidean distance, Jaccard coefficient, value cosine, etc. [25]. In addition, advanced machine learning algorithms such as deep learning [26,27] train on data, which means they need several images with their ground truths (labels) to have a good result [28].
There are two tree segmentation cases in the literature: the non-overlapping orchards and the overlapping orchards. The non-overlapping orchard case has attracted the attention of many scientific researchers for a long time. In 1996, Ref. [29] tried to calculate the canopy area of tomato plants from RGB images using machine vision techniques. Their segmentation process adopted a thresholding technique, followed by an Otsu method, and extended it to an efficient multimode operation. Moreover, the canopy area determined by machine vision was compared to those measured by human operators using interactive video tracing to obtain the results. Another study [30] extracted Gabor filter texture features from remote sensing images and applied K-means [31] clustering analysis to extract the tree’s crown. After that, they used some morphological operations to reduce false detection. Their approach has a score of 0.79 F-measure. Valeriano et al. [32] tried to estimate the orange yield based on laser scanner data and the k-means algorithm. Their system used a laser scanner to obtain the point belonging to the tree’s crown.
After processing the data, they got a 3D image of each tree which was used to input the k-means algorithm. The model counts the fruits and estimates their diameter by counting the pixels of each fruit. Finally, they compared the fruit quantity obtained by the algorithm with the actual amount in the field, and they got a score of 0.88 R2 (coefficient of determination). Zhao et al. [33] used a UAV to collect a large dataset of pomegranate tree images and segmented the canopy using deep learning. The study was done in an orchard with non-overlapping trees, and they tested two architectures, mask-rcnn [34] and Unet [35], and got 0.98 and 0.61 (recall), respectively. To sum up, several studies attempt to segment the tree canopy in a non-overlapping orchard with good results that do not exceed an error of 10% [16,18,36,37].
On the other hand, it is challenging to distinguish each tree individually using remote sensing images in overlapping orchards. Because sometimes the trees are in rows (semi-overlapping), while in other cases, the canopy covers all the parcels, which means it is impossible to see the soil in the image (fully overlapping). Muhammad Moshiur et al. [38] used high-resolution satellite images to predict the mono yield in a semi-overlapping orchard and showed that canopy size is strongly correlated with it. In this process, the parcel’s canopy was segmented using the thresholding method to predict the yield. When only spectral information was used, a low score of 0.3 R2 was found, however, when the canopy size information was added, the results improved to 0.93 R2, which means that the segmentation was excellent and robust.
Paolo et al. [39] took a vineyard orchard and tried to segment the canopy to optimize resources, pruning, and plant health monitoring. They used UAVs to collect the multispectral vineyard images, which had several rows. In a visual comparison between some clustering algorithms such as Digital Elevation Model (DEM) [40], and K-means [41], they found that the k-means algorithm is more precise in canopy segmentation than the other algorithms. HyMap aerial hyperspectral images and light sensing and ranging (LiDAR) data were used by [42] to classify urban forest tree species. After the exhaustive pre-processing techniques, an image with a spatial resolution of three meters was obtained and used to segment the trees using the object-based image analysis algorithm (OBIA) [43], The validation was made with an ortho-image with a spatial resolution of 0.09 m, and a good segmentation was found. Afterward, Random Forest [44] and Multi-Class [45] algorithms were used to classify the tree species, and an overall accuracy of 87.0% and 88.9%, was reached, respectively.
Usually, when we have an overlapping orchard, it is challenging to identify each tree individually in the canopy. However, the trees rows or the total parcel canopy can be segmented [46,47].
To sum up, the orchard type (overlapping or non-overlapping) influences the segmentation verification and validation results. Since obtaining a scientific score in overlapping trees is difficult, most studies used visual or field comparisons to validate their models. On the other hand, in non-overlapping trees, the images can be labeled manually, and the model is trained to compare the model’s results with the labels to have a scientific score. This is why supervised learning, such as deep learning algorithms, was used [48].
So, this paper aims to present a new approach for extracting the tree’s canopy and classifying them according to their size in overlapping orchards.

2. Materials and Methods

2.1. Study Region

Our studied region is Ouargha Orchard, located in Morocco, and it is the property of Les Domaines Agricoles, a Moroccan company. This orchard (Figure 2) contains citrus trees with a variety of tangerine. The trees are presented in several parcels, and the rows are clearly defined. However, the orchard is classified as semi-overlapping because it is difficult to separate each tree crown.

2.2. Data

The remote sensing field used images of the Earth taken by artificial satellites. In agriculture, satellite images are essential data sources. This data consists of multispectral or hyperspectral band information in several spectral resolutions [49].
In this study, we used the Moroccan Mohammed VI satellite [50], which has two satellites, A and B, launched on 8 November 2017, and 21 November 2018, respectively. One of the primary satellite objectives is to develop the agriculture field in Morocco. Mohammed VI satellite system collects multispectral images in several bands: 450–530 nm (blue), 510–590 nm (green), 620–700 nm (red), and 775–915 nm (near-infrared). This satellite is similar to the Pléiades satellite, which has a 0.5 m spatial resolution.
The OUARGHA orchard has 25 parcels ranging from 30,000 to 80,000 m2, with thousands of semi-overlapping trees. So, to carry out this research, an image captured by the Mohammed VI satellite in 2020 (Figure 3) was used and clipped by each parcel using open-source software QGIS 3.16.3 [51].
Technically we used the ’create new shapefile’ extension to create a kml file of the parcel, and we split it using the ’clip raster by mask layer’ extension.
In Figure 3, you can see an image of the study region captured by the Mohammed VI satellite.

2.3. Our Approach

In this study, we are interested in a semi-overlapping orchard, which means that trees are presented as rows with a uniform distribution, which means that there is the same distance between all tree trunks and between rows. Our methodology used satellite images and some field measurements and consists of five steps presented in (Figure 4), which are: (1) segment the rows, (2) locate the trees, (3) create a box for each location, (4) estimate the canopy inside the boxes, and finally (5) classify these boxes according to the canopy size. Each of the steps are described in further detail.

2.3.1. Tree’s Rows Segmentation

Tree segmentation is an important technique that helps farmers count their trees, get an idea about the yield based on the canopy size, and control the orchard’s health. However, this approach presents many difficulties in overlapping trees. In our study, the field contains several parcels with very condensed rows, which makes it challenging to segment each tree canopy individually. However, the rows can be identified and counted. So, we used the high spatial resolution Mohammed VI satellite images [50] to do this process. In fact, we tried two segmentation approaches using thresholding and clustering algorithms. Visually the thresholding algorithm did not get a good result, mainly because not all of the image bands were used. Otherwise, the clustering algorithms, especially k-means, presented a better result. To construct the model, we start by extracting the parcel from the original image. This technique gives each parcel an image with three pieces of information: trees, soil, and the space outside it, which is the image’s background. An example of two parcel images is presented in (Figure 3). Based on the k-means algorithm, our model takes all the image’s pixels and creates several clusters based on the K number. This k represents the number of centroids that the algorithm will develop—the cluster’s number. The elements in each of them have a small euclidean distance, which means they are very similar.
In our case, we use K = 3 to get three clusters (trees, soil, and outside parcel). The results obtained are an image with one band (Figure 5). We excluded the part outside the box to get a binary mask image, with pixels of 0 and 1 values that represent the soil and the trees.

2.3.2. Tree Localization

In tree crops, the planting process usually follows uniform distances [52]. In our study area, the orchard has rows of trees with six meters between them and two meters between the tree trunks inside each row. In addition, the size of the canopies is heterogeneous, with large and small canopies and sometimes with weak or missing trees. The satellite spatial resolution of 50 cm means that each pixel in the image is a square with a side of half a meter. Thus, a distance of two meters between trees inside the rows equals four pixels in the image, while six-meter lengths equal 12 pixels (Figure 6).
With this information, a semi-automatic algorithm was created to help us obtain the distribution of the trees’ locations in the image. The model takes three tree trunk positions as input, the first two positions of the first row and the first position of the second row, the distance between the trees inside the rows, and the distance between the rows. The algorithm loops on all pixels of the images and locates the center positions of the trees. Finally, when we compared the tree’s positions number in the image by the tree’s number in the orchard, we found that they are the same, which means that the algorithm considers all the tree’s positions in the image precisely (Figure 7).

2.3.3. Tree Crown Detection

As shown in (Figure 5), it is difficult to separate each tree inside the rows. So, our solution was to build boxes that surround the tree’s location, and with them, get the information about each tree canopy. The first step was to find the best box sizes. In the field, trees presented a circle shape, which means a square box could be used with some caveats. In the orchard, the maximum diameter of a tree is three meters, and the distance between them is two meters. Thus, there is an overlapping of one meter between every two trees. Based on this measure, we used boxes of three meters, which have a common part between them (Figure 8).

2.3.4. Tree Classification by Size

After locating the trees and creating the boxes, we moved to get the crowns by counting the pixels that describe the canopy. The model loops through all the boxes and outputs a CSV file containing each crown size and its identifier. We have created three tree classes: missing or weak, normal, and big canopies to make a classification. The canopy sizes of the trees in each one of them is calculated using the formulas follow:
class 1: min _ value canopy   size < min _ value + avg _ value / 5
class 2: min _ value + avg _ value / 5 canopy   size < max _ value avg _ value / 5
class 3: max _ value avg _ value / 5 canopy   size max _ value
min_value = the minimum canopy in the parcel
max_value = the maximum canopy in the parcel
avg_value = the tree’s canopies average in the parcel

3. Results and Validation

We took an empty image with zero in all pixels to map the classification results. Each box location was defined, so after classifying each tree, we assigned a fixed number to all the pixels of the same box in the empty image. Finally, a matrix with four pixels of value, the three classes, and the background was obtained and colored by a specific color to visualize the results as a single map (Figure 9).
We counted the segmented rows to compare them with the absolute number of rows in the field to validate the results. After having visited all the parcels, we found the same number of rows. The next step was to validate the segmentation algorithm results. In this process, five plots with different varieties were visited in the field, with 24 of them representing 20%. Then, three samples in several positions were taken from each parcel seen in the orchards (Figure 10). Finally, we obtained 15 samples manually labeled and used them to validate the segmentation.
Finally, the results obtained by the algorithm and the mask that we manually created were compared. The metrics used were the F1 and Precision score.
Precision = TP TP + FP
F 1 - score = TP TP + 1 2 ( FP + FN )
TP = true positives;
FP = false positives;
FN = false negatives.
The scores obtained for the five parcels are presented in (Table 1).
The results obtained during the segmentation in the five parcels were, on average, 0.93 F1-score. This part tried to compare the map obtained with the actual distribution of trees in the orchards. Indeed, it is not easy to visit all the trees and parcels because the orchard is extensive. However, we took a random tree’s sample in different parcels to validate the results.
To validate a random tree, we located it in the resulting map with the row number to which it belongs and its order in this row to quickly facilitate their visit to the field. After visiting many trees in multiples locations with several canopy sizes, (see some examples in Figure 11), we compared the map results with the field measurements, and good accuracy in the model was found.
In (Figure 11), the blue circle means a normal tree, the yellow a large tree, and the red a missing or weak tree. It is difficult to distinguish between weak and absent trees with this approach because, in the missing tree location, the side trees find space to branch out, which means that there is a high probability that the box contains a canopy from the neighboring trees. Finally, based on this approach, farmers can easily monitor the situation of an overlapping orchard by identifying the places that contain weak, missing, and large trees.

4. Conclusions

Remote sensing has been used to identify tree crops without overlaps for a long time. However, when there is some tree overlapping in the orchards, a new techniques must be used. This study proposes a new approach to localize and classify trees using high-resolution satellite images and field measures in an overlapping orchard. The results were presented in a map with the distribution of the trees organized by crown size. Our methodology can be used in several other tree crops in Morocco or worldwide. However, this requires high-resolution remote sensing images, with 0.7 m at minimum, and the pattern distances between trees must be uniform. Therefore, our approach can be an alternative solution to tree monitoring problems found in overlapping orchards. For example, the vegetation mapping for each tree could be obtained by combining the map of the tree obtained by our model and the vegetation indexes obtained from the spectral information, similarly for water stress. Our approach can also provide data such as the number of trees missing, normal and big trees in each row, the total canopy of the parcel, the vegetation by tree class, etc. Finally, it is also essential to mention that this data can estimate the yield, which is a precision challenge in agriculture.

Author Contributions

Conceptualization, A.M., S.E.F. and Y.Z.; methodology, A.M., S.E.F. and Y.Z.; software, A.M.; validation, A.M., S.E.F. and Y.Z.; formal analysis, A.M., S.E.F. and Y.Z.; investigation, A.M.; resources, Y.Z.; data curation, A.M., S.E.F. and Y.Z.; writing—original draft preparation, A.M.; writing—review and editing, A.M., S.E.F. and Y.Z.; visualization, A.M.; supervision, S.E.F. and Y.Z.; project administration, Y.Z.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Hassan II Academy of Science and Technology under the project entitled "multispectral satellite imagery, data mining, and agricultural applications".

Acknowledgments

We acknowledge the Moroccan company Les Domaines Agricole for letting us use the Ouargha orchard as a study region for this research.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Mazumdar, S.; Dunshea, A.; Chong, S.; Jalaludin, B. Tree Canopy Cover Is Best Associated with Perceptions of Greenspace: A Short Communication. Int. J. Environ. Res. Public Health 2020, 17, 6501. [Google Scholar] [CrossRef]
  2. Chen, J.; Jin, S.; Du, P. Roles of horizontal and vertical tree canopy structure in mitigating daytime and nighttime urban heat island effects. Int. J. Appl. Earth Obs. Geoinf. 2020, 89, 102060. [Google Scholar] [CrossRef]
  3. Goodman, R.C.; Phillips, O.L.; Baker, T.R. The importance of crown dimensions to improve tropical tree biomass estimates. Ecol. Appl. 2014, 24, 680–698. [Google Scholar] [CrossRef] [Green Version]
  4. Ke, Y.; Quackenbush, L.J. A review of methods for automatic individual tree-crown detection and delineation from passive remote sensing. Int. J. Remote Sens. 2011, 32, 4725–4747. [Google Scholar] [CrossRef]
  5. Rahman, M.F.F.; Fan, S.; Zhang, Y.; Chen, L. A Comparative Study on Application of Unmanned Aerial Vehicle Systems in Agriculture. Agriculture 2021, 11, 22. [Google Scholar] [CrossRef]
  6. Robson, A.; Rahman, M.M.; Muir, J. Using worldview satellite imagery to map yield in Avocado (Persea americana): A case study in Bundaberg, Australia. Remote Sens. 2017, 9, 1223. [Google Scholar] [CrossRef] [Green Version]
  7. Tazima, Z.H.; Neves, C.S.V.J.; Yada, I.F.U.; Júnior, R.P.L. Performance of ‘Okitsu’satsuma mandarin trees on different rootstocks in Northwestern Paraná State. Semin. Ciênc. Agrár. 2014, 35, 2297–2308. [Google Scholar] [CrossRef] [Green Version]
  8. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef] [Green Version]
  9. Hamuda, E.; Glavin, M.; Jones, E. A survey of image processing techniques for plant extraction and segmentation in the field. Comput. Electron. Agric. 2016, 125, 184–199. [Google Scholar] [CrossRef]
  10. Dyson, J.; Mancini, A.; Frontoni, E.; Zingaretti, P. Deep learning for soil and crop segmentation from remotely sensed data. Remote Sens. 2019, 11, 1859. [Google Scholar] [CrossRef] [Green Version]
  11. Maddikunta, P.K.R.; Hakak, S.; Alazab, M.; Bhattacharya, S.; Gadekallu, T.R.; Khan, W.Z.; Pham, Q.V. Unmanned aerial vehicles in smart agriculture: Applications, requirements, and challenges. IEEE Sens. J. 2021. [Google Scholar] [CrossRef]
  12. Laliberte, A.S.; Goforth, M.A.; Steele, C.M.; Rango, A. Multispectral remote sensing from unmanned aircraft: Image processing workflows and applications for rangeland environments. Remote Sens. 2011, 3, 2529–2551. [Google Scholar] [CrossRef] [Green Version]
  13. Karydas, C.; Gewehr, S.; Iatrou, M.; Iatrou, G.; Mourelatos, S. Olive plantation mapping on a sub-tree scale with object-based image analysis of multispectral UAV data; Operational potential in tree stress monitoring. J. Imaging 2017, 3, 57. [Google Scholar] [CrossRef] [Green Version]
  14. Cheng, H.; Damerow, L.; Sun, Y.; Blanke, M. Early yield prediction using image analysis of apple fruit and tree canopy features with neural networks. J. Imaging 2017, 3, 6. [Google Scholar] [CrossRef]
  15. Weinstein, B.G.; Marconi, S.; Bohlman, S.; Zare, A.; White, E. Individual tree-crown detection in RGB imagery using semi-supervised deep learning neural networks. Remote Sens. 2019, 11, 1309. [Google Scholar] [CrossRef] [Green Version]
  16. Csillik, O.; Cherbini, J.; Johnson, R.; Lyons, A.; Kelly, M. Identification of citrus trees from unmanned aerial vehicle imagery using convolutional neural networks. Drones 2018, 2, 39. [Google Scholar] [CrossRef] [Green Version]
  17. Ok, A.; Ozdarici-Ok, A. Detection of citrus trees from UAV DSMs. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 4, 27. [Google Scholar] [CrossRef] [Green Version]
  18. Maschler, J.; Atzberger, C.; Immitzer, M. Individual tree crown segmentation and classification of 13 tree species using airborne hyperspectral data. Remote Sens. 2018, 10, 1218. [Google Scholar] [CrossRef] [Green Version]
  19. Li, W.; Fu, H.; Yu, L.; Cracknell, A. Deep learning based oil palm tree detection and counting for high-resolution remote sensing images. Remote Sens. 2017, 9, 22. [Google Scholar] [CrossRef] [Green Version]
  20. Zortea, M.; Macedo, M.M.; Mattos, A.B.; Ruga, B.C.; Gemignani, B.H. Automatic citrus tree detection from UAV images based on convolutional neural networks. In Proceedings of the 31th Sibgrap/WIA—Conference on Graphics, Patterns and Images, SIBGRAPI’18, Foz do Iguacu, Brazil, 29 October–1 November 2018. [Google Scholar]
  21. Ampatzidis, Y.; Partel, V. UAV-based high throughput phenotyping in citrus utilizing multispectral imaging and artificial intelligence. Remote Sens. 2019, 11, 410. [Google Scholar] [CrossRef] [Green Version]
  22. Abdollahnejad, A.; Panagiotidis, D.; Surovỳ, P. Estimation and extrapolation of tree parameters using spectral correlation between UAV and Pléiades data. Forests 2018, 9, 85. [Google Scholar] [CrossRef] [Green Version]
  23. Alganci, U.; Sertel, E.; Kaya, S. Determination of the olive trees with object based classification of Pleiades satellite image. Int. J. Environ. Geoinform. 2018, 5, 132–139. [Google Scholar] [CrossRef]
  24. Lee, S.U.; Chung, S.Y.; Park, R.H. A comparative performance study of several global thresholding techniques for segmentation. Comput. Vis. Graph. Image Process. 1990, 52, 171–190. [Google Scholar] [CrossRef]
  25. Ghahramani, Z. Unsupervised learning. In Summer School on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2003; pp. 72–112. [Google Scholar]
  26. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Garcia-Rodriguez, J. A review on deep learning techniques applied to semantic segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  27. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image segmentation using deep learning: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021. [Google Scholar] [CrossRef] [PubMed]
  28. Zaitoun, N.M.; Aqel, M.J. Survey on image segmentation techniques. Procedia Comput. Sci. 2015, 65, 797–806. [Google Scholar] [CrossRef] [Green Version]
  29. Ling, P.; Ruzhitsky, V. Machine vision techniques for measuring the canopy of tomato seedling. J. Agric. Eng. Res. 1996, 65, 85–95. [Google Scholar] [CrossRef]
  30. Huihui, S.; Ni, W.; Wenxiu, T. Tree canopy extraction method of high resolution image based on Gabor filter and morphology. J. Geo-Inf. Sci. 2019, 21, 249–258. [Google Scholar]
  31. Likas, A.; Vlassis, N.; Verbeek, J.J. The global k-means clustering algorithm. Pattern Recognit. 2003, 36, 451–461. [Google Scholar] [CrossRef] [Green Version]
  32. Méndez, V.; Pérez-Romero, A.; Sola-Guirado, R.; Miranda-Fuentes, A.; Manzano-Agugliaro, F.; Zapata-Sierra, A.; Rodríguez-Lizana, A. In-Field Estimation of Orange Number and Size by 3D Laser Scanning. Agronomy 2019, 9, 885. [Google Scholar] [CrossRef] [Green Version]
  33. Zhao, T.; Yang, Y.; Niu, H.; Wang, D.; Chen, Y. Comparing U-Net convolutional network with mask R-CNN in the performances of pomegranate tree canopy segmentation. In Multispectral, Hyperspectral, and Ultraspectral Remote Sensing Technology, Techniques and Applications VII; International Society for Optics and Photonics: Bellingham, WA, USA, 2018; Volume 10780, p. 107801J. [Google Scholar]
  34. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Cambridge, MA, USA, 20–23 June 2017; pp. 2961–2969. [Google Scholar]
  35. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  36. Liu, L.; Cheng, X.; Dai, J.; Lai, J. Adaptive threshold segmentation for cotton canopy image in complex background based on logistic regression algorithm. Trans. Chin. Soc. Agric. Eng. 2017, 33, 201–208. [Google Scholar]
  37. Chen, Y.; Hou, C.; Tang, Y.; Zhuang, J.; Lin, J.; He, Y.; Guo, Q.; Zhong, Z.; Lei, H.; Luo, S. Citrus tree segmentation from UAV images based on monocular machine vision in a natural orchard environment. Sensors 2019, 19, 5558. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Rahman, M.M.; Robson, A.; Bristow, M. Exploring the potential of high resolution worldview-3 Imagery for estimating yield of mango. Remote Sens. 2018, 10, 1866. [Google Scholar] [CrossRef] [Green Version]
  39. Cinat, P.; Di Gennaro, S.F.; Berton, A.; Matese, A. Comparison of unsupervised algorithms for Vineyard Canopy segmentation from UAV multispectral images. Remote Sens. 2019, 11, 1023. [Google Scholar] [CrossRef] [Green Version]
  40. Mukherjee, S.; Joshi, P.K.; Mukherjee, S.; Ghosh, A.; Garg, R.; Mukhopadhyay, A. Evaluation of vertical accuracy of open source Digital Elevation Model (DEM). Int. J. Appl. Earth Obs. Geoinf. 2013, 21, 205–217. [Google Scholar] [CrossRef]
  41. Krishna, K.; Murty, M.N. Genetic K-means algorithm. IEEE Trans. Syst. Man, Cybern. Part B 1999, 29, 433–439. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Zhang, Z.; Kazakova, A.; Moskal, L.M.; Styers, D.M. Object-based tree species classification in urban ecosystems using LiDAR and hyperspectral data. Forests 2016, 7, 122. [Google Scholar] [CrossRef] [Green Version]
  43. Blaschke, T. Object based image analysis for remote sensing. ISPRS J. Photogramm. Remote Sens. 2010, 65, 2–16. [Google Scholar] [CrossRef] [Green Version]
  44. Liaw, A.; Wiener, M. Classification and regression by randomForest. R News 2002, 2, 18–22. [Google Scholar]
  45. Hastie, T.; Rosset, S.; Zhu, J.; Zou, H. Multi-class adaboost. Stat. Its Interface 2009, 2, 349–360. [Google Scholar] [CrossRef] [Green Version]
  46. Tu, Y.H.; Johansen, K.; Phinn, S.; Robson, A. Measuring canopy structure and condition using multi-spectral UAS imagery in a horticultural environment. Remote Sens. 2019, 11, 269. [Google Scholar] [CrossRef] [Green Version]
  47. Wang, L.; Gong, P.; Biging, G.S. Individual tree-crown delineation and treetop detection in high-spatial-resolution aerial imagery. Photogramm. Eng. Remote Sens. 2004, 70, 351–357. [Google Scholar] [CrossRef] [Green Version]
  48. Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Volume 1. [Google Scholar]
  49. Soproni, D.; Maghiar, T.; Molnar, C.; Hathazi, I.; Bandici, L.; Arion, M.; Krausz, A. Study of Electromagnetic Properties of the Agricultural Products. J. Electr. Electron. Eng. 2008, 1, 130–133. [Google Scholar]
  50. El-Harti, A.; Bannari, A.; Manyari, Y.; Nabil, A.; Lahboub, Y.; El-Ghmari, A.; Bachaoui, E. Capabilities of the New Moroccan Satellite Mohammed-VI for Planimetric and Altimetric Mapping. In Proceedings of the IGARSS 2020—2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 6105–6108. [Google Scholar]
  51. QGIS, D.T. QGIS Geographic Information System, Open Source Geospatial Foundation. Available online: http://qgis.osgeo.org (accessed on 31 August 2015).
  52. Lavee, S.; Haskal, A.; Avidan, B. The effect of planting distances and tree shape on yield and harvest efficiency of cv. Manzanillo table olives. Sci. Hortic. 2012, 142, 166–173. [Google Scholar] [CrossRef]
Figure 1. Different tree’s distribution in an orchard: (a) non-overlapping, (b) semi-overlapping, and (c) fully overlapping.
Figure 1. Different tree’s distribution in an orchard: (a) non-overlapping, (b) semi-overlapping, and (c) fully overlapping.
Jimaging 07 00241 g001
Figure 2. OUARGHA orchard - LES DOMAINES AGRICOLES (citrus trees).
Figure 2. OUARGHA orchard - LES DOMAINES AGRICOLES (citrus trees).
Jimaging 07 00241 g002
Figure 3. The false color (R,G,B,NIR) of the OUARGHA orchard obtained by the Mohammed VI satellite in (23 July 2020). (a,b) two examples of parcels. Projected Coordinate System: EPSG:26191—Merchich/Nord Maroc; Projection: Transverse Mercator; Linear Unit: Meter; Datum: Merchich; Prime Meridian: Greenwich; Angular Unit: Degree.
Figure 3. The false color (R,G,B,NIR) of the OUARGHA orchard obtained by the Mohammed VI satellite in (23 July 2020). (a,b) two examples of parcels. Projected Coordinate System: EPSG:26191—Merchich/Nord Maroc; Projection: Transverse Mercator; Linear Unit: Meter; Datum: Merchich; Prime Meridian: Greenwich; Angular Unit: Degree.
Jimaging 07 00241 g003
Figure 4. Flowchart of the methodology.
Figure 4. Flowchart of the methodology.
Jimaging 07 00241 g004
Figure 5. Example of tree identification using thresholding and clustering (k-means) segmentation.
Figure 5. Example of tree identification using thresholding and clustering (k-means) segmentation.
Jimaging 07 00241 g005
Figure 6. Trees distribution in the field.
Figure 6. Trees distribution in the field.
Jimaging 07 00241 g006
Figure 7. Tree localization and distribution in the parcel.
Figure 7. Tree localization and distribution in the parcel.
Jimaging 07 00241 g007
Figure 8. Tree crown detection using the boxes distribution in the parcel.
Figure 8. Tree crown detection using the boxes distribution in the parcel.
Jimaging 07 00241 g008
Figure 9. Tree classification by size in the parcel.
Figure 9. Tree classification by size in the parcel.
Jimaging 07 00241 g009
Figure 10. Examples of the samples, maks, and the segmented image results.
Figure 10. Examples of the samples, maks, and the segmented image results.
Jimaging 07 00241 g010
Figure 11. Comparison between tree classification in the map and field validation.
Figure 11. Comparison between tree classification in the map and field validation.
Jimaging 07 00241 g011
Table 1. Segmentation scores.
Table 1. Segmentation scores.
ParcelF1-ScorePrecision
Parcel 10.930.97
Parcel 20.910.95
Parcel 30.930.96
Parcel 40.930.97
Parcel 50.940.98
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Moussaid, A.; Fkihi, S.E.; Zennayi, Y. Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms. J. Imaging 2021, 7, 241. https://doi.org/10.3390/jimaging7110241

AMA Style

Moussaid A, Fkihi SE, Zennayi Y. Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms. Journal of Imaging. 2021; 7(11):241. https://doi.org/10.3390/jimaging7110241

Chicago/Turabian Style

Moussaid, Abdellatif, Sanaa El Fkihi, and Yahya Zennayi. 2021. "Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms" Journal of Imaging 7, no. 11: 241. https://doi.org/10.3390/jimaging7110241

APA Style

Moussaid, A., Fkihi, S. E., & Zennayi, Y. (2021). Tree Crowns Segmentation and Classification in Overlapping Orchards Based on Satellite Images and Unsupervised Learning Algorithms. Journal of Imaging, 7(11), 241. https://doi.org/10.3390/jimaging7110241

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop