Next Article in Journal
TPE-RBF-SVM Model for Soybean Categories Recognition in Selected Hyperspectral Bands Based on Extreme Gradient Boosting Feature Importance Values
Next Article in Special Issue
Investigating Sentinel-1 and Sentinel-2 Data Efficiency in Studying the Temporal Behavior of Wheat Phenological Stages Using Google Earth Engine
Previous Article in Journal
Effects of Conservation Tillage on Soil Properties and Maize Yield in Karst Regions, Southwest China
Previous Article in Special Issue
Combination of Sentinel-2 Satellite Images and Meteorological Data for Crop Water Requirements Estimation in Intensive Agriculture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Segmentation and Stratification Methods of Field Maize Terrestrial LiDAR Point Cloud

1
College of Resources and Environment, Huazhong Agricultural University, Wuhan 430070, China
2
College of Informatics, Huazhong Agricultural University, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Agriculture 2022, 12(9), 1450; https://doi.org/10.3390/agriculture12091450
Submission received: 13 July 2022 / Revised: 18 August 2022 / Accepted: 12 September 2022 / Published: 13 September 2022
(This article belongs to the Special Issue Recent Advances in Agro-Geoinformatics)

Abstract

:
Three-dimensional (3D) laser point cloud technology is an important research method in the field of agricultural remote sensing research. The collection and processing technology of terrestrial light detection and ranging (LiDAR) point cloud of crops has greatly promoted the integration of agricultural informatization and intelligence. In a smart farmland based on 3D modern agriculture, the manager can efficiently and conveniently achieve the growth status of crops through the point cloud collection system and processing model integrated in the smart agricultural system. To this end, we took field maize as the research object in this study and processed four sets of field maize point clouds, named Maize-01, Maize-02, Maize-03, and Maize-04, respectively. In this research, we established a field individual maize segmentation model with the density-based clustering algorithm (DBSCAN) as the core, and four groups of field maize were used as research objects. Among them, the value of the overall accuracy (OA) index, which was used to evaluate the comprehensive performance of the model, were 0.98, 0.97, 0.95, and 0.94. Secondly, the multi-condition identification method was used to separate different maize organ point clouds from the individual maize point cloud. In addition, the organ stratification model of field maize was established. In this organ stratification study, we take Maize-04 as the research object and obtained the recognition accuracy rates of four maize organs: tassel, stalk, ear, and leaf at 96.55%, 100%, 100%, and 99.12%, respectively. We also finely segmented the leaf organ obtained from the above-mentioned maize organ stratification model into each leaf individual again. We verified the accuracy of the leaf segmentation method with the leaf length as the representative. In the linear analysis of predicted values of leaf length, R 2 was 0.73, R M S E was 0.12 m, and M A E was 0.07 m. In this study, we examined the segmentation of individual crop fields and established 3D information interpretations for crops in the field as well as for crop organs. Results visualized the real scene of the field, which is conducive to analyzing the response mechanism of crop growth and development to various complex environmental factors.

1. Introduction

Maize has wide application and has high economic value as food, fodder, and industrial and pharmaceutical raw materials. Maize is the crop with the largest cultivation area in the world [1]. The 3D laser point cloud data of the target scene obtained by terrestrial laser scanning (TLS) can be used for visualization and 3D reconstruction [2,3,4,5,6,7]. In this context, TLS is widely used in maize field phenotyping. The maize phenotype information obtained from maize point cloud data is more reliable than images [8,9]. The resulting problem is that the high-density point cloud data of the target is difficult to interpret. In other words, it is of great significance to develop a method that can interpret single-plant point clouds and maize organ point clouds from maize fields.
Conducting digital agriculture research and applying laser scanning and 3D laser point cloud technology is of great significance [10,11,12]. Both can reconstruct crops and scenes in the field environment, extract phenotypic features of crops, and automate stratification management of field [4,8,13,14,15]. Lin [16] demonstrated the necessity of combining 3D laser point cloud technology with plant phenotyping research from multiple dimensions. Lin [16] focused on the advantages and challenges of LiDAR-based plant phenotyping. As described by the author, for plant phenotypes, the use of LiDAR technology can calculate the various properties of the target with the least data information, the fastest speed, and is lossless.
The traditional field crop phenotype research is restricted by factors such as time, season, site, growth cycle, and non-reproducibility. Moreover, meeting the current research on agricultural production practice is difficult. Compared with traditional plant phenotyping research, the 3D laser point cloud can improve experimental safety, shorten the experimental period, and reduce experimental costs. The 3D laser point cloud can also provide simulation methods and basic data for agricultural crop phenotyping research [17]. Many scholars at home and abroad have poured into the field of high-throughput field phenotyping based on the 3D laser point cloud [14,18,19]. For example, Miao et al. [20] conducted a series of studies in the field of maize point cloud stem and leaf segmentation and ear recognition. They used a point cloud skeleton extraction and optimization algorithm to obtain the skeleton structure of maize plants to measure the phenotypic shape of target maize plants. Then, Elnashef et al. [19] reported a segmentation model for plant organ segmentation. The model provides a new segmentation algorithm based on first- and second-order tensors to achieve a more accurate organ-level assessment. G. Wu et al. [21] improved the segmentation and recognition accuracy of fruit point cloud based on hue, saturation, value (HSV) color model, and viewpoint feature histogram (VFH) descriptor in the study. Therefore, high-throughput field phenotyping can be considered a new frontier in crop breeding [22].
Among the existing crop phenotyping studies based on the laser point cloud, the method of plant point cloud skeleton extraction is widely used [20,23,24,25]. However, skeleton extraction cannot be effectively applied to agricultural field phenotype research. The fundamental reason is that complex natural factors will lead to fragmentation of 3D point cloud data of target crops. The existing point cloud inpainting technology has limitations. Moreover, the broken crop organ point cloud will greatly interfere with the accuracy of the skeleton extraction algorithm [26]. It is worth mentioning that these experiments rarely take into account the versatility and efficiency of research methods in field experiments, and deliberately avoid targets that are not easy to experiment in the selection of materials [27,28,29,30].
In recent years, after Qi et al. (2016) and Qi et al. (2017) [31,32] proposed deep learning-based 3D point cloud segmentation networks PointNet and PointNet++, many researchers have applied deep learning technology combined with 3D point cloud to the field of crop research [11,33,34,35]. Among them, Y. Li et al. [11] proposed a plant point cloud segmentation technology that integrates high-throughput data acquisition and deep learning, named DeepSeg3Dmaize. This method refers to the mature PointNet to achieve stem, leaf, and organ instance segmentation. Although this type of point cloud segmentation methods is well-structured and has good performance, it is still hard to be widely applied to field phenotyping. The complex real field environment is uncontrollable in time and space [36]. In addition, the controlled conditions in the laboratory environment cannot be compared with those in the field environment. Most important studies such as plant segmentation models, high-throughput phenotypic parameter algorithms, and plant organ clustering algorithms have been validated under certain constraints or scenarios [11,20,37]. The applications in these special scenarios cannot be replicated in a field environment.
In summary, a crop segmentation model that can work smoothly in the field environment needs to be designed urgently. The model should preferably have high computational efficiency while segmenting the point cloud of farmland crops. Therefore, this paper aims to achieve the following objectives:
  • Go deep into the farmland and use reliable methods to complete the collection, registration, and cropping of field maize point clouds;
  • Explore and verify, based on 3D laser point cloud technology, the spatial morphological characteristics of field maize;
  • Create an individual maize segmentation model that can automatically identify and segment each maize plant from the scanned point cloud of the maize field;
  • Create a maize organ stratification model that can accurately segment and visualized all maize organs from the field maize point cloud.

2. Materials and Methods

2.1. Material

The research area of this experiment is located in the maize planting field (114.3667° E, 30.4742° N) in the experimental field (Figure 1). This area is located in central China and is the main producing area of high-quality maize. The breeding method of maize adopts seedling transplanting, and the point cloud data collection time is summer. To ensure the capacity scale of the dataset and the diversity of maize varieties, four maize crop fields with a moderate size were selected in the maize production base. Table 1 shows the basic attributes of four groups of maize fields point cloud data, includes Maize-01, Maize-02, Maize-03, and Maize-04 type of maize, acquisition date, maize numbers, growing stage, experimental area size, number of scan stations, number of scan points, and average position deviation.
Many experiments have shown that 3D laser scanners perform well in field scenes, and various working parameters of 3D laser scanners have also been added to various agronomic measurement variables [8,38,39]. In this experiment, Figure 2 shows the maize field point cloud obtained by FARO Focus S70 terrestrial laser scanner in the maize experimental field, named Maize-01, Maize-02, Maize-03, and Maize-04. We completed point cloud collection under natural conditions with clear weather and no wind. At the same time, in order to avoid overexposure of the camera built into the scanner, the instrument should be avoided under strong light.
The FARO Focus S70 is capable of capturing scans with up to 165 megapixels color information, and the system measurement error at 10 m was ±1 mm. The accuracy of the instrument is sufficient to restore the real field environment. Three sets of target balls and a brackets are used to assist in the field collection of the target point cloud data. Moreover, a Lenovo laptop is utilized to complete the office processing of the point cloud data. The computing configuration is Windows 10 (×64) operating system, 16 GB running memory, NVIDIA RTX 3070 Graphics card, and AMD Ryzen 7 5800H processor. Figure 2 shows the follow-up data process after obtaining the experimental data. The process mainly includes point cloud data preprocessing, individual maize segmentation model, and maize organ stratification model.
Considering that the real field environment is complex, the target distribution is relatively dense, and the occlusion is serious. To completely record the target, this study carried out 3D laser scanning from multi height and multi angle stack of the target. Then, the point cloud data obtained by all stations are stitched together. Given that the origin coordinates of the 3D point clouds scanned by the multi-site are inconsistent, they cannot be overlapped. Ordinary point cloud registration algorithms are lacking in the ability of multi-site cloud scan registration. Therefore, we used the target ball to assist in the registration in this study. By registering the scanned target ball, the point cloud scanned by each station is converted into a unified coordinate system. Table 1 shows the average positional deviation of the target ball fitting in the four groups of point cloud registration process, all of which are less than 1 mm. Then, the FARO Focus3D professional post-processing software SCENE is used for point cloud stitching and preliminary cutting.
After the original point cloud data are obtained by multi-site cloud registration, the number of point clouds in the study area can reach tens of millions, containing a large amount of redundant data. In this study, a 3D voxel grid is created for the input target point cloud data based on point cloud library (PCL, A large crossplatform open source C++ programming library). Thus, the point cloud data volume is sampled to 40% of the original. The voxel center point was replaced by the point nearest to the barycenter point in the original point cloud data, thus improving the accuracy of high point cloud data expression.

2.2. Individual Maize Segmentation Model

In this section, Maize-01, Maize-02, Maize-03, and Maize-04 are used as research data. Based on the existing maize point cloud data set, we extracted the oriented bounding box (OBB) of 300 maize plants and extracted the OBB length and width measurements. OBB determines the size and orientation of the box according to the geometry of the object itself. The bounding box does not need to be perpendicular to the coordinate axis, which is the most compact bounding box. We set each point to p i = x i , y i , z i . The position mean m and the covariance matrix A of the target point cloud p 1 , p 2 , , p n are calculated by Equations (1) and (2). The covariance matrix A is diagonalized to obtain its eigenvector, which is the coordinate axis of the OBB.
m = 1 N i = 1 n p i ,
A = c o v x , x c o v x , y c o v x , z c o v x , y c o v y , y c o v y , z c o v x , z c o v y , z c o v z , z   =   1 N i = 1 n ( x i m x ) 2 1 N i = 1 n ( x i m x ) ( y i m y ) 1 N i = 1 n ( x i m x ) ( z i m z ) 1 N i = 1 n ( x i m x ) ( y i m y ) 1 N i = 1 n ( y i m y ) 2 1 N i = 1 n ( y i m y ) ( z i m z ) 1 N i = 1 n ( x i m x ) ( z i m z ) 1 N i = 1 n ( y i m y ) ( z i m z ) 1 N i = 1 n ( z i m z ) 2 .
After the aspect ratio of the OBB of the 300 maize point clouds obtained by the above equations, the data histogram is statistically obtained, as shown in Figure 3. In the normality test results (Table 2), the kurtosis and skewness values, which describe the shape of the data distribution, are 0.113 and 0.161, respectively. The test results show that the aspect ratio of the maize point cloud OBB obeys a normal distribution, which proves that the OBB-type bounding box of maize plants is generally a cuboid. That is, as long as the 3D parameters of the OBB of an individual maize plant are determined, an individual maize plant from the field point cloud can be segmented.
According to each maize plant seed point q i , a rectangular box for individual plant segmentation is developed, wherein the rectangle is parallel to the horizontal plane. Then, a bounding box D is generated according to this rectangle, and the point cloud segmented by the bounding box is divided into point sets C i , where i represents the maize number.
Figure 4 demonstrates a method of accurately determining the plant position, taking Maize-04-No. 9 (crop number is 277) as an example. The seed point q i is located at the center of the maize stalk on the same horizontal plane. For its tracking method, in the field point cloud data, slices are made at a level 5 cm above the soil plane (Figure 4a). In Figure 4a, W represents the slice target, and two equidistant planes W l and W r are generated in the normal vector direction of W . The distance between W l and W r is the section thickness δ . Moreover, clustering is performed to cluster individual plants in the sliced point cloud. Then, outliers are filtered in the point cloud of each individual plant slice (Figure 4b). Considering that withered leaves in mature maize plants droop down, after filtering out outliers, the circular fitting should be performed on the slice point cloud (Figure 4d). Through this method, the point cloud of withered leaves is filtered, and then, the center point q i of the point cloud of each individual plant slice is calculated (Figure 4g).
The growth position of each maize plant is determined to calculate the OBB of the maize point cloud. First, the point set C i is projected to the horizontal plane, and the radius nearest neighbors (Radius-NN) research is implemented with the projection point of each seed point q i as the center. Then, the projection point within the radius of 30 cm can be obtained. Principal component analysis was used to analyze the linear distribution of projected data points. After centralization processing (Equation (3)) and maximizing Equation (4), the fitting straight line of the projected point cloud is obtained to preliminarily judge the horizontal growth direction of leaves, where u 1 is the first principal axis of the point set p 1 , p 2 , , p j , μ is the data center, x 1 , x 2 , , x j are the centralized point sets, and j is the number of points in the point set C i . Based on a large number of maize prior growth parameters, this study generates a cropbox with length, width, and height of a , b , and c , respectively. This bounding box is used for the preliminary segmentation of field maize plants. Among them, the cropbox size a of the maize plant at the flowering stage is 0.9 m, b is 0.7 m, and c is the range of point cloud elevation.
μ = 1 n j = 1 n p j x 1 , x 2 , , x j = p 1     μ , p 2     μ , , p j     μ x j · u 1 = x j T u 1 ,
1 n j = 1 n x j · u 1 2 = 1 n j = 1 n x j · u 1 2 .
Given that the number of target individuals obtained by the crop box is unknown, an algorithm that does not need to input the number of prior individuals should be selected. Moreover, in the process of individual plant segmentation, the time complexity should be kept as low as possible. Therefore, in this research, we used DBSCAN (time complexity O is n · log n , n is the number of points) algorithm to segment individuals from population crops [40,41,42]. In the part of organ stratification model, the maize organs were segmented from the point cloud of single maize plant by spectral clustering algorithm (time complexity O is n 3 ) [43].
The DBSCAN [44] algorithm can divide regions with sufficient density into clusters and find irregular-shaped clusters in a noisy point cloud space. Among them, a cluster is defined as the largest collection of point clouds with a point cloud density greater than a threshold. To quickly complete the cluster search, the KDTree module is added to the algorithm to perform the Radius-NN search [45].
The research in this section takes the mainstream algorithm as the core and constructs the above individual maize segmentation model. Among them, these algorithms have been mentioned in many studies, and the research results have proved that the algorithms have good performance.

2.3. Maize Organ Stratification Model

In this section, Maize-04 is used as research data. Among them, Maize-04 is in the Full ripening stage, where all organs of maize have basically stopped growing. When the beam of the 3D laser scanner Focus S70 used in this study scans a target object within a range of 10 m, the point cloud accuracy can be controlled within 2 mm. In addition, the data quality is sufficient to support the organ-level segmentation of crop point clouds. In this study, the spectral clustering algorithm was used to achieve individual organ clustering in maize. First, the point cloud of the maize stalk is separated to disconnect the spatial connection among the various organs of the maize plant, which will facilitate the further completion of organ clustering. Then, the clustering of individual organs was completed through the spectral clustering algorithm. Finally, the classification of clustered organs is achieved by establishing organ identification conditions.
In this section, the focus is on using spectral clustering, an unsupervised algorithm, to achieve clustering of different organs in maize. Among them, the spectral clustering algorithm is based on the K-nearest neighborhood (KNN) composition method and the Normalized cut (Ncut) method [46]. In addition, the Kmeans algorithm is used to add labels to the point cloud to complete the clustering. The spectral clustering algorithm is sensitive to parameter settings. The number of neighbors K of KNN needs to be adjusted through several experiments. Moreover, the normalization of the laplacian matrix needs to be selected through certain prior knowledge. As the time complexity of the spectral clustering algorithm is extremely high ( O n 3 ), this research completes the second filtering process on the object point cloud through the voxel filtering method in this part. This case can preserve the surface features of the object point cloud as much as possible while improving the efficiency of the spectral clustering algorithm. The number of target point clouds after voxel filtering is 10% of the initial point cloud.
The organ labeling operation can be completed by setting the following label discrimination conditions.
  • Stem segmentation
The stalk of the maize plant is thick and grows straight upward, and the scanning results of the high precision laser scanner show a relatively evident deformation of the cylindrical geometry. Therefore, directly fitting the target cylinder is not feasible. With this in mind, the stalk segmentation in this study consists of the following two steps (Figure 5). First, the linear features of the maize point cloud were identified in PCL. Moreover, the method of cylinder space segmentation was used to preliminarily separate the stalk point cloud in a single plant and the leaf point cloud that was partially connected to the stalk. A polynomial curve fit was then performed for the curved stalk in cylinder space using the RANSAC algorithm. Considering the overall cylindrical geometry of the stem, the fitting algorithm in this section adopts radius sampling.
2.
Identification of ear point sets
According to the observation, compared with the non-stem organs, the shape of the ear has conical and cylindrical features. The RANSAC algorithm is used to perform cone and cylinder fitting in the selected organ points. The point set that meets the point number threshold is identified as the ear point set.
3.
Identification of tassel and leaf point sets
The tassel organs are located at the top of the plant. Therefore, the organ at the highest point of the plant is the tassel, and the remaining organs are marked as plant leaves. All leaf point clouds in the individual maize plant form a leaf point cloud set, which can be divided into point cloud subsets of each leaf through a clustering algorithm. Figure 6 shows the final result of adding the maize organ point cloud label:
In this study, leaf length was used as a measure of leaf stratification accuracy. Figure 7 illustrates the fast calculation method of leaf length. The key points on the leaf veins are determined by leaf length path point, and the adjacent key points are linked [27,30]. Moreover, the key point connecting line is defined as the construction line l i ( i 1 , m ), and the vector sum l i of all construction lines is 0. The construction line where the key point p m of the blade tip lies is l m . The scalar sum of all construction lines l i minus l m is defined as the leaf length l l e a f (Equation (5)).
l l e a f = l i l m .

2.4. Model Evaluation Metrics

In this study, the precision rate P and the recall rate R were set to verify the accuracy of individual plant segmentation. In addition, the global accuracy of the target segmentation was calculated directly by the overall accuracy ( O A ) (Equation (6)), where n is the total number of points in the standard set. Compared with the field image data and the target photographed by the laser scanner, the point cloud of the maize plant and its organs was calibrated in PCL as the standard value to obtain the confusion matrix of the segmentation layered model. In Equations (7) and (8), true-positive ( T P ) represents a true positive example, false-positive ( F P ) represents a false positive example, and false-negative ( F N ) represents a false negative example. Figure 8 shows an example of the T P , F N , and F P decisions for the maize segmentation result. Among them, Figure 8c highlights the difference between Figure 8a,b.
O A = i = 1 n T P i = 1 n T P + F P + F N ,
P = T P T P + F P ,
R = T P T P + F N .
To comprehensively weigh the two indicators of P and R , a new indicator F-score is introduced here to comprehensively consider the harmonic value of Precision and Recall (Equation (9)). In this study, the weight coefficient β of F-score is 1.
F-score = 1 + β 2 × P × R β 2 × P + R .
In addition, the spectral clustering algorithm in the organ clustering module needs to perform sub-sampling on the point cloud. Thus, the intersection over union ( I o U ) based on the standard value and the predicted value is used as the accuracy evaluation index of organ clustering in the research. The I o U calculation is shown in Equation (10), where I   T P represents the intersection of the bounding box between the predicted value and the standard value. Moreover, U T P + F P + F N represents the union of the bounding box between the predicted and standard values.
I o U = I T P U T P + F P + F N .
In this study, the linear regression analysis method was also used to explore the linear relationship between the predicted and standard values of the plant phenotypic parameters based on the stratification model. The coefficient of determination of regression evaluation indicators, including R squared ( R 2 ), root mean squared error ( R M S E ), and mean absolute error ( M A E ), was used to quantitatively evaluate the segmentation performance of the maize organ clustering model in field scenarios (Equations (11)–(13)).
R 2 = 1 i = 1 n y i ^ y i 2 i = 1 n y i ¯ y i 2 ,
R M S E = 1 n i = 1 n y i ^ y i 2 ,
M A E = 1 n i = 1 n y i ^ y i .
Among them, y i ^ is the standard value of the parameter, y i is the predicted value of the parameter, y i ¯ is the sample mean, i = 1 n y i ^ y i 2 is the error caused by the prediction, and i = 1 n y i ¯ y i 2 is the error caused by the mean.

3. Results and Discussion

3.1. Segmentation Results and Analysis

In this research, a two-dimensional linear equation was used to represent the growth direction of maize plants. Figure 9 shows the growth trend of individual maize plants obtained based on the cylindrical space projection method.
Figure 10 shows that the prediction error of the growing trend of an individual plant during the experiment is generally less than 0.2 radian. This result indicates that the plant direction prediction in this study is more accurate, and the crop box divided based on this has higher accuracy. In Figure 10, there are a few maize plants whose growth direction prediction deviation is greater than 0.4 radian. However, based on the segmentation results in Figure 11 and Figure 12, it can be seen that the limited prediction bias did not cause significant interference to the final segmentation results of maize individual plants. For example, the prediction of the growth direction of Maize-04-No. 25 (crop number is 293) has a large deviation, and the angle between it and the standard direction vector is 41.12°. However, in the follow-up cluster model accuracy analysis, the P value and R value of No. 25 maize were 0.99 and 0.93, respectively. This phenomenon indicates that the prediction bias value of the direction within a certain range does not have much influence on the clustering accuracy.
Figure 11 depicts the target individual segmentation preview. We render T P , F P , and F N as yellow, red, and blue, respectively, and display them in the same coordinate system. F P mainly comes from the extreme adhesion phenomenon of some adjacent maize point clouds. However, the reasonable setting of the crop box in the segmentation model will control this error within a small range. F N is mainly derived from the maize point cloud in the broken state and the under-segmentation of the crop box. The former might be caused by the influence of complex field environments, such as wind blowing blades and others in the process of collecting target point cloud. The latter is because of the large span of individual maize plants in the horizontal direction, and the uniform size of the crop box will lead to a small amount of under-segmentation.
Figure 12a,b shows the precision and recall numerical statistics of the individual maize segmentation results obtained by this research method, and Figure 11c depicts the F1-score numerical statistics. The results show that the segmentation results of most single maize samples were better. Among them, the segmentation accuracy of 97.31% maize samples per plant is higher than 80%. The segmentation recall rate of 97.64% maize samples is higher than 80%. The F1-score of 99.66% individual maize samples is higher than 80%. When the weight coefficient is 1, F1-score performs well, which indicates that the individual plant segmentation model in this study can overcome the complex field environment and effectively achieve outdoor individual-maize segmentation. During the analysis of the results (Table 3), the segmentation accuracy rate of several maize plants was less than 80%, which was because of the intricate distribution of the crowded maize leaves in the space. The leaves were often stuck together. The existence of the crop box in this model can reduce the sticking of leaves to a certain extent, thereby reducing the segmentation error. This also resulted in the F1-score of only one of our 297 maize samples being less than 80%.
According to Equation (6), the O A value of Maize-01, Maize-02, Maize-03, and Maize-04 are 0.98, 0.97, 0.95, and 0.94. It is obvious that the overall clustering accuracy is relatively high. The above-mentioned accuracy analysis results all verify the feasibility and stability of the individual-maize segmentation model in this study.
Different from conventional unsupervised identification and segmentation methods, the segmentation model in this study firstly determines in which region a single maize plant is located, and separates the point cloud of this region from the whole. Finally, a single maize plant is obtained from the clustering of the point cloud in this area. This research idea fundamentally guarantees the accuracy of crop segmentation and provides a premise for the subsequent segmentation of maize organs. It is conceivable that this fast and effective segmentation method will become the preferred choice for digital agriculture platforms.

3.2. Organ Stratification Model Accuracy Analysis

Figure 13 shows the stratification visualization effect of Maize-04 leaves based on the organ stratification model. In the analysis in this section, to fully verify the accuracy of the maize organ stratification model, we evaluated the leaf organ clustering accuracy using the linear analysis results between the predicted and standard values of leaf length. The clustering accuracy of stem organs was evaluated using the accuracy rate of stem height prediction. The clustering accuracy of ear and tassel organs was evaluated using the I o U metric. Among them, the top ear of maize was used as the research object of ear organs.
After obtaining the above-mentioned four kinds of organ clustering statistics (Figure 14), we calculated the recognition accuracy of the four kinds of organs of maize in the study area. The recognition accuracy of tassel, stem, ear, and leaf was 96.55%, 100%, 100%, and 99.12%, respectively. The experimental results show that the maize organ stratification model in this section can accurately complete the task of organ segmentation for maize plants in the field environment.
As shown in the histogram (Figure 14a), the tassel of maize plant Maize-04-No. 2 (crop number is 270) is not accurately identified. Figure 15 shows the visualization of the target, which is because of the breakage of the tip of the second maize plant, resulting in confusion between the first leaf and the tassel. This phenomenon once again confirms the difficulty of crop organ stratification in natural field environments.
In the histogram of the I o U value for ear recognition (Figure 14b), all ears were correctly identified, and all had a high segmentation accuracy. This case shows that the organ segmentation model should be able to accurately locate ear targets in maize fields in practical applications.
In the histogram of maize stalk height prediction errors (Figure 14c), the prediction errors of stalk height are generally less than 0.1 m, and only two maize plants have errors greater than 0.1 m but less than 0.16 m. Therefore, in the actual application process, this segmentation model can accurately estimate the maize stalk height.
In the scatter plot between the predicted value of leaf length and the true value (Figure 14d), an evident linear relationship exists between the predicted value of the model and the true value. The R 2 of the linear fitting result is 0.71, the R M S E is 0.12 m, and the M A E is 0.07 m. The R M S E and M A E performed well in the analysis results. The value of R 2 is mainly restricted by two factors: one is because of the missed identification of three leaves, and the other is under-clustering caused by the fragmentation of the point cloud of a few leaves.
The output accuracy of segmentation and stratification models in this study is restricted by two factors. First, the rationality of the segmentation and stratification method. Second, the data quality of the target 3D point cloud. This study only considers the performance of the segmented and stratification model. However, in the actual application process of agricultural production, a certain degree of data quality compensation is equally important. Examples include using higher-precision experimental instruments, more complete data acquisition methods, and more rigorous data repair methods. In addition, this study is limited to only a few varieties of maize and four medium-scale maize experimental field for data collection, modeling, and verification, which is slightly insufficient in terms of experimental data scale. In the future, we plan to test and update this study in larger field crop phenotypic experiments.

4. Conclusions

In this study, based on the point cloud of maize in the field, a model that can complete the individual segmentation and organ stratification of maize plants is established. After counting a large number of maize OBB aspect ratios, we found that it had obvious normal distribution characteristics. We found a reliable method for extracting this feature in this study, with an average error of only 0.06 rad in direction prediction (Section 2.2). Based on this law of maize growth, we finally established an individual maize segmentation model that can directly process field maize point clouds. In Section 3.1, the OA output values of this model for Maize-01, Maize-02, Maize-03, and Maize-04 are 0.98, 0.97, 0.95, and 0.94, respectively.
Based on the individual maize segmentation model described above, this study automatically segmented each maize point cloud into organ-level point sets, thereby proposing a maize organ stratification model. Taking Maize-04 as an example, the recognition accuracy of tassel, stem, ear, and leaf of this model was 96.55%, 100%, 100%, and 99.12%, respectively. In addition, the mean I o U of tassel segmentation output by the stratification model was 0.69, the mean I o U of ear segmentation was 0.75, and the mean error of plant height estimation was 0.03 m. In a linear regression analysis of maize leaf length estimates, R 2 is 0.71, R M S E is 0.12 m, and M A E is 0.07 m. Finally, the stratification model was also able to visualize maize organs in a hierarchical manner (Figure 13), further exploiting the advantages of 3D visualization techniques in agricultural phenotyping.
Both the individual maize segmentation model and the maize organ stratification model in the paper can run with high efficiency in ordinary computer equipment. The research results in this paper can be applied to the 3D study of maize in the field environment, thus providing support for the 3D research field of digital agriculture.

Author Contributions

Conceptualization, C.L., F.H. and R.Z.; Methodology, F.H.; software, F.H. and J.P.; Resources, C.L. and R.Z.; Data curation, F.H. and J.W.; Writing—original draft preparation, F.H.; Writing—review and editing, C.L. and R.Z.; Supervision, C.L. and R.Z.; Validation, J.W. and J.P.; Funding Acquisition, C.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (41301522), Natural Science Foundation of Hubei Province (2014CFB940), and Fundamental Research Funds for the Central Universities (Program No. 2662018JC054).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The main PCL/C++ code in this paper are available online at https://github.com/1117ismore/HZAU-Segmentation-Stratification-pcd.git [accessed on 12 July 2022]. The data are available online at https://www.scidb.cn/en/detail?dataSetId=f1c63789b8fe4832b30dccc9511d034e (accessed on 13 July 2022). As we are still conducting more research on the dataset, we will upload our dataset to the same link later.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ranum, P.; Peña-Rosas, J.P.; Garcia-Casal, M.N. Global Maize Production, Utilization, and Consumption. Ann. N. Y. Acad. Sci. 2014, 1312, 105–112. [Google Scholar] [CrossRef] [PubMed]
  2. Indirabai, I.; Nair, M.V.H.; Jaishanker, R.N.; Nidamanuri, R.R. Terrestrial Laser Scanner Based 3D Reconstruction of Trees and Retrieval of Leaf Area Index in a Forest Environment. Ecol. Inform. 2019, 53, 100986. [Google Scholar] [CrossRef]
  3. Indirabai, I.; Nair, M.V.H.; Nair, J.R.; Nidamanuri, R.R. Direct Estimation of Leaf Area Index of Tropical Forests Using LiDAR Point Cloud. Remote Sens. Appl. Soc. Environ. 2020, 18, 100295. [Google Scholar] [CrossRef]
  4. Friedli, M.; Kirchgessner, N.; Grieder, C.; Liebisch, F.; Mannale, M.; Walter, A. Terrestrial 3D Laser Scanning to Track the Increase in Canopy Height of Both Monocot and Dicot Crop Species under Field Conditions. Plant Methods 2016, 12, 9. [Google Scholar] [CrossRef]
  5. Calders, K.; Adams, J.; Armston, J.; Bartholomeus, H.; Bauwens, S.; Bentley, L.P.; Chave, J.; Danson, F.M.; Demol, M.; Disney, M.; et al. Terrestrial Laser Scanning in Forest Ecology: Expanding the Horizon. Remote Sens. Environ. 2020, 251, 112102. [Google Scholar] [CrossRef]
  6. Barbeito, I.; Dassot, M.; Bayer, D.; Collet, C.; Drössler, L.; Löf, M.; del Rio, M.; Ruiz-Peinado, R.; Forrester, D.I.; Bravo-Oviedo, A.; et al. Terrestrial Laser Scanning Reveals Differences in Crown Structure of Fagus Sylvatica in Mixed vs. Pure European Forests. For. Ecol. Manag. 2017, 405, 381–390. [Google Scholar] [CrossRef]
  7. De Conto, T.; Olofsson, K.; Görgens, E.B.; Rodriguez, L.C.E.; Almeida, G. Performance of Stem Denoising and Stem Modelling Algorithms on Single Tree Point Clouds from Terrestrial Laser Scanning. Comput. Electron. Agric. 2017, 143, 165–176. [Google Scholar] [CrossRef]
  8. Hosoi, F.; Nakabayashi, K.; Omasa, K. 3-D Modeling of Tomato Canopies Using a High-Resolution Portable Scanning Lidar for Extracting Structural Information. Sensors 2011, 11, 2166–2174. [Google Scholar] [CrossRef]
  9. Rosell, J.R.; Sanz, R. A Review of Methods and Applications of the Geometric Characterization of Tree Crops in Agricultural Activities. Comput. Electron. Agric. 2012, 81, 124–141. [Google Scholar] [CrossRef]
  10. Rose, J.; Paulus, S.; Kuhlmann, H. Accuracy Analysis of a Multi-View Stereo Approach for Phenotyping of Tomato Plants at the Organ Level. Sensors 2015, 15, 9651–9665. [Google Scholar] [CrossRef] [Green Version]
  11. Li, Y.; Wen, W.; Miao, T.; Wu, S.; Yu, Z.; Wang, X.; Guo, X.; Zhao, C. Automatic Organ-Level Point Cloud Segmentation of Maize Shoots by Integrating High-Throughput Data Acquisition and Deep Learning. Comput. Electron. Agric. 2022, 193, 106702. [Google Scholar] [CrossRef]
  12. Zhou, J.; Fu, X.; Zhou, S.; Zhou, J.; Ye, H.; Nguyen, H.T. Automated Segmentation of Soybean Plants from 3D Point Cloud Using Machine Learning. Comput. Electron. Agric. 2019, 162, 143–153. [Google Scholar] [CrossRef]
  13. Paproki, A.; Sirault, X.; Berry, S.; Furbank, R.; Fripp, J. A Novel Mesh Processing Based Technique for 3D Plant Analysis. BMC Plant Biol. 2012, 12, 63. [Google Scholar] [CrossRef]
  14. Paulus, S.; Dupuis, J.; Mahlein, A.-K.; Kuhlmann, H. Surface Feature Based Classification of Plant Organs from 3D Laserscanned Point Clouds for Plant Phenotyping. BMC Bioinform. 2013, 14, 238. [Google Scholar] [CrossRef]
  15. Mortensen, A.K.; Bender, A.; Whelan, B.; Barbour, M.M.; Sukkarieh, S.; Karstoft, H.; Gislum, R. Segmentation of Lettuce in Coloured 3D Point Clouds for Fresh Weight Estimation. Comput. Electron. Agric. 2018, 154, 373–381. [Google Scholar] [CrossRef]
  16. Lin, Y. LiDAR: An Important Tool for next-Generation Phenotyping Technology of High Potential for Plant Phenomics? Comput. Electron. Agric. 2015, 119, 61–73. [Google Scholar] [CrossRef]
  17. Comba, L.; Biglia, A.; Ricauda Aimonino, D.; Gay, P. Unsupervised Detection of Vineyards by 3D Point-Cloud UAV Photogrammetry for Precision Agriculture. Comput. Electron. Agric. 2018, 155, 84–95. [Google Scholar] [CrossRef]
  18. White, J.W.; Andrade-Sanchez, P.; Gore, M.A.; Bronson, K.F.; Coffelt, T.A.; Conley, M.M.; Feldmann, K.A.; French, A.N.; Heun, J.T.; Hunsaker, D.J.; et al. Field-Based Phenomics for Plant Genetics Research. Field Crops Res. 2012, 133, 101–112. [Google Scholar] [CrossRef]
  19. Elnashef, B.; Filin, S.; Lati, R.N. Tensor-Based Classification and Segmentation of Three-Dimensional Point Clouds for Organ-Level Plant Phenotyping and Growth Analysis. Comput. Electron. Agric. 2019, 156, 51–61. [Google Scholar] [CrossRef]
  20. Miao, T.; Zhu, C.; Xu, T.; Yang, T.; Li, N.; Zhou, Y.; Deng, H. Automatic Stem-Leaf Segmentation of Maize Shoots Using Three-Dimensional Point Cloud. Comput. Electron. Agric. 2021, 187, 106310. [Google Scholar] [CrossRef]
  21. Wu, G.; Li, B.; Zhu, Q.; Huang, M.; Guo, Y. Using Color and 3D Geometry Features to Segment Fruit Point Cloud and Improve Fruit Recognition Accuracy. Comput. Electron. Agric. 2020, 174, 105475. [Google Scholar] [CrossRef]
  22. Araus, J.L.; Cairns, J.E. Field High-Throughput Phenotyping: The New Crop Breeding Frontier. Trends Plant Sci. 2014, 19, 52–61. [Google Scholar] [CrossRef]
  23. Zhou, C.; Yang, G.; Liang, D.; Yang, X.; Xu, B. An Integrated Skeleton Extraction and Pruning Method for Spatial Recognition of Maize Seedlings in MGV and UAV Remote Images. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4618–4632. [Google Scholar] [CrossRef]
  24. Wu, S.; Wen, W.; Xiao, B.; Guo, X.; Du, J.; Wang, C.; Wang, Y. An Accurate Skeleton Extraction Approach From 3D Point Clouds of Maize Plants. Front. Plant Sci. 2019, 10, 248. [Google Scholar] [CrossRef]
  25. Jiang, A.; Liu, J.; Zhou, J.; Zhang, M. Skeleton Extraction from Point Clouds of Trees with Complex Branches via Graph Contraction. Vis. Comput. 2021, 37, 2235–2251. [Google Scholar] [CrossRef]
  26. Cao, W.; Wu, J.; Shi, Y.; Chen, D. Restoration of Individual Tree Missing Point Cloud Based on Local Features of Point Cloud. Remote Sens. 2022, 14, 1346. [Google Scholar] [CrossRef]
  27. Liu, F.; Song, Q.; Zhao, J.; Mao, L.; Bu, H.; Hu, Y.; Zhu, X. Canopy Occupation Volume as an Indicator of Canopy Photosynthetic Capacity. New Phytol. 2021, 232, 941–956. [Google Scholar] [CrossRef]
  28. Xiao, F.; Li, W.; Xiao, M.; Yang, Z.; Cheng, W.; Gao, S.; Li, G.; Ding, Y.; Paul, M.J.; Liu, Z. A Novel Light Interception Trait of a Hybrid Rice Ideotype Indicative of Leaf to Panicle Ratio. Field Crops Res. 2021, 274, 108338. [Google Scholar] [CrossRef]
  29. Qiu, Q.; Sun, N.; Bai, H.; Wang, N.; Fan, Z.; Wang, Y.; Meng, Z.; Li, B.; Cong, Y. Field-Based High-Throughput Phenotyping for Maize Plant Using 3D LiDAR Point Cloud Generated with a “Phenomobile". Front. Plant Sci. 2019, 10, 554. [Google Scholar] [CrossRef]
  30. Jin, S.; Su, Y.; Wu, F.; Pang, S.; Gao, S.; Hu, T.; Liu, J.; Guo, Q. Stem–Leaf Segmentation and Phenotypic Trait Extraction of Individual Maize Using Terrestrial LiDAR Data. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1336–1346. [Google Scholar] [CrossRef]
  31. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  32. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  33. Li, T.; Feng, Q.; Qiu, Q.; Xie, F.; Zhao, C. Occluded Apple Fruit Detection and Localization with a Frustum-Based Point-Cloud-Processing Approach for Robotic Harvesting. Remote Sens. 2022, 14, 482. [Google Scholar] [CrossRef]
  34. Schmohl, S.; Narváez Vallejo, A.; Soergel, U. Individual Tree Detection in Urban ALS Point Clouds with 3D Convolutional Networks. Remote Sens. 2022, 14, 1317. [Google Scholar] [CrossRef]
  35. Jin, S.; Su, Y.; Gao, S.; Wu, F.; Hu, T.; Liu, J.; Li, W.; Wang, D.; Chen, S.; Jiang, Y.; et al. Deep Learning: Individual Maize Segmentation From Terrestrial Lidar Data Using Faster R-CNN and Regional Growth Algorithms. Front. Plant Sci. 2018, 9, 866. [Google Scholar] [CrossRef] [PubMed]
  36. Deery, D.; Jimenez-Berni, J.; Jones, H.; Sirault, X.; Furbank, R. Proximal Remote Sensing Buggies and Potential Applications for Field-Based Phenotyping. Agronomy 2014, 4, 349–379. [Google Scholar] [CrossRef]
  37. Magney, T.S.; Eusden, S.A.; Eitel, J.U.H.; Logan, B.A.; Jiang, J.; Vierling, L.A. Assessing Leaf Photoprotective Mechanisms Using Terrestrial LiDAR: Towards Mapping Canopy Photosynthetic Performance in Three Dimensions. New Phytol. 2014, 201, 344–356. [Google Scholar] [CrossRef] [PubMed]
  38. Eitel, J.U.H.; Magney, T.S.; Vierling, L.A.; Brown, T.T.; Huggins, D.R. LiDAR Based Biomass and Crop Nitrogen Estimates for Rapid, Non-Destructive Assessment of Wheat Nitrogen Status. Field Crops Res. 2014, 159, 21–32. [Google Scholar] [CrossRef]
  39. Kjaer, K.; Ottosen, C.-O. 3D Laser Triangulation for Plant Phenotyping in Challenging Environments. Sensors 2015, 15, 13533–13547. [Google Scholar] [CrossRef]
  40. Deng, X.; Tang, G.; Wang, Q. A Novel Fast Classification Filtering Algorithm for LiDAR Point Clouds Based on Small Grid Density Clustering. Geod. Geodyn. 2022, 13, 38–49. [Google Scholar] [CrossRef]
  41. Czerniawski, T.; Sankaran, B.; Nahangi, M.; Haas, C.; Leite, F. 6D DBSCAN-Based Segmentation of Building Point Clouds for Planar Object Classification. Autom. Constr. 2018, 88, 44–58. [Google Scholar] [CrossRef]
  42. Ferrara, R.; Virdis, S.G.P.; Ventura, A.; Ghisu, T.; Duce, P.; Pellizzaro, G. An Automated Approach for Wood-Leaf Separation from Terrestrial LIDAR Point Clouds Using the Density Based Clustering Algorithm DBSCAN. Agric. For. Meteorol. 2018, 262, 434–444. [Google Scholar] [CrossRef]
  43. Sugiyama, M. Clustering. In Introduction to Statistical Machine Learning; Elsevier: Amsterdam, The Netherlands, 2016; Volume 37, pp. 447–456. [Google Scholar]
  44. Ester, M.; Kriegel, H.-P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In KDD-96 Proceedings; Simoudis, E., Ed.; AAAI Press: Palo Alto, CA, USA, 1996. [Google Scholar]
  45. Wald, I.; Havran, V. On Building Fast Kd-Trees for Ray Tracing, and on Doing That in O(N Log N). In Proceedings of the 2006 IEEE Symposium on Interactive Ray Tracing, Salt Lake City, UT, USA, 18–20 September 2006; pp. 61–69. [Google Scholar]
  46. Shi, J.; Malik, J. Normalized Cuts and Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 888–905. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Experimental scene diagram of target 3D point cloud acquisition. The picture contains the placement of the laser scanner and the target ball. (a) Maize-01; (b) Maize-02; (c) Maize-03; (d) Maize-04.
Figure 1. Experimental scene diagram of target 3D point cloud acquisition. The picture contains the placement of the laser scanner and the target ball. (a) Maize-01; (b) Maize-02; (c) Maize-03; (d) Maize-04.
Agriculture 12 01450 g001
Figure 2. Research technology roadmap.
Figure 2. Research technology roadmap.
Agriculture 12 01450 g002
Figure 3. Histogram of the aspect ratio of OBB (X-axis represents the aspect ratio of OBB per maize plant).
Figure 3. Histogram of the aspect ratio of OBB (X-axis represents the aspect ratio of OBB per maize plant).
Agriculture 12 01450 g003
Figure 4. Effect of the coordinate calibration of the Maize-04 plant. (a) Schematic diagram of point cloud slicing algorithm. (b) Top view of sliced maize plant in field; (c) Maize-04-No. 9; (d) Slice of Maize-04-No. 9; (e) 3D view of slice; (f) 3D view of stalk slice; (g) Stem spatial positioning points of Maize-04-No. 9. The blue point cloud show the slice point cloud of maize stalk obtained by fitting from the slice point cloud. The red point cloud is the center of the fitting circle. That’s where the maize stalks are.
Figure 4. Effect of the coordinate calibration of the Maize-04 plant. (a) Schematic diagram of point cloud slicing algorithm. (b) Top view of sliced maize plant in field; (c) Maize-04-No. 9; (d) Slice of Maize-04-No. 9; (e) 3D view of slice; (f) 3D view of stalk slice; (g) Stem spatial positioning points of Maize-04-No. 9. The blue point cloud show the slice point cloud of maize stalk obtained by fitting from the slice point cloud. The red point cloud is the center of the fitting circle. That’s where the maize stalks are.
Agriculture 12 01450 g004
Figure 5. Schematic diagram of the stem division method. The spatial position of the green cylinder is determined by the red point cloud in Figure 4. RANSAC algorithm was used to extract stem point cloud from point cloud in cylinder.
Figure 5. Schematic diagram of the stem division method. The spatial position of the green cylinder is determined by the red point cloud in Figure 4. RANSAC algorithm was used to extract stem point cloud from point cloud in cylinder.
Agriculture 12 01450 g005
Figure 6. Schematic diagram of Maize-04 point cloud data label.
Figure 6. Schematic diagram of Maize-04 point cloud data label.
Agriculture 12 01450 g006
Figure 7. Schematic diagram of blade length calculation.
Figure 7. Schematic diagram of blade length calculation.
Agriculture 12 01450 g007
Figure 8. Determination method of T P , F N , and F P of single maize segmentation. (a) Standard value of individual plant segmentation, the blue boxes mark the location of F N ; (b) Segmentation model output value, the red box marks the location of F P ; (c) Decisions of T P , F N , and F P . T P renders as yellow point clouds, F N renders as blue point clouds, and F P renders as red point clouds.
Figure 8. Determination method of T P , F N , and F P of single maize segmentation. (a) Standard value of individual plant segmentation, the blue boxes mark the location of F N ; (b) Segmentation model output value, the red box marks the location of F P ; (c) Decisions of T P , F N , and F P . T P renders as yellow point clouds, F N renders as blue point clouds, and F P renders as red point clouds.
Agriculture 12 01450 g008
Figure 9. Prediction of plant growth trends. (a) Standard value of growth trend for Maize-01; (b) Predicted value of growth trend for Maize-01; (c) Standard value of growth trend for Maize-02; (d) Predicted value of growth trend for Maize-02; (e) Standard value of growth trend for Maize-03; (f) Predicted value of growth trend for Maize-03; (g) Standard value of growth trend for Maize-04; (h) Predicted value of growth trend for Maize-04.
Figure 9. Prediction of plant growth trends. (a) Standard value of growth trend for Maize-01; (b) Predicted value of growth trend for Maize-01; (c) Standard value of growth trend for Maize-02; (d) Predicted value of growth trend for Maize-02; (e) Standard value of growth trend for Maize-03; (f) Predicted value of growth trend for Maize-03; (g) Standard value of growth trend for Maize-04; (h) Predicted value of growth trend for Maize-04.
Agriculture 12 01450 g009
Figure 10. Scatter plot of growth direction prediction error.
Figure 10. Scatter plot of growth direction prediction error.
Agriculture 12 01450 g010
Figure 11. Individual maize plant segmentation. T P renders as yellow point clouds, F N renders as blue point clouds, and F P renders as red point clouds. (a) Maize-01; (b) Maize-02; (c) Maize-03; (d) Maize-04.
Figure 11. Individual maize plant segmentation. T P renders as yellow point clouds, F N renders as blue point clouds, and F P renders as red point clouds. (a) Maize-01; (b) Maize-02; (c) Maize-03; (d) Maize-04.
Agriculture 12 01450 g011
Figure 12. Individual maize segmentation accuracy; (a) Precision rate; (b) Recall rate; (c) F1-score .
Figure 12. Individual maize segmentation accuracy; (a) Precision rate; (b) Recall rate; (c) F1-score .
Agriculture 12 01450 g012
Figure 13. Stratification visualization of maize leaves. Sort by leaf p 1 point elevation value in descending order and render the colors one by one (Figure 7). (a) Standard value; (b) Predictive value.
Figure 13. Stratification visualization of maize leaves. Sort by leaf p 1 point elevation value in descending order and render the colors one by one (Figure 7). (a) Standard value; (b) Predictive value.
Agriculture 12 01450 g013
Figure 14. Clustering accuracy analysis of maize organs. (a) Column chart of I o U value of maize tassel; (b) Column chart of I o U value of maize ear; (c) Column chart of maize stalk height error value; (d) Scatter plot of maize leaf length.
Figure 14. Clustering accuracy analysis of maize organs. (a) Column chart of I o U value of maize tassel; (b) Column chart of I o U value of maize ear; (c) Column chart of maize stalk height error value; (d) Scatter plot of maize leaf length.
Agriculture 12 01450 g014
Figure 15. Maize-04-No. 2 visualization. The blue point cloud is the topmost leaf that was mistakenly identified as the tassel, and the green is the naturally broken tassel.
Figure 15. Maize-04-No. 2 visualization. The blue point cloud is the topmost leaf that was mistakenly identified as the tassel, and the green is the naturally broken tassel.
Agriculture 12 01450 g015
Table 1. Field maize point cloud data attributes.
Table 1. Field maize point cloud data attributes.
NumberType of MaizeAcquisition Date (Day/Month/Year)Maize NumbersGrowing StageArea SizeNumber of Scan StationsNumber of Scan PointsAverage Position Deviation
Maize-01Feed grade yellow maize13 July 202189Seedling stage5.9 × 3.4 m812,585,0280.61 mm
Maize-02Sweet maize17 July 202189heading stage6.1 × 3.9 m811,686,5600.64 mm
Maize-03Sweet maize29 July 202190heading stage5.9 × 3.6 m810,688,5900.73 mm
Maize-04Waxy maize30 July 202129Full ripening stage4.0 × 2.7 m812,920,6430.69 mm
Table 2. Normality tests: skewness and kurtosis.
Table 2. Normality tests: skewness and kurtosis.
NSkewnessKurtosis
StatisticStd. ErrorStatisticStd. Error
Valid N3000.1130.1410.1610.281
Table 3. Individual maize segmentation result statistics table.
Table 3. Individual maize segmentation result statistics table.
TP
(Yellow)
FN
(Blue)
FP
(Red)
PRF-ScoreOA
[0, 80%][80%, 95%][95%, 100%][0, 80%][80%, 95%][95%, 100%][0, 80%][80%, 95%][95%, 100%]
Maize-015,727,64273,57962,7810137632264028610.98
Maize-028,826,515157,479108,0643167012365027620.97
Maize-039,815,061268,093285,1774206631572123660.95
Maize-045,913,240228,412123,82212260111806230.94
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, C.; Hu, F.; Peng, J.; Wang, J.; Zhai, R. Segmentation and Stratification Methods of Field Maize Terrestrial LiDAR Point Cloud. Agriculture 2022, 12, 1450. https://doi.org/10.3390/agriculture12091450

AMA Style

Lin C, Hu F, Peng J, Wang J, Zhai R. Segmentation and Stratification Methods of Field Maize Terrestrial LiDAR Point Cloud. Agriculture. 2022; 12(9):1450. https://doi.org/10.3390/agriculture12091450

Chicago/Turabian Style

Lin, Chengda, Fangzheng Hu, Junwen Peng, Jing Wang, and Ruifang Zhai. 2022. "Segmentation and Stratification Methods of Field Maize Terrestrial LiDAR Point Cloud" Agriculture 12, no. 9: 1450. https://doi.org/10.3390/agriculture12091450

APA Style

Lin, C., Hu, F., Peng, J., Wang, J., & Zhai, R. (2022). Segmentation and Stratification Methods of Field Maize Terrestrial LiDAR Point Cloud. Agriculture, 12(9), 1450. https://doi.org/10.3390/agriculture12091450

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop