Next Article in Journal
Remote Sensing Technologies Using UAVs for Pest and Disease Monitoring: A Review Centered on Date Palm Trees
Previous Article in Journal
Changes in Urban Green Spaces in the Pearl River Delta Urban Agglomeration: From the Perspectives of the Area, Spatial Configuration, and Quality
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Techniques for Canopy to Organ Level Plant Feature Extraction via Remote and Proximal Sensing: A Survey and Experiments

1
College of Engineering and Computer Science, Texas A&M University-Corpus Christi, Corpus Christi, TX 78412, USA
2
Department of Mechanical Engineering, Texas A&M University, College Station, TX 77840, USA
3
Digital Agriculture Research Lab, Texas A&M AgriLife Research, Corpus Christi, TX 77843, USA
4
Department of Engineering Technology and Industrial Distribution, Texas A&M University, College Station, TX 77840, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(23), 4370; https://doi.org/10.3390/rs16234370
Submission received: 11 October 2024 / Revised: 28 October 2024 / Accepted: 4 November 2024 / Published: 22 November 2024

Abstract

:
This paper presents an extensive review of techniques for plant feature extraction and segmentation, addressing the growing need for efficient plant phenotyping, which is increasingly recognized as a critical application for remote sensing in agriculture. As understanding and quantifying plant structures become essential for advancing precision agriculture and crop management, this survey explores a range of methodologies, both traditional and cutting-edge, for extracting features from plant images and point cloud data, as well as segmenting plant organs. The importance of accurate plant phenotyping in remote sensing is underscored, given its role in improving crop monitoring, yield prediction, and stress detection. The review highlights the challenges posed by complex plant morphologies and data noise, evaluating the performance of various techniques and emphasizing their strengths and limitations. The insights from this survey offer valuable guidance for researchers and practitioners in plant phenotyping, advancing the fields of plant science and agriculture. The experimental section focuses on three key tasks: 3D point cloud generation, 2D image-based feature extraction, and 3D shape classification, feature extraction, and segmentation. Comparative results are presented using collected plant data and several publicly available datasets, along with insightful observations and inspiring directions for future research.

1. Introduction

Remote sensing is a cornerstone for plant phenotyping in smart agriculture, enabling high-throughput, non-invasive, and precise monitoring of crop traits across large fields. Technologies such as multispectral and hyperspectral imaging, UAV platforms, and satellite observations allow for real-time data collection on plant health, growth patterns, and stress responses, which are crucial for optimizing agricultural practices and enhancing crop yields. Traditional phenotyping methods, which are often labor-intensive and prone to error, are significantly improved by remote sensing. This technology allows for the rapid assessment of critical traits like plant height, leaf area index, and chlorophyll content, facilitating targeted interventions that enhance resource efficiency and reduce environmental impacts.

1.1. Background and Motivation

Plant phenotyping is a crucial component in modern agriculture and crop management, enabling researchers and agronomists to gain a deeper understanding of plant traits, growth dynamics, and responses to environmental stressors. The primary objective of phenotyping is to quantitatively measure various plant traits, such as the canopy structure, leaf area, plant height, and biomass, which are vital for plant breeding, yield prediction, and resource management. Traditionally, plant phenotyping has relied on manual measurements, which are time-consuming, labor-intensive, and prone to human error. With the rise of digital agriculture, remote and proximal sensing technologies have become essential tools for the high-throughput, non-destructive monitoring of crops at various scales, from the canopy to organ level [1].
Remote sensing platforms, such as satellites, unmanned aerial vehicles (UAVs), and ground-based systems, offer new opportunities for large-scale phenotyping by capturing diverse datasets using multispectral, hyperspectral, and thermal imaging sensors. These technologies provide a wealth of information on plant health, nutrient status, water content, and structural characteristics, enabling more precise and timely interventions in crop management [2,3]. However, the complexity of plant architectures and the variability in environmental conditions present significant challenges for extracting and interpreting phenotypic traits from these datasets. Thus, developing robust feature extraction and segmentation methods that can handle multi-scale data, such as 2D imagery and 3D point clouds, is crucial for advancing plant phenotyping research [4].
Plant feature extraction and segmentation from 2D images is a rapidly growing field in plant phenotyping, enabling the analysis of plant traits and characteristics from digital images [5]. With the advancement of computer vision and machine learning techniques, 2D image-based methods offer a non-invasive, cost-effective, and high-throughput approach to extract valuable information from plant images, such as the leaf shape, size, color, and texture [6]. Accurate segmentation and feature extraction from 2D images facilitate the quantification of plant growth, development, and responses to environmental stimuli, ultimately supporting plant breeding, crop monitoring, and precision agriculture [7]. Various approaches have been proposed for plant feature extraction and segmentation, including texture-based leaf identification [8], shape and texture feature-based classifications [9], and color histogram-based classification [10]. Recently, multi-descriptor approaches have also been developed for leaf classification [11]. Among the many techniques used in plant phenotyping is the 3D point cloud technology in conjunction with deep neural networks.
Figure 1 illustrates the trend in publications on canopy, plant, and organ-level feature extraction methods using remote sensing techniques, and an analysis from 2000 to 2024 of publication trends in feature extraction methods for plant phenotyping demonstrates significant growth across organ-level, plant-level, and canopy-level studies. Plant-level research holds the highest total publications, indicating a predominant focus on this area.
Notably, all levels experienced a peak in publications around 2021, likely due to technological advancements. The surge in canopy-level studies post-2015 reflects the adoption of remote sensing technologies like drones and satellites for large-scale analyses. These trends underscore the importance of integrating feature extraction methods across different organizational levels and leveraging advanced technologies such as hyperspectral imaging, deep learning, and machine learning algorithms.
This paper aims to provide a holistic view of the latest advancements in remote sensing for plant phenotyping, covering the entire spectrum from canopy-level to organ-level sensing. The insights presented here will serve as a valuable resource for researchers and practitioners looking to develop new phenotyping methods or enhance existing ones for precision agriculture and sustainable crop management.
Through empirical evaluation and case studies, these techniques demonstrate their efficacy in enhancing agricultural productivity, sustainability, and resilience to environmental challenges. Table 1 shows the techniques used for plant phenotyping in remote sensing.

1.2. Scope and Objectives

The primary focus of this review is to provide a comprehensive overview of the latest techniques for plant phenotyping using remote and proximal sensing across different scales—canopy, plant, and organ levels. This survey covers satellite-based, UAV-based, and ground-based sensing methods, evaluating their effectiveness in extracting critical phenotypic traits for various applications. The scope includes an in-depth analysis of feature extraction methodologies, such as multispectral and hyperspectral imaging, LiDAR (Light Detection and Ranging), synthetic aperture radar (SAR), and 3D point cloud segmentation. Furthermore, this paper aims to bridge the gap between these multi-scale sensing techniques by providing a unified view of how different data acquisition methods and analytical frameworks can be integrated to achieve high-resolution, multi-modal plant phenotyping.
The objectives are threefold: (1) to highlight the advancements in remote sensing technologies for automated phenotyping, (2) to compare the strengths and limitations of each method for specific phenotypic traits and environmental contexts, and (3) to provide a critical assessment of future directions for research and development in this rapidly evolving field. The overarching goal is to support the development of scalable, efficient, and accurate phenotyping methods that can be deployed across different agricultural landscapes.

1.3. Contributions of This Paper

This paper makes several unique contributions to the field of plant phenotyping using remote and proximal sensing techniques. First, it systematically reviews the current state-of-the-art in feature extraction and segmentation methods, covering both 2D image-based and 3D point cloud-based approaches. Second, the survey provides a detailed comparison of various sensing platforms, from satellite-based large-scale monitoring to UAV-based field-level assessments and organ-level analyses using ground-based systems. This multi-scale perspective helps identify the most suitable techniques for specific phenotyping tasks and highlights the potential of integrating different modalities to capture complex plant traits more effectively.
Additionally, this review identifies key research gaps and challenges in current methodologies, such as the need for improved accuracy in organ-level segmentation and the development of robust algorithms that can handle varying spatial resolutions and data noise. By synthesizing findings from recent studies and experiments, this paper offers valuable insights into how emerging technologies like AI, machine learning, and advanced 3D modeling can be leveraged to enhance the precision and scalability of plant phenotyping systems. It not only reviews progress in remote/proximal sensing for plant phenotyping but also demonstrates various levels of plant feature extraction techniques, ranging from remote and proximal data collection to feature extraction and segmentation of plants. The insights of each experiment provided are expected to help researchers in the field and to inform future research and support the design of next-generation phenotyping platforms that combine the strengths of remote and proximal sensing technologies.

1.4. Structure of the Paper

This paper is structured as follows: Section 2, Applications of Plant Phenotyping in Remote Sensing, provides a detailed exploration of various remote sensing technologies and their applications at the canopy, plant, and organ levels. Section 3, Technical Progress of Machine Learning and 3D Point Cloud Classification in Remote Sensing, delves into advancements in 3D point cloud generation and classification, which are essential for capturing detailed plant structures, particularly when integrated with LiDAR and other imaging systems. Section 4, Plant Feature Extraction, discusses the techniques used to extract meaningful features from remote sensing data, emphasizing the use of machine learning and image processing for an accurate phenotypic trait analysis. Section 5, Remote Sensing and Plant Phenotyping: Insights, presents a synthesis of experimental results from recent studies that have validated these technologies in practical applications. Section 6 is the Discussion. Finally, Section 7, Conclusions, summarizes the key findings, highlights the current challenges, and outlines potential future directions for developing remote sensing technologies for plant phenotyping.
Remote sensing technologies for plant phenotyping can be categorized into three main groups: satellite-based, UAV-based, and ground-based (proximal) sensing (Figure 2). Each platform has unique advantages and challenges based on the spatial resolution, data acquisition cost, and the ability to capture specific phenotypic traits.

2. Overview of Plant Phenotyping Using Remote/Proximal Sensing Technologies

Remote sensing technologies have significantly transformed plant phenotyping by providing high-throughput, non-destructive methods for monitoring and analyzing plant traits across different scales, from the canopy down to the organ level [12,13]. These technologies leverage advancements in imaging and data analysis to offer precise and actionable insights into plant health, growth, and productivity. This section explores recent advancements in satellite and UAV-based remote sensing techniques for plant feature extraction, highlighting their applications, strengths, and limitations. In Table 2, the focus of study and techniques at different levels of plant level phenotyping have been reviewed.

2.1. Satellite-Based Plant Feature Extraction

Satellite remote sensing offers a macroscopic view of agricultural landscapes, enabling extensive monitoring of crops and vegetation across large areas. With recent technological advancements, satellites now provide improved spatial, spectral, and temporal resolutions, facilitating detailed extraction of plant phenotypic traits.

2.1.1. Multispectral and Hyperspectral Imaging with Respect to Satellites

Satellites equipped with multispectral and hyperspectral sensors capture data across a range of wavelengths, which is critical for a detailed analysis of plant health and phenotypic traits. Multispectral imaging, which typically captures a limited number of broad spectral bands, is useful for basic vegetation monitoring through indices like the Normalized Difference Vegetation Index (NDVI) and the Enhanced Vegetation Index (EVI). These indices help assess overall plant health, chlorophyll content, and stress levels by analyzing differences in red and near-infrared reflectance, as demonstrated by the Sentinel-2 satellite’s capabilities [14]. Hyperspectral imaging, in contrast, provides a much finer spectral resolution by capturing hundreds of narrow spectral bands, allowing for a detailed analysis of the biochemical properties of plants, such as the nitrogen content, pigment concentrations, and water status. This high spectral resolution makes hyperspectral imaging particularly powerful for the early detection of diseases, nutrient deficiencies, and water stress, providing a comprehensive view of plant physiology [15].

2.1.2. Thermal Imaging with Respect to Satellites

Thermal sensors on satellites measure the thermal radiation emitted by plants, offering insights into the plant water status and stress levels. This method is particularly effective in detecting water stress because stressed plants often close their stomata to conserve water, leading to an increase in leaf temperature. Thermal imaging thus becomes an invaluable tool for irrigation management and drought monitoring. By analyzing thermal data, researchers can create water stress maps to guide irrigation decisions, optimize water use, and prevent crop losses due to drought [16]. Additionally, thermal imaging can detect heat stress and its impacts on photosynthetic efficiency, enabling better management of crops under varying climatic conditions.

2.1.3. Synthetic Aperture Radar (SAR)

SAR technology provides all-weather, day-and-night imaging capabilities, which are critical for consistent monitoring of plant growth and health, regardless of the weather conditions. SAR’s ability to penetrate cloud cover makes it a reliable tool for estimating biomass, monitoring soil moisture levels, and assessing vegetation structure. SAR sensors are particularly useful in regions with frequent cloud cover or during rainy seasons when optical sensors may be less effective. SAR data have been applied successfully to monitor changes in plant structure, detect lodging in crops, and assess the impacts of flooding on agricultural fields [17]. The typical spatial resolution of SAR is in the range of 1–100 m, depending on the system configuration, while satellite optical imaging systems typically achieve spatial resolutions between 10 and 30 m. SAR complements optical and thermal data by providing structural information about plant canopies, offering a more holistic view of crop conditions.

2.1.4. LiDAR (Light Detection and Ranging)

Although less common on satellite platforms due to its weight and power requirements, LiDAR technology provides high-resolution 3D models of plant canopies. These models offer detailed insights into the plant height, volume, and structural characteristics, which are crucial for understanding plant growth dynamics, biomass accumulation, and canopy architecture [18]. In terrestrial or airborne applications, LiDAR is highly effective for creating precise digital elevation models (DEMs) and for applications such as forest inventory, where accurate measurements of tree height and volume are needed. When used in conjunction with other sensors, LiDAR data can enhance the understanding of the vertical canopy structure and contribute to more accurate estimates of crop yield and biomass. Unlike satellite-based or SAR imaging, LiDAR sensors offer the highest possible resolution for 3D structural mapping, with resolutions up to 0.1 m

2.2. UAV-Based Plant Feature Extraction

Unmanned aerial vehicles (UAVs) provide unparalleled flexibility and high-resolution data collection capabilities at lower altitudes, making them particularly well-suited for detailed phenotyping at the canopy, plant, and organ levels. UAVs allow for more frequent and targeted data collection, which are essential for monitoring dynamic changes in plant phenotypes over time.

2.2.1. Multispectral and Hyperspectral Imaging with Respect to UAVs

UAVs equipped with multispectral and hyperspectral cameras offer high-resolution images that can be tailored to specific phenotyping tasks, such as detecting diseases, assessing the nutrient status, and monitoring plant health. Multispectral UAVs can capture several specific bands that are crucial for evaluating vegetation indices, such as NDVI and the Normalized Difference Red Edge (NDRE), which are indicators of plant vigor and stress [19]. Hyperspectral UAVs, which capture continuous spectral data across many bands, provide even more detailed information on plant biochemistry and physiology, allowing for the detection of subtle stress responses, early disease symptoms, and nutrient deficiencies. UAV-based hyperspectral imaging has been particularly valuable in precision agriculture, where detailed spectral information is used to apply inputs (like fertilizers or pesticides) more precisely and efficiently.

2.2.2. Thermal Imaging with Respect to UAVs

Thermal cameras mounted on UAVs are used to monitor plant temperature, a direct indicator of water stress and transpiration rates. UAV-based thermal imaging provides high spatial resolution and temporal flexibility, enabling frequent monitoring of the water status across different parts of the field [20]. By creating thermal maps, farmers can identify areas of the field that are under water stress and adjust irrigation practices accordingly. Additionally, thermal imaging can help detect plant diseases that alter transpiration rates, thereby providing an early warning system for disease management.

2.2.3. RGB Imaging and Structure from Motion (SfM)

High-resolution RGB cameras on UAVs, combined with Structure from Motion (SfM) algorithms, are used to generate detailed 3D models of plant canopies. These models are essential for measuring the plant height, canopy cover, and biomass, which are important phenotypic traits. The SfM technique reconstructs 3D structures from a series of overlapping 2D images taken from different angles, providing a cost-effective alternative to LiDAR for 3D modeling. RGB imaging is also used to detect visual symptoms of stress or disease and to monitor growth stages by analyzing changes in color and texture over time [21].

2.2.4. LiDAR and Depth Sensing

UAV-mounted LiDAR systems provide high-resolution 3D data that are invaluable for precise measurements of plant structure and morphology. Unlike passive optical sensors, LiDAR can penetrate the canopy and capture the 3D structures of plants, allowing for accurate assessments of the plant height, canopy density, and leaf area. This technology is particularly useful for phenotyping traits related to the plant architecture, such as the leaf angle distribution and branching patterns, which are important for understanding light interception and photosynthesis in crops [21]. The highest possible resolution of UAV-mounted LiDAR systems is approximately 0.1–0.3 m.
Fluorescence Imaging:
UAVs equipped with fluorescence sensors can detect specific wavelengths emitted by plants that are indicative of photosynthetic activity and plant health. Chlorophyll fluorescence, in particular, is a direct indicator of photosynthetic efficiency and can reveal early stress responses that are not visible through other imaging methods [22,23]. This technique is used to monitor plant stress, detect diseases, and assess the effectiveness of treatments. Fluorescence imaging can provide real-time feedback on plant health, enabling rapid decision-making in precision agriculture. Fluorescence imaging typically provides a resolution of up to 0.5 m. The use of remote sensing technologies for plant phenotyping provides valuable insights into crop health, growth, and productivity, empowering farmers and researchers to make precise, data-driven decisions. Advancements in satellite and UAV-based remote sensing technologies have enabled the monitoring and analysis of plant traits at various scales, from the canopy to organ level. By leveraging these tools, modern agriculture can improve efficiency, reduce input costs, and enhance sustainability.

2.3. Phenotyping Scope in Remote and Proximal Sensing

Canopy-Level Phenotyping: Crop yield Estimations, Disease Detection, and Canopy Cover.
At the canopy level, remote sensing through UAVs and satellites primarily focuses on whole-plant or vegetation-level analyses. Multispectral and hyperspectral imaging provide essential data for calculating vegetation indices like NDVI and EVI, which are used for monitoring overall plant health, estimating biomass, and assessing the canopy structure. Thermal imaging helps detect water stress at the canopy level, while LiDAR and RGB images enable 3D reconstructions for measuring plant height and canopy density. More features for canopy-level phenotyping often used in remote sensing are discussed in Table 3.
The integration of satellite and UAV-based remote sensing technologies, along with controlled environment setups (CESs), has revolutionized plant phenotyping at all levels—canopy, plant, and organ. These advancements provide detailed, high-throughput methods for assessing plant traits, contributing to more precise and efficient agricultural practices. With ongoing improvements in sensor technologies and data analysis techniques, the application of remote sensing in phenotyping will continue to enhance crop management and sustainability.
The diverse methodologies for canopy-level phenotyping are essential for advancing agricultural research and enabling precise crop management strategies. Spectral features such as the Normalized Difference Vegetation Index (NDVI), Green NDVI (GNDVI), Enhanced Vegetation Index (EVI), and Soil-Adjusted Vegetation Index (SAVI) are widely utilized for monitoring vegetation health, estimating biomass, and assessing crop stress. These indices leverage various spectral bands (e.g., NIR, red, and green) to capture subtle changes in photosynthetic activity and the chlorophyll content. The NDVI, for example, compares reflectance in the NIR and red bands to assess plant vigor, making it a powerful tool for evaluating crop growth and development [47,54]. Such indices are crucial for yield prediction and growth monitoring under varying environmental conditions.
Additionally, integrating these spectral indices with evapotranspiration (ET) models enhances the understanding of crop water use and helps optimize irrigation strategies. ET, which represents the combined effect of soil evaporation and plant transpiration, is a key indicator of crop water requirements and overall health. UAV-based imaging combined with spectral indices and thermal imagery provides precise ET monitoring for effective water management in precision agriculture. For instance, the two-source energy balance model (TSEB) can partition soil and canopy ET, improving water use efficiency across different crop stages, which is vital for preventing water stress and enhancing crop yields. By integrating a time series analysis of VIs with ET models, a comprehensive assessment of water use and crop development can be achieved, leading to more efficient agricultural practices [47].
Textural features, such as the Gray-Level Co-occurrence Matrix (GLCM), entropy, and local binary patterns (LBPs), provide insights into crop structure and stress conditions by analyzing texture patterns in plant canopies. These methods are particularly useful for distinguishing healthy from stressed crops based on canopy uniformity and texture variations, making them effective for detecting crop type variations and monitoring stress levels. Structural features such as the canopy height, volume, and leaf area index (LAI) are vital for estimating the biomass, potential yield, and canopy architecture, contributing to accurate phenotyping of plant growth and development. Temporal features, including the growth rate and phenological metrics, are essential for tracking key developmental stages such as flowering and senescence, which are crucial for predicting crop yield and identifying stress responses over time. Environmental features like the soil moisture content and evapotranspiration (ET), derived from the NDWI and thermal images, play significant roles in effective irrigation management and water balance modeling, providing detailed insights into the soil water content and crop water status.
Deep Neural Network techniques, including convolutional neural networks (CNNs), long short-term memory (LSTM) networks, and attention mechanisms, have significantly enhanced feature extraction, phenotyping accuracy, and the simulation of crop growth patterns. These approaches can analyze high-dimensional data, generate synthetic images, and predict physiological changes in crops under various environmental conditions. For instance, integrating deep learning models with a time series analysis of vegetation indices allows growers to assess crop conditions over an entire growing season, leading to improved yield prediction models and the early detection of crop stress.
Finally, 3D reconstruction methods using point cloud data enable detailed analyses of the canopy structure and distribution, providing precise structural information for phenotyping and yield estimations. Point cloud data capture spatial details of plant canopies, allowing for accurate modeling of the canopy architecture, biomass estimation, and the detection of subtle changes in the canopy structure over time. Overall, the integration of these diverse feature extraction techniques provides a comprehensive toolkit for effective canopy-level phenotyping, supporting improved crop management, yield prediction, and agricultural decision-making.

2.3.1. Plant-Level Phenotyping: Plant Height, Biomass Estimation, Plant Health Monitoring

At the individual plant level, UAVs equipped with high-resolution cameras and sensors can monitor specific plant traits such as growth patterns, shape, and nutrient deficiencies. The 3D models constructed from RGB and LiDAR data provide detailed insights into plant morphology, while multispectral and hyperspectral imaging reveal biochemical traits such as photosynthetic activity and stress responses. UAV-mounted systems offer flexibility, enabling continuous data collection throughout different growth stages significantly 3D point clouds.
Skeletonization [55] of plants structure from huge 3D point clouds are more advantageous for plant-level phenotyping. The skeletonization process for plant phenotyping involves denoising the point cloud, applying iterative Laplacian contraction to reduce it to a skeletal structure, adaptively sampling key points, and establishing connections to form a cohesive skeleton while avoiding loops. This method enables precise measurements of morphological traits like plant height and leaf angles in under 100 s, making it efficient for automated maize phenotyping. Skeletonization offers significant advantages: it simplifies data processing, supports semantic segmentation of plant organs, improves trait measurement accuracy, enhances visualization, and is compatible with functional-structural plant models. Its robustness to noise and occlusion makes it a powerful tool for scalable, high-throughput phenotyping across diverse environments, although additional refinement is needed for more complex plant architectures and field conditions.

2.3.2. Organ-Level Phenotyping: Leaf Area Index, Stem Thickness, and Flower and Fruit Counting

For phenotyping plant organs, such as leaves, stems, and fruits, high-resolution imaging techniques are critical. In indoor controlled environments (CESs) or UAV settings, fluorescence imaging plays a significant role in monitoring the chlorophyll content and detecting early stress responses at the organ level. Three-dimensional imaging and detailed organ segmentation allow for the measurement of traits like the leaf area, fruit detection, and stem architecture. Advanced deep learning techniques are often employed to automate the segmentation and analysis of individual organs, providing researchers with highly specific insights into plant physiology. More details regarding organ-level phenotyping methods and techniques are listed in Table 4.
Plant and organ-level feature extraction techniques play crucial roles in precision agriculture, plant phenotyping, and disease detection. Each technique provides unique insights into various aspects of plant morphology, health, and physiological processes. For instance, hyperspectral imaging and NIR imaging capture detailed spectral data and internal structures, making them ideal for identifying subtle changes in plant health and detecting early stress. Terrestrial laser scanning and point cloud data provide accurate 3D structural models, allowing for the quantification of plant organs and growth analysis. Microscopic approaches enable cellular-level studies, focusing on stomata behavior and other tissue characteristics that are essential for understanding physiological responses to environmental factors. In feature extraction, techniques such as morphological feature analysis and textural feature analysis are widely applied to assess disease impacts, leaf shape, and species classification. GLCMs and Wavelets excel in capturing textural and high-resolution details, making them valuable for segmenting complex patterns in leaves and stems. Moreover, advanced methods like graph-based techniques (PHOGs) and multi-fractal analysis effectively represent the complex geometry and multifractal nature of plant organs, enabling accurate identification and segmentation under diverse conditions.
Additionally, deep learning and optimization methods, such as CNNs, LSTM, and attention mechanisms, are integrated to automate feature extraction and improve the detection accuracy. These models can process multi-temporal data and selectively emphasize key features, enhancing the model’s robustness in real-world applications. Genetic algorithms and other optimization techniques (e.g., particle swarm optimization and Rider Cuckoo Search) further refine feature selection and segmentation, resulting in the more precise classification of diseases or growth stages. By utilizing these diverse methodologies, researchers can extract critical plant features, enabling early disease detection, species classification, and optimized phenotyping, and ultimately contributing to sustainable agricultural practices.

3. Technical Progress in 3D Point Cloud Classification in Remote Sensing

In this section, we review the technical advancements in 3D point cloud classification techniques, one of the most critical and popular techniques for plant phenotyping for both remote and proximal sensing, tracing the evolution from conventional methods to machine learning and deep learning approaches. Recognizing PointNet as a significant milestone in the deep learning-based 3D point cloud classification, we present comparative test results for an urban dataset with complex structures, including vegetation, to highlight the differences between conventional methods and PointNet.

3.1. Conventional Methods

In the early days of 3D point cloud processing, researchers developed traditional methods to analyze and understand these complex datasets. In addition to these histogram-based methods, researchers also explored curvature-based and normal-based approaches. The authors of [121] introduced curvature-based methods, which estimate curvature values at each point to describe the local shape. Similarly, normal-based methods estimate surface normals at each point to describe the local orientation. These approaches enable feature description, segmentation, and registration, and have been used in various applications. Finally, voxel-based methods [122] were developed to discretize the 3D space into voxels, allowing for the efficient processing and analysis of point clouds. These methods enable fast and efficient processing but may lose some accuracy due to the discretization process.
In addition to these local shape descriptor or voxel methods, in 2004, Belongie et al. [123] developed the Shape Context method, which uses 2D histograms to encode the spatial distribution of points. This approach has been successful in feature description and object recognition tasks, particularly in the presence of clutter and occlusions. Shape Context creates a 2D histogram of the angles and distances between neighboring points, capturing the underlying shape and structure of the point cloud. One of the pioneering methods in this field was the Point Feature Histogram (PFH) technique, introduced by Rusu et al. in 2008 [124,125]. The PFH encodes the spatial distribution of points in a local neighborhood, allowing for feature description and registration. The method creates a histogram of the angles and distances between neighboring points, capturing the underlying structure of the point cloud. This approach enables robust feature description and registration, even in the presence of noise and missing data. Building on this work, Rusu et al. [126] later developed the Fast Point Feature Histogram (FPFH) method, which improves upon the PFH by reducing computational complexity. The FPFH uses a smaller set of neighboring points, making it faster and more efficient while still maintaining accurate feature description and registration capabilities. Around the same time, Johnson and Hebert introduced the Spin Images method, which uses 2D histograms to encode the spatial distribution of points. This approach enables feature description and object recognition and has been widely used in various applications. Spin Images creates a 2D histogram of the angles and distances between neighboring points, capturing the underlying shape and structure of the point cloud.
Despite these limitations, researchers continued to develop new methods to process and analyze 3D point cloud data. However, these traditional methods were often cumbersome, computationally expensive, and limited in their ability to handle the complexity and variability of point cloud data. These approaches enabled feature description, segmentation, and registration, but were often sensitive to noise and missing data. These traditional methods laid the foundation for 3D point cloud processing, enabling feature description, segmentation, registration, and object recognition (Table 5). While they have been largely surpassed by deep learning techniques, they remain important milestones in the development of this field. Below is a list of conventional point cloud processing methods deemed as milestones of technical development.

3.2. Deep Learning Approaches

Volumetric CNNs, multiview CNNs, spectral CNNs, and feature-based DNNs are all methods of 3D point cloud processing that have improved upon traditional analytic methods. Volumetric CNNs [128] apply 3D convolutional neural networks to voxel shapes, allowing for faster processing times and improved accuracy compared to traditional methods. However, this method is limited by resolution due to data sparsity and computation costs. Multiview CNNs [129] render 3D point clouds into 2D images and then apply 2D convolutional neural networks, making it easier to analyze complex shapes. However, this method is limited to shape classification and retrieval tasks and may lose important 3D information. Spectral CNNs [130] utilize meshes, specifically suited for manifold meshes like organic objects, offering greater efficiency compared to traditional methods. Feature-based DNNs [131] convert 3D data into vectors by extracting traditional shape features, which can be used in fully connected networks for classification. However, this method is limited by the representation power of the extracted features and may not capture complex shapes. Compared to traditional analytic methods, these methods offer improved efficiency and accuracy, but may have limitations in terms of resolution, shape complexity, and feature representation (Table 6).
Before the introduction of PointNet [132], the disorder and irregularity of point clouds made it impossible for deep learning technology to directly process point clouds. Early point cloud processing used hand-designed rules for feature extraction, which were time-consuming and limited in their ability to capture complex shapes. The introduction of PointNet and PointNet++ [133] allowed for the direct processing of point clouds, which improved the efficiency and accuracy of 3D point cloud processing. These methods offer significant advantages over traditional analytic methods, which relied on hand-designed rules and were limited in their ability to capture complex shapes. However, PointNet and PointNet++ require large amounts of training data and computational resources, which can be limitations. Overall, the choice of method depends on the specific application and the trade-off between efficiency, accuracy, and complexity. While traditional analytic methods are still useful for simple shapes and small datasets, deep learning methods offer improved efficiency and accuracy for complex shapes and large datasets.

3.3. Deep Learning Advances in 3D Point Cloud Processing

The development of segmentation models specifically tailored for 3D point cloud data has revolutionized the field of plant phenotyping, providing researchers with unprecedented capabilities for high-throughput and automated analyses of complex plant traits. Over the past decade, the focus has shifted from traditional machine learning algorithms to deep learning-based architectures, each designed to tackle unique challenges associated with plant structures, such as overlapping organs, variable point densities, and complex morphologies. Early models like PointNet (Qi et al., 2016) [132] and PointNet++ (Qi et al., 2017) [133] set the foundation by introducing point-wise feature extraction and hierarchical grouping strategies, which proved effective in controlled and less cluttered environments. However, these architectures struggled to capture intricate local and global patterns, making them less suitable for dense canopies and highly variable field conditions. To overcome these limitations, subsequent models such as DGCNN (Dynamic Graph CNN) (Yue et al., 2019) [134], PAConv (Xu et al., 2021) [135], and transformer-based architectures emerged, significantly enhancing the robustness and accuracy of segmentation in more challenging scenarios.
PointNet and PointNet++ employ multi-layer perceptron (MLP) networks to individually process each point, with PointNet++ extending this by incorporating local neighborhood information through hierarchical feature extraction. This design, while effective in relatively simple environments, fails to account for complex inter-point dependencies and cannot maintain context in large-scale plant datasets. On the other hand, DGCNN addresses these issues by employing a graph-based structure that dynamically updates neighboring points using EdgeConv, thereby enabling better modeling of local relationships. This makes DGCNN highly effective at capturing complex patterns, such as those found in dense maize or cotton fields. However, DGCNN’s reliance on local neighborhood aggregation makes it computationally intensive and susceptible to noise in low-density regions. Conversely, CurveNet (Xiang et al., 2021) [136] leverages curve-based features to capture connectivity and context in curved plant structures, improving its performance on datasets with prominent geometric features. Yet, its vulnerability to noisy data significantly limits its applicability in real-world agricultural settings.
The introduction of transformer-based models, namely, Point Transformer (PT) (Zhao et al., 2021) [137] and Stratified Transformer (ST) (Lai et al., 2022) [138] marked a substantial leap in segmentation performance by incorporating self-attention mechanisms for long-range feature aggregation. These architectures excel at handling complex inter-point interactions over large distances, making them ideal for segmenting crops with overlapping and intertwined organs, such as tomato or rose plants. By combining self-attention with multi-scale feature integration, the Stratified Transformer adapts to varying point densities and structural complexities, setting new benchmarks for accuracy in plant organ segmentation. However, these models come with the drawback of significantly higher computational costs, which can hinder their deployment in real-time applications and resource-constrained environments.
Recently, PAConv and Mask3D (Hou, et al., 2023) [139] have emerged as strong contenders for specific segmentation tasks. PAConv introduces a dynamic kernel construction method, where convolutional weights are adaptively generated based on the geometry of local point sets, allowing the model to generalize across different plant scales and structures. This adaptability makes PAConv particularly effective for plants with variable organ sizes, such as in cotton or potato datasets. However, the model’s complexity can lead to an increased training times and a need for more computational resources. Meanwhile, Mask3D extends the capabilities of Mask R-CNN to 3D data using sparse convolutions, making it one of the first methods to achieve robust instance segmentation on large-scale field datasets. Its performance on complex scenes, such as those captured using UAVs, highlights its potential for large-scale agricultural monitoring. Yet, Mask3D’s reliance on detailed pre-computed features limits its flexibility and increases preprocessing overhead, reducing its efficiency in dynamic environments.
The novel SP-LSCnet (Mertoğlu et al., 2024) [140] segmentation method, which integrates unsupervised clustering techniques with an adaptive classification network, offers a promising alternative by partitioning the input point cloud into superpoints, thereby reducing computational load and enabling more efficient segmentation of large-scale datasets. By employing attention-based modules to refine receptive fields, SP-LSCnet is able to dynamically adapt to varying organ geometries, making it particularly suited for scenarios with diverse plant species and complex structural variations. However, the reliance on a two-stage pipeline introduces challenges in maintaining superpoint homogeneity and mitigating misclassifications in regions with ambiguous geometry, such as thin petioles or small leaflets, which can degrade its performance in highly cluttered environments.
The development of segmentation models specifically tailored for 3D point cloud data has revolutionized the field of plant phenotyping, providing researchers with unprecedented capabilities for high-throughput and automated analyses of complex plant traits. Over the past decade, the focus has shifted from traditional machine learning algorithms to deep learning-based architectures, each designed to tackle unique challenges associated with plant structures, such as overlapping organs, variable point densities, and complex morphologies. Early models like PointNet [132] and PointNet++ [133] set the foundation by introducing point-wise feature extraction and hierarchical grouping strategies, which proved effective in controlled and less cluttered environments. However, these architectures struggled to capture intricate local and global patterns, making them less suitable for dense canopies and highly variable field conditions. To overcome these limitations, subsequent models such as DGCNN (Dynamic Graph CNN) [134], PAConv [135], and transformer-based architectures emerged, significantly enhancing the robustness and accuracy of segmentation in more challenging scenarios.
PointNet and PointNet++ employ multi-layer perceptron (MLP) networks to individually process each point, with PointNet++ extending this by incorporating local neighborhood information through hierarchical feature extraction. This design, while effective in relatively simple environments, fails to account for complex inter-point dependencies and cannot maintain context in large-scale plant datasets. On the other hand, DGCNN addresses these issues by employing a graph-based structure that dynamically updates neighboring points using EdgeConv, thereby enabling better modeling of local relationships. This makes DGCNN highly effective in capturing complex patterns, such as those found in dense maize or cotton fields. However, DGCNN’s reliance on local neighborhood aggregation makes it computationally intensive and susceptible to noise in low-density regions. Conversely, CurveNet [136] leverages curve-based features to capture connectivity and context in curved plant structures, improving its performance on datasets with prominent geometric features. Yet, its vulnerability to noisy data significantly limits its applicability in real-world agricultural settings.
The introduction of transformer-based models, namely, Point Transformer (PT) [137] and Stratified Transformer (ST) [138], marked a substantial leap in segmentation performance by incorporating self-attention mechanisms for long-range feature aggregation. These architectures excel at handling complex inter-point interactions over large distances, making them ideal for segmenting crops with overlapping and intertwined organs, such as tomato or rose plants. By combining self-attention with multi-scale feature integration, the Stratified Transformer adapts to varying point densities and structural complexities, setting new benchmarks for accuracy in plant organ segmentation. However, these models come with the drawback of significantly higher computational costs, which can hinder their deployment in real-time applications and resource-constrained environments.
Recently, PAConv [135] and Mask3D [139] have emerged as strong contenders for specific segmentation tasks. PAConv introduces a dynamic kernel construction method, where convolutional weights are adaptively generated based on the geometry of local point sets, allowing the model to generalize across different plant scales and structures. This adaptability makes PAConv particularly effective for plants with variable organ sizes, such as in cotton or potato datasets. However, the model’s complexity can lead to increased training times and a need for more computational resources. Meanwhile, Mask3D extends the capabilities of Mask R-CNN to 3D data using sparse convolutions, making it one of the first methods to achieve robust instance segmentation on large-scale field datasets. Its performance on complex scenes, such as those captured using UAVs, highlights its potential for large-scale agricultural monitoring. Yet, Mask3D’s reliance on detailed pre-computed features limits its flexibility and increases preprocessing overhead, reducing its efficiency in dynamic environments.
The novel SP-LSCnet [140] segmentation method, which integrates unsupervised clustering techniques with an adaptive classification network, offers a promising alternative by partitioning the input point cloud into superpoints, thereby reducing the computational load and enabling more efficient segmentation of large-scale datasets. By employing attention-based modules to refine receptive fields, SP-LSCnet is able to dynamically adapt to varying organ geometries, making it particularly suited for scenarios with diverse plant species and complex structural variations. However, the reliance on a two-stage pipeline introduces challenges in maintaining superpoint homogeneity and mitigating misclassifications in regions with ambiguous geometry, such as thin petioles or small leaflets, which can degrade its performance in highly cluttered environments.
The state-of-the-art segmentation models have evolved significantly, with each model offering unique strengths and limitations. While transformer-based architectures provide superior accuracy in complex environments, their high computational costs pose significant challenges. Conversely, models like PAConv and DGCNN offer a balanced approach for simpler plant structures but may falter in more intricate scenarios. The development of novel approaches such as SP-LSCnet demonstrates the potential of integrating unsupervised and attention-based methods to achieve robust segmentation across diverse agricultural contexts, highlighting the need for future research to focus on hybrid models that balance accuracy, computational efficiency, and adaptability to various plant phenotyping tasks. Contributions and limitations of advanced deep learning approaches are highlighted in Table 7.

3.4. Progress in Performance (Complexity/Data Size)

The performance of point cloud processing has dramatically improved with the advent of neural network-based methods, marking a significant leap over conventional techniques. Earlier methods, prior to 2010, were marked by several constraints that limited their efficiency and adaptability. These conventional approaches were often slow, relying on hand-crafted features and rule-based algorithms. These properties made it difficult to handle large-scale point clouds or complex scenes efficiently. Additionally, their limited ability to adapt to new data or scenarios made them less flexible, as they often required manual tuning to work effectively. Neural network-based methods, introduced after 2010, represent a major advancement in the field. As mentioned earlier, PointNet, developed in 2017, was a game-changer because it allowed direct processing of point clouds without needing voxelization or rasterization, which conventional methods often required. This new approach achieved state-of-the-art results in tasks like classification, segmentation, and semantic parsing, demonstrating significant improvements in processing speed and efficiency. PointNet could handle large-scale point clouds more efficiently, opening new possibilities for applications like autonomous vehicles and robotics. Following the success of PointNet, PointNet++ was introduced in 2018. It improved upon the original design by incorporating hierarchical feature learning and multi-scale aggregation, leading to better performance in challenging datasets and complex scenes. This approach also proved more efficient in handling large-scale point clouds, making it suitable for a wider range of applications.
Recent neural network-based methods, developed from 2020 onwards, have introduced even more sophisticated techniques. These newer approaches leverage attention mechanisms, graph neural networks, and transformer architectures, further enhancing performance and efficiency. They allow for real-time processing, which is crucial for applications in autonomous vehicles, robotics, and mobile devices. Additionally, these advanced methods are adaptable and can learn from data, allowing them to adjust to new scenarios without the rigid constraints of conventional methods. Neural network-based methods offer several advantages over traditional approaches. They are significantly faster, achieving real-time performance in some cases, and demonstrate much higher accuracy, consistently delivering state-of-the-art results across various point cloud processing tasks. Their ability to scale and adapt to large datasets and complex scenes makes them highly versatile, while their adaptability to new data and scenarios reflects a key shift in the field’s approach to processing point clouds. These advancements represent a new era of efficiency, accuracy, and scalability in point cloud processing.
Table 8 shows the significant improvements in speed and data size that have been achieved with the introduction of neural network-based methods. The processing time has been reduced by a factor of 10–50, the frame rate has increased by a factor of 5–20, the data size has been reduced by a factor of 5–10, and the storage requirements have decreased by a factor of 10–50.
Note the following:
  • Processing Time—the time it takes to process a single point cloud frame.
  • Frame Rate—the number of point cloud frames that can be processed per second.
  • Data Size—the size of a single point cloud frame.
  • Storage Requirements—the amount of storage needed to store one minute of point cloud data.

3.5. Exploring Point Cloud Segmentation via PointNet and SVM

The main difference between PointNet and other neural networks for 3D point cloud processing is its ability to directly handle point clouds without the need for transformation into regular 3D voxel grids or collections of images [133]. Additionally, PointNet is designed to respect the permutation invariance of points in the input and provides a unified architecture for various applications such as object classification, part segmentation, and scene semantic parsing. Permutation invariance of points in the input refers to the fact that the order of the points in a point cloud does not affect the outcome of the neural network. In other words, the network is invariant to the permutation of the input points. Point clouds are typically represented as a set of 3D points, and the order of these points is arbitrary. The same point cloud can be represented in different ways by permuting the order of the points. A neural network that is permutation invariant can process point clouds in any order and produce the same output. PointNet achieves permutation invariance using symmetric functions such as max pooling that aggregate information from all points in the point cloud. This allows the network to capture global information about the point cloud, regardless of the order of the points. Permutation invariance is an important property for point cloud processing, as it allows the network to focus on the geometric structure of the point cloud, rather than the specific order of the points. This makes the network more robust and able to generalize better to new point clouds.
In this section, we compare a conventional method, Support Vector Machine (SVM), and PointNet for 3D point cloud processing capability. This literature review explores the application of 3D deep learning techniques and SVM methods for segmenting point cloud data in urban environments, with a focus on a case study in Voorst, the Netherlands, utilizing the Actueel Hoogtebestand Nederland (AHN) dataset [148,149]. The AHN dataset, with over 32 million point records, provides a comprehensive representation of the terrain, simplified into 16 different class labels, from which five labels are used for analysis. The study adopts a systematic approach outlined in fundamental steps, including data curation, labeling, feature selection, sub-tiling, and normalization.

3.5.1. Data Processing

Normalization and outlier removal are essential preprocessing steps in handling point cloud data to ensure consistency and eliminate noise. Normalization involves scaling the point cloud data to a standardized range, typically between zero and one, to facilitate uniform processing across different datasets. Outlier removal aims to identify and discard data points that deviate significantly from the rest of the dataset, often indicating errors or anomalies in data acquisition. Downsampling is a technique used to reduce the size of point cloud data while preserving its essential features. This process is crucial for managing large datasets and improving computational efficiency in subsequent analysis tasks. One common method of down sampling is voxel downsampling [150], where points are grouped into voxels, and each occupied voxel generates one point by averaging all points within it. This method effectively reduces the number of points while maintaining the overall structure and characteristics of the dataset.
Weighted random sampling is a downsampling technique that prioritizes certain data points based on their importance or relevance to the task at hand. In the context of point cloud data, weighted random sampling involves computing the weight distribution of unique labels present in the dataset. These weights are then used to guide the random downsampling process, ensuring that points from different classes are represented proportionally in the downsampled dataset. This approach helps maintain the class balance and prevent bias in subsequent analysis or modeling tasks. Then, noise and data augmentation techniques are employed to enhance the robustness and generalization capability of models trained on point cloud data. These techniques involve introducing variations such as translation, scaling, and random rotations to the data to simulate real-world conditions and increase the diversity of the training set. By exposing the model to augmented data, it becomes more resilient to noise and variations encountered during inference.

3.5.2. Methodology

The approach taken in this project for the segmentation of outdoor 3D point clouds involves a multi-step process. First, the study begins with meticulous preprocessing of high-quality airborne LIDAR data obtained from Actueel Hoogtebestand Nederland [149]. This preprocessing step is essential for cleaning and organizing the raw point cloud data before segmentation. It involves tasks such as noise removal, outlier detection, and data normalization to ensure the data’s quality and consistency.
First, the study employs a conventional machine learning approach for semantic segmentation, which involves the integration of Fast Point Feature Histograms (FPFHs) (Rusu 2009) and Support Vector Machines (SVMs) (Figure 3). FPFHs are utilized for their precision in extracting local geometric features from the point cloud data. SVMs are chosen for their classification prowess and robustness against overfitting, complementing the FPFH-SVM framework. While FPFHs process data to generate histograms of point cloud data based on neighboring data with various features, SVMs cluster them into different classes, thus segmenting 3D point cloud data.
In contrast to the machine learning approach, this study also explores the advanced PointNet deep learning framework for semantic segmentation. PointNet offers an end-to-end solution for processing raw point cloud data directly, eliminating the need for handcrafted features. Its neural architecture autonomously identifies and learns representative features from the multi-dimensional nature of LIDAR data, making it exceptionally suitable for semantic segmentation tasks. Both the machine learning and PointNet approaches are evaluated based on their accuracy, computational efficiency, and scalability. The Focal Loss function [151], also known as the Modified Categorical Cross Entropy Loss, is a variation of the standard cross-entropy loss commonly used in classification tasks and is employed in PointNet architecture. It is particularly effective in addressing class imbalance problems, where certain classes in the dataset are heavily outnumbered by others. The focal loss function below is derived from the standard categorical cross-entropy loss but introduces a modulating factor to down-weight the loss assigned to well-classified examples
C C E ( S n , Y n ) = α y n log s n C = 1 m S n 1 y n
where Sn represents the predicted class score, yn represents the true class label, α y n represents the class weight for the true class, and 1(yn) is the indicator function. In the context of segmentation tasks, such as segmenting point clouds using PointNet, the Focal Loss function plays a crucial role in addressing the imbalance in class distributions. Point cloud data often contain classes with varying frequencies, such as rare objects or background clutter. By assigning higher weights to underrepresented classes, the Focal Loss function helps the model focus more on learning from these minority classes, improving the overall segmentation performance. By incorporating the Focal Loss function into the training process of models like PointNet, researchers have observed significant improvements in segmentation accuracy, especially for datasets with imbalanced class distributions. The ability to effectively handle a class imbalance ensures that the model learns to differentiate between classes more effectively, leading to more accurate and reliable segmentation results. Focal Loss incorporation leads to improved segmentation accuracy and robustness, making it a valuable enhancement in the field of point cloud analysis and semantic segmentation.

3.5.3. Results

In our experiment, PointNet shows significant discrepancies in its ability to classify different classes (Figure 4). Notably, it performs well in identifying ‘Ground’ and ‘Vegetation’, ‘Buildings’ classes, with a notable confusion between ‘Vegetation’ and other classes (Table 9). In addition, it struggles with the ‘Unclassified’ and ‘Water’ classes, showing almost no ability to identify them correctly (Figure 5). Whereas SVM with FPFH displays a strong performance on ‘Ground’ and ‘Vegetation’ classes similar to PointNet, it struggles with ‘Unclassified’, ‘Buildings’, and ‘Water’. Moreover, the resolution of the results also deteriorated. The confusion is notably less for ‘Ground’ and ‘Vegetation’, suggesting better discrimination between these and other classes (Figure 5). For a comparative analysis with traditional machine learning models, we have chosen to evaluate it based on the standard metrics shown in Figure 5, such as precision, recall, F1 score, and overall accuracy (see Table 9 for a comparison). This comparative analysis of SVM with FPFH and PointNet for the semantic segmentation of LIDAR point clouds reveals important insights for practical geospatial applications. While SVM with FPFH achieves higher accuracy in recognizing ‘Ground’ and ‘Vegetation’, it struggles with other classes such as ‘Buildings’ and ‘Water’, indicating challenges with a class imbalance. The diagram in Figure 6 shows the training and testing loss and accuracy for the PointNet architecture, which achieves approximately 81% accuracy.
Conversely, PointNet demonstrates broader class recognition but lower overall accuracy, suggesting its potential for applications requiring diverse class identification. However, both models face challenges in generalizing across diverse classes, highlighting the need for further research to enhance their performance. The impact of the class imbalance is evident in the significantly lower performance on minority classes like ‘Buildings’ and ‘Water’, suggesting the necessity of techniques such as SMOTE [152] or adjusting class weights for improved performance. While FPFH features excel in distinguishing ‘Ground’ and ‘Vegetation’, they may be less effective for complex structures like buildings or sparse classes like water. Exploring additional features or alternative feature extraction techniques could enhance the classification capabilities. For deployment, SVM with FPFH appears robust for applications requiring high accuracy in specific classes, while the computational efficiency of PointNet could be advantageous for real-time processing, despite its lower accuracy. One key observation from above experiments is that, the red, green, blue features along with the x, y, x features of the point cloud helped to segment the trees from the cone shaped buildings. Future research directions include adding more features and exploring hybrid approaches combining the strengths of both models and addressing the class imbalance through synthetic data generation or modified training objectives.

4. Plant Feature Extraction via 2D and 3D Data Modalities

In this section, we review 2D image and 3D point cloud-based plant feature extraction techniques, providing an in-depth analysis of recent technical advancements. Unlike industrial objects with standardized shapes, the complexity of plant structures poses significant challenges for extracting the comprehensive features necessary for operations such as harvesting. Therefore, more advanced data processing techniques are required to address these challenges and improve feature extraction for practical applications.

4.1. Two-Dimensional (2D) Image-Based Plants Organ-Level Feature Extraction

Plant image processing aims to enhance image quality, extract meaningful information, or transform images into different formats by performing specific operations on the original plant images. Typically, 2D RGB images are often used as the input, and various processing techniques, such as color conversion to convert them into RGB images or into multiple other formats such as HSL, HSV or image segmentation, and shape detection, are applied for accurate object recognition. This process is crucial in autonomous and precision agriculture, where images taken from agricultural fields must be transformed into useful information. This information can be then integrated with machine-learning-based methods for autonomous feature extraction and segmentation for application-specific requirements.
When harvesting plants, it is crucial to distinguish between ripe and unripe fruits or vegetables before picking them from the plant. Such decision-making tasks are usually based on human perception and judgment. Emulating this behavior is essential for autonomous agricultural robots to accomplish the allotted task successfully. Computer vision and image processing are vital in emulating the above-described behavior. Various image processing techniques have been developed to accomplish these tasks. These techniques include image segmentation, edge detection, blob analysis, color conversion, feature extraction, and various algorithms such as the Euclidean clustering algorithm, the Random Sample Consensus (RANSAC) algorithm, methods based on deep learning, and neural networks. These techniques and their agricultural applications are reviewed below.

4.1.1. Color Space Analysis and Conversion

Color space is essentially a mathematical model for describing the wide range of colors captured by the camera. Color space analysis is application-oriented and is classified into two categories, device-dependent (RGB, HSV, HSL, HIS, YUV, YIQ, CMYK, YCbCr, and YPbPr) and device-independent (XYZ and CIELAB), as described by Velastegui et al. [153].
For example, in the article by Al-Mashhadani et al. [154], OpenCV was adopted for color conversion (RGB to HSV) to detect fruits and distinguish between ripe and unripe ones. The steps involved in fruit detection included preprocessing techniques such as image resizing and bicubic interpolation, and image segmentation achieved using thresholding and histogram analysis. The process also involved color conversion and feature extraction based on color and shape, as well as training and classifying algorithms. This work utilized the HSV color space, with hue values used to establish the thresholding that distinguishes between ripe and transitioning (unripe) fruits. By combining these techniques, a model was developed to differentiate between ripe and unripe tomatoes, as described by Al-Mashhadani et al. [154]. In the study by Rao et al. [155], a color subtraction module (subtraction value of red and blue) was applied to distinguish cotton fruits from the rest of the plants, based on the observation that the RGB values of the cotton fruits differ from that of the other parts of the plants.
Blob (binary large object) analysis exploits color and shape information for object detection, as described in the article by Ganesan et al. [156]. In the research performed by Dewi et al. [157,158], a BLOB analysis was utilized to group pixels with similar brightness and color and detect the fruits.

4.1.2. Shape Analysis and Morphological Features

Shape analysis within an image extracts meaningful information and features by identifying, describing, and quantifying the geometric characteristics of objects in an image. This process is crucial in various applications such as object recognition, image retrieval, and classification. Various techniques utilized in agriculture for the shape analysis of crops are described in this section. For example, edge detection methods such as Sobel, Prewitt, Canny, and the Laplacian edge detection methods are described in the papers by Dewi et al. [157,158], Bulanon et al. [159], and Fernandes et al. [160], which detect the edges of objects or features within an image to define their boundaries. Extracting the contour or boundary of objects requires techniques like the Douglas–Peucker method used by Fernandes et al. [160], the split and merge algorithm used in the study by Zhang et al. [161], or the Probabilistic Hough Transform algorithm applied by Ge et al. [162].

4.1.3. Segmentation Techniques

Image segmentation is the process of performing a set of operations, such as deleting, adding, or enhancing certain parts of an image, with the aim of separating and distinguishing the desired area of interest, as described by Kodagali et al. [163]. In most cases, image segmentation is the very first image processing operation carried out during the process of object recognition. Thresholding and histogram analysis segmentation techniques were utilized to detect targets by Dewi et al. [157]. In the research work conducted by Vasconcelos et al. [164], semi-autonomous annotation of the acquired dataset was performed. Different segmentation networks were utilized to train the acquired dataset, and benchmark results were obtained. Some researchers aimed to develop a synthetic dataset to advance computer vision performance in agricultural robotics. A synthetic dataset was created to generate an exhaustive image dataset, which was difficult to obtain using the manual data acquisition process. Barth et al. [165] used semantic segmentation to detect objects and compare the similarity between synthetic and original images.
In the published study by Jiao et al. [166], the acquired image is transformed into Lab color space, and the K-means segmentation algorithm is utilized. The outlines of apples are extracted using erosion and dilation techniques (morphological processes). The minimum distance of internal points from the outline is calculated by applying a fast algorithm. The maximum distance among the points obtained above represents the center of the apple. Subsequently, radii are calculated, successfully finishing the localization of the overlapped apple quickly.

4.1.4. Deep Learning for 2D Image-Based Feature Extraction

Single-stage deep learning combined with other image-processing techniques have been utilized to distinguish between ripe and unripe fruits in complex environments. In the research by Wang et al. [167], the authors added the ECA attention mechanism to the backbone network of YOLOv8 to enhance its performance. Focal-EIOU loss was applied to the loss function, which aided in balancing easy- and difficult-to-classify samples. A similar method has been applied by Bu et al. [168], employing the YOLOv4 detection algorithm. This paper resulted in an acceptable processing speed and target detection. After observing the results, the authors noted that errors in detection may have been caused by the small bounding boxes. In the future, preprocessing of the images shall be performed by scaling and cropping the acquired images. Also, the prediction results are often based on the number of training images used to train the model. The image preprocessing method based on YOLOv5, involving HSV color space transformation, thresholding, edge detection using the Canny operator, and a combination of image processing and deep learning methods, was performed to determine the exact cutting point on the stem in the article by Miao et al. [169]. Another research team, Yin et al. [170], also utilized YOLOv4 tiny to detect citrus fruits and fruit navel point locations. Onishi et al. [171] utilized a single deep neural network SSD based on CNN for feature extraction due to its superiority in speed and accuracy compared to other single deep neural network methods. The images were captured from below to reduce occlusion by leaves, branches, and other fruits. However, fruits on the edge of the image or at a farther distance from the camera could not be detected using the above method.
Semantic segmentation using DeepLabV3+ was employed to detect strawberries and plant features in the study conducted by Fujinaga et al. [172]. This preprocessing did not aid in distinguishing ripe and unripe fruits. Other plant features were extracted by applying classical image processing, which essentially involves the color (conversion of the RGB color space to the HSV color space) and shape of the plant features. A similar approach was adopted to distinguish litchi from plants by Peng et al. [173]. A convolutional neural network was utilized for image processing. The DeepLabV3+ model with Atrous Spatial Pyramid Pooling (ASPP) and the Xception feature extraction network were adopted for image segmentation. Fujinaga et al. [174] utilized ResNet-50 as the backbone network for DeepLabV3+ semantic segmentation. The study also utilized ASPP, aiding in feature extraction. The study aimed to detect and distinguish between parts of the plant but did not perform any research on distinguishing between ripe and unripe fruits. The results obtained from the semantic segmentation were further processed to reduce noise (remove noise based on the area size) and false positives (morphological operations, skeletonization [55] feature extraction methods popular in 3D point clouds, and dilation—only for trusses). Another study by Giang et al. [175] fused depth data from an RGB-D camera for the fast detection of tomato suckers in real time by training a two-stage neural network.
The study performed by Kalampokas et al. [176] evaluated different CNN models to compare the segmentation accuracy of each model. Another study by Ge et al. [162] utilized deep CNN (Mask R-CNN) to detect both bounding boxes and masks for each instance, for instance, the segmentation of images to detect strawberries. Localization accuracy was improved by implementing density-based point clustering that removed noise points in 3D point clouds. The study conducted by Lin et al. [177] utilized a fine-tuning strategy to train a fully convolutional neural network (FCN) to segment guava and its branches. Further processing was performed to count the number of detected fruits and to estimate the fruit pose. In the paper by Onishi et al. [170], instance segmentation was achieved by employing Mask R-CNN with ResNet-101 as the backbone and the original model of MS COCO. Principal Component Analysis (PCA) was employed to estimate the orientation of fruit and stem masks. Another study by Hussain et al. [178] utilized DeepLabV3 for semantic segmentation to detect litchi and twigs. Later, morphological processes were applied to detect fruit-bearing branches. The picking position of the fruit-bearing branch was obtained using spatial clustering (revised density-based clustering of applications with noise (DBSCAN)) and Principal Component Analysis (PCA) straight-line fitting. Table 10 summarizes the contributions and limitations of various 2D-based feature extraction techniques.

4.2. Three-Dimensional (3D) Point Cloud-Based Plant Feature Extraction

In this section, we advance to point cloud data processing for plant structures, focusing on the complexities involved in 3D point cloud processing. Major challenges in processing 3D plant data include shape ambiguity and occlusion, which complicate the accurate representation and analysis of plant structures. There are several techniques and data processing pipelines for 3D image acquisition of plants. Briefly, the data acquisition techniques can be classified as active and passive approaches [180]. In an active approach, the sensors emit their own energy towards the object and detect the reflected signals. The characteristic of the object is formed based on the time delay and intensity of the reflected signal. A common example is a LiDAR (Light Detection and Ranging) system that emits laser pulses and detects reflected pulses to calculate the distance to the object. Another example is the work by Um [181], introducing a novel method to calculate the 3D depth and the surface angle of an object at the same time. The enabling technology is the multiple intensity differentiation method. Single camera-based 3D depth technology demonstrated a real-time feature extracting and gesture processing as well. On the other hand, passive approaches do not use specific structured energy but rely on the light that is reflected, emitted, or scattered by objects to form an image [182]. Both approaches and sensing systems have their own advantages and disadvantages. Briefly, active sensors are relatively accurate and more expensive than passive systems. Data captured by both methods need to go through a series of processes to develop intermediate data products such as 3D point clouds, depth maps, and voxels from which 3D structures of the plants or canopy can be generated. Harandi et al. [180] and Paulus [68] conducted a detailed comparison of these 3D imaging processes/techniques with their advantages, disadvantages, and potential uses for field conditions or controlled environments. In summary, 3D laser scanners and imagery are the most commonly used data collection techniques for 3D plant phenotyping. LiDAR operates based on time of flight (ToF), where it determines the distance to an object by measuring the time it takes for a laser pulse to travel to the object and back. The distance is calculated using the speed of light multiplied by half of the time-of-flight duration. Some of the key advantages include faster data collection, light independence, high precision, and higher resolution data.
Panjvani et al. [183] designed a low-cost LiDAR-based framework (LiDARPheno) and developed a 3D model of plants to extract leaf traits such as length, width, and area. Patel et al. [184] utilized 3D point cloud data of sorghum plants obtained from LiDAR to test deep learning models for extracting phenotypic traits such as plant height, plant crown diameter, plant compactness, stem diameter, panicle length, and panicle width. Zhu et al. [185] explored a large-scale phenotyping approach using backpack LiDAR to acquire a 3D point cloud of a field and extract phenotypic features. Paulus et al. [186] tested the potential use of 3D laser scanners to generate point cloud data for grapevines and wheat. They were able to classify grapevines and leaves with 98% accuracy. Additionally, wheat ears were separated from other organs, and the ear volume was calculated and correlated with the ear weight, kernel number, and weight. Imaging techniques utilize depth cameras, Structure from Motion (SfM), multi-view stereo, or a combination to generate 3D architectures of plants. Han et al. [187] utilized a depth camera, Azure Kinect, mounted directly above melon seedlings to gather point clouds and proposed a novel point cloud segmentation framework for measuring leaf area.
The Microsoft Azure Kinect DK combines a high-precision time-of-flight (ToF) depth sensor with a 12 MP RGB color camera, a 7-microphone array for audio capture, and an Inertial Measurement Unit (IMU), including an accelerometer and gyroscope, all integrated into one device. It offers a depth resolution up to 1024 × 1024 pixels with a field of view of 120° × 120° and a depth range from 0.5 to 5 m. The 4K color camera captures at 3840 × 2160 pixels with a 75° × 65° field of view, supporting up to 30 frames per second. Intel RealSense depth cameras are also used in plant phenotyping [188]. Common models of Intel RealSense depth cameras include D415 and D435. D415 features a field of view of 69.4° × 42.5° × 77°, a depth resolution up to 1280 × 720 pixels, and a global shutter, making it ideal for capturing detailed scans of small objects or fast-moving features. In contrast, D435 offers a wider field of view at 91.2° × 65.5° × 100.6°, the same resolution, and a rolling shutter, which suits a broader scene understanding and provides a comprehensive view for larger environment reconstruction.
SfM is a photogrammetry and computer vision technique that uses a series of overlapping 2D images captured from multiple angles to reconstruct the 3D structure of an object [186,189,190]. Ghahremani et al. [5] used PhenoGreen rotary tables for capturing high-resolution multi-view images of grapevine and brassica plants. Collected images were processed using SfM algorithms to generate 3D point clouds. Rossi et al. [191] evaluated the performance of a low-cost phenotyping platform consisting of an RGB camera, rotating plates, and a sliding system. An RGB camera was installed on a sliding system consisting of two one-meter-long aluminum tubes, and plants were placed on the rotating plates. The images captured were further processed to generate 3D plants. A novel segmentation algorithm for the automatic collection of data capturing plant structural features such as height, angle, and area was proposed [192].
Gao et al. [193] performed a detailed study on the development of imaging systems for plant 3D modeling and compared factors such as the impacts of the number of cameras used to collect images, the number of images captured, the use of color checkboards during image capture, and the system’s stability. It was found that images captured using more than one camera resulted in better 3D models of plants. Multiple cameras at different heights were also used in systems developed by Tanabata et al. [194] and later adopted by Pongpiyapaiboon et al. [195]. Using a rotating system, the study found that at least 60 images are required to generate reliable 3D models. Additionally, increasing the number of images improves the model but comes with increased computational costs. Using a colored checkboard during data collection also improves the quality of 3D reconstruction. The system’s stability was higher when the camera was rotated compared to rotating the plants, especially when detecting non-rigid parts of the plants.
The evolution of 3D point cloud-based plant feature extraction has been significant as technology matures in fields like autonomous vehicles, robotics, and other automation applications. However, one of the primary challenges in plant 3D point cloud technology is incomplete data acquisition. Despite advancements in scanning technology, certain areas remain challenging or impossible to reach, resulting in gaps within the dataset and complicating the accurate analysis and understanding of plant structures. Additionally, extracting phenotypic parameters from point clouds presents challenges due to inaccuracies introduced by environmental factors and plant occlusions, which can adversely affect data accuracy. This is similar to measuring the size of a leaf while it is blown by the wind, complicating reliable parameter extraction, which in turn impacts plant breeding and crop management. Furthermore, point clouds exhibit rotation invariance, complicating the ascertainment of the point cloud’s orientation, akin to assembling a 3D puzzle without reference points. This poses a challenge in aligning point clouds with real-world coordinates, which is essential for precise analyses and measurements. The disordered and irregular nature of point clouds further complicates feature extraction, making it challenging to identify specific features, such as leaves or stems, which can result in inaccurate analyses and measurements.
Lastly, noise in data presents a significant challenge. Sensor errors and environmental factors can introduce noise, which complicates the extraction of accurate information. This noise can lead to incorrect analysis and measurement, with potential consequences for plant breeding and crop management. Gomathi et al. [196] introduced the Point Sampling Method to improve instance segmentation in point clouds, particularly for plant growth phenotyping in agriculture. Utilizing an occupancy grid representation and single symmetric function max pooling, they achieved semantic and leaf instance segmentation with remarkable accuracy, contributing to plant growth analysis and disease detection.
Continual semantic segmentation is essential for maintaining accurate spatial annotations in dynamic environments. Roggiolani et al. [197] proposed a self-supervised pre-training approach for 3D leaf instance segmentation, reducing labeling efforts in agricultural plant phenotyping. Their method combines this approach with automatic post-processing to handle overlapping petioles, showing consistent improvements in segmentation accuracy. LatticeNet [145] also addresses the challenges of 3D semantic segmentation of plant point clouds. The study employed data augmentation techniques to enhance plant part segmentation, demonstrating significant improvements. This work focuses on addressing leaf crossover challenges, contributing to plant architecture reconstruction. Li et al. [198] introduced PlantNet, a dual-function deep learning network for semantic and instance segmentation in multiple plant species from point clouds. Innovative strategies like 3D Edge-Preserving Sampling (3DEPS) and Local Feature Extraction Operation (LFEO) modules enhanced performance, as validated on datasets of tobacco, tomato, and sorghum, indicating its potential impacts on plant phenotype extraction and intelligent agriculture. Additional studies, such as Weyler et al. [7], further advanced the field with methods for joint plant instance detection and leaf count estimation in agricultural fields using mobile robots. They utilized a single-stage object detection method to localize crops and weeds while estimating the leaf count and tested it on sugar beet fields, where it outperformed traditional methods like Mask R-CNN.
Masuda [199] presents a novel approach for estimating the leaf area in tomato plants grown in a sunlight-type plant factory. Utilizing RGB-D sensors and point cloud segmentation techniques, Masuda achieves promising results, with a relative error of about 20% in the leaf area estimation, thereby contributing significantly to automated cultivation management. This method emphasizes the application of 3D point cloud technology in plant phenotyping, enabling more precise management of plant growth in controlled environments. Ghahremani et al. [5] focus on a direct point cloud analysis using the RANSAC algorithm. Their robust methodology achieves high accuracy in measuring critical plant attributes, such as branch diameters and leaf angles, further advancing the utility of point cloud technology in agricultural research. Gu et al. [199] introduce a method to analyze cabbage plant phenotypes through 3D point cloud segmentation. The ASAP-PointNet model, tested in this study, demonstrates improved semantic segmentation accuracy, with high IoU scores. The DBSCAN algorithm is employed to enhance cabbage instance segmentation, providing significant contributions to the field of phenotypic analysis.
Imabuchi et al. [200] propose a combined method for semantic and volumetric 3D modeling from point cloud data in plant environments. This method enhances accuracy in radiation dose distribution modeling, particularly in decommissioning scenarios, by integrating 2D image-based deep learning with volumetric reconstruction. Gu et al. [199] also present a virtual design-based 3D reconstruction technique for wheat plants, addressing limitations in traditional methods. By combining point cloud data with optimization algorithms, the study offers an effective approach to model the wheat plant architecture, despite challenges in precise reconstruction due to the absence of organ templates. Wang et al. [201] develop a method for segmenting tomato plant stems and leaves using multi-view image sequences from a solar greenhouse, extracting six key phenotypic parameters. The study highlights both superior segmentation accuracy and computational efficiency, though it also acknowledges the need for further research to address challenges like canopy adhesion.
Giang et al. [10] propose a method for estimating sweet pepper leaf area using semantic 3D point clouds generated from RGB-D images. Their technique achieves high accuracy and efficiency, with strong correlations between the point cloud data and actual leaf area. However, challenges remain in fully capturing taller plants and varying point cloud resolutions, indicating that further validation across diverse conditions is necessary. Table 11 summarizes the contributions and limitations of various 3D point cloud-based feature extraction techniques.

5. Remote Sensing and Plant Phenotyping: Insights

In this section, we present our experience in plant feature extraction, encompassing the entire process from RGB data collection and point cloud generation to 3D point cloud segmentation. Our aim is to provide valuable insights for scholars engaged in the study of plant structure classification and segmentation. We begin by detailing our methods for RGB data acquisition for both remote sensing and proximal sensing, followed by an explanation of image labeling and model training for 2D image-based segmentation. Finally, we discuss our approach to 3D point cloud-based plant structure segmentation, sharing the lessons learned and best practices identified throughout the process.

5.1. Canopy-Level Remote Sensing Data Collection and Analysis

To demonstrate remote sensing data collection and processing in the field, we captured raw RGB images of a sorghum field in Driscoll, Texas, and converted them into 3D point cloud data for plant phenotyping. The raw images obtained from the UAV were processed using Agisoft Metashape Software (2.1.3) (Agisoft, LLC, St. Petersburg, Russia), which utilizes a Structure from Motion (SfM) algorithm. This process generated geospatial data products, including a digital surface model (DSM) that represents the surface elevation of ground objects [204], orthomosaic images, and finally a 3D point cloud of plants (Figure 7).
The ground sampling distance (GSD) for the orthomosaic images was 2 cm per pixel. For the data collection experiment in the sorghum field, we used a DJI Phantom 4 Real-Time Kinematics (RTK) (SZ DJI Technology Co. Ltd., Shenzhen, China) as the primary platform. This UAV is equipped with a 1-inch 20 MP (megapixel) CMOS (complementary metal oxide-semiconductor) sensor (FC6310R) that captures RGB image data. Additionally, the Phantom 4 RTK features a GPS module that stores geospatial information for RTK corrections. To further enhance the accuracy of the geospatial data products, a ground control point (GCP) was installed in the field. The flight height during data collection was maintained at 120 m above the ground, with a 75 percent forward and side overlap. As show in Figure 8, raw RGD data are processed to generate 3D point cloud of a sorghum field. Environmental conditions, including wind and rainfall, are critical factors for ensuring high-quality data collection, essential for generating precise 3D point clouds and accurately extracting phenotypic traits from each plant section. Despite significant technical advancements in UAV technology, data collection across multi-acre fields remains a labor-intensive and high-cost operation. Further automation of UAV operations, including streamlined setup and autonomous data acquisition, will be a critical challenge to address in future remote sensing research.
In the following section, we demonstrate how to generate 3D point clouds and process 3D point cloud data for plant phenotyping.

5.2. Plant-Level 3D Data Collection for Phenotyping

In this study, a 3D data collection system was developed to capture detailed images of tomato plants at the plant level for a precise 3D reconstruction and canopy height analysis. The setup, as depicted in Figure 9, integrated a RAUBAY 360° motorized turntable with a 6 W, 110 V, 50/60 Hz motor operating at 15 rpm, paired with a speed controller for precise rotational control. To facilitate image capture at consistent angles, a Canon EOS 1000D camera with a 10 MP sensor was mounted on a rotating arm, which allowed for incremental rotations at various predefined angles around the plant (Figure 9a). The turntable rotation and image capture were synchronized through a custom Python script that communicated with an Arduino microcontroller via the PySerial library.
This script controlled a sequential “stop-take-go” action, ensuring that the camera captured high-quality images at set intervals as the turntable moved. To optimize image quality and minimize shadows or reflections, a professional-grade photo booth setup (Figure 9b) was constructed using a 47″ × 39″ × 78″ adjustable LED lightbox. The controlled lighting environment provided uniform illumination around the plant, reducing unwanted noise and artifacts during the image processing stage. Additionally, a target marker sheet from Alex Elvis Badillo [206] was placed on the base of the turntable, serving as a reference for improved alignment and key point detection in the subsequent photogrammetry process using Agisoft Metashape software.
The image acquisition process began by positioning the potted tomato plant on the RAUBAY turntable inside the photo booth. The Arduino microcontroller controlled the motor to rotate in small, incremental steps (5–10 degrees), ensuring sufficient overlap between consecutive images. After each rotational step, the Python script sent a command to the camera to capture a high-resolution image. This “stop-take-go” sequence was repeated for a complete 360-degree view around the plant, resulting in 60 to 70 high-quality overlapping images. The use of this automated system significantly enhanced the precision and efficiency of the image capture process, ensuring consistency in image overlap, angle increments, and focus.
The collected images shown in Figure 9c were processed using the Structure from Motion (SfM) technique in Agisoft Metashape software to generate a dense 3D point cloud of a plant (Figure 9d). Metashape aligns the images by identifying key features and matching points across overlapping regions. Once the sparse point cloud was generated, the software reconstructed a detailed dense point cloud using depth information from the matched points. This dense point cloud was then further refined to create an accurate 3D representation of the plant structure.

5.3. Two-Dimensional (2D) Image-Based Organ Level Feture Extraction

This section highlights the initial findings and future directions for improving the 2D image-based plant feature extraction process, addressing both the contributions and limitations identified during the study. The 2D image-based plant feature extraction processes are based on the experimental data captured in the indoor tomato farm at Texas A&M University—AgriLife Center. First, the collected images were sorted into a training dataset with manual annotation, followed by plant organ segmentation via CNN model training.

5.3.1. Image Annotation and Segmentation

The dataset images were annotated for semantic segmentation, focusing on branches, stems, and suckers, using LabelStudio (1.14.0). Each image annotation took approximately 2–4 min. Figure 10a illustrates the annotated image, where green, purple, and red masked areas represent the stem, branches, and suckers, respectively.
Model training was conducted over 50 epochs using a total of 262 images for training and validation. The training utilized full plant images, making suckers visible when observed closely. Consequently, the accuracy of the results, as measured by the F1 confidence score, did not exceed 0.65 in our experiment. The main issue was the lack of a sufficient dataset for training. Typically, a model needs to be trained with over 1000 images to achieve a higher F1 score

5.3.2. Model Training and Implementation

The YOLO v8 [167]-based machine learning model was trained on a 64-bit Windows 11 Pro operating system with 16 GB RAM on an Intel i7-9750H CPU. The training program was developed in Python 3.11, leveraging PyTorch 2.2.1, TorchVision 0.17.1, CUDA 1.21, and YOLO v8. Figure 10b demonstrates the F1–confidence curve for the trained model. The data obtained from the annotation process were used to train the machine learning model. Training was implemented on a high-performance computing system, ensuring efficient processing and model training. Despite the limited dataset, the initial results provided valuable insights into the segmentation capabilities of the model. The experimental study revealed several contributions and limitations, which are summarized in Table 12.

5.3.3. Future Directions in 2D Analysis

Future work will focus on enhancing the dataset by increasing the number of annotated images, specifically targeting branches, stems, and suckers. This will involve collecting additional images and annotating them to build a more comprehensive dataset. With a larger dataset, the model training can be improved, leading to higher accuracy in plant feature extraction. Additionally, exploring advanced segmentation techniques and integrating other machine learning models could further enhance the performance of the plant feature extraction process. For higher accuracy, training will need to focus specifically on images of branches, stems, and suckers rather than complete plant images, which is expected to enhance the prediction accuracy.

5.4. Organ-Level Feature Extraction Using High-Resolution 3D Data

PointNet, a pioneering deep learning architecture, directly processes point clouds without converting them into other representations like voxels or images, preserving inherent details but presenting unique challenges. By focusing on its application to the Pheno4D plant phenotyping dataset [207], we explored its advantages, limitations, and preprocessing steps to optimize performance in segmentation tasks. PointNet uses multi-layer perceptrons (MLPs) and a symmetric function to aggregate information from all points with classification and segmentation networks. The classification network classifies the entire point cloud into a single class, such as a chair or car, starting with a T-Net that learns an affine transformation matrix to align the input. Feature extraction is performed through multiple 1 × 1 convolutions, typically with increasing depths, followed by another T-Net for feature space alignment. A max-pooling operation aggregates the global feature vector, which then passes through fully connected layers with dropout and batch normalization, leading to the final classification layer. TransformNet includes the Input TransformNet and Feature TransformNet, outputting 3 × 3 and 64 × 64 affine transformation matrices, respectively. Both use convolutional layers, batch normalization, ReLU, fully connected layers, and regularization to maintain orthogonality. For segmentation, PointNet assigns specific labels to each point in the cloud. The architecture is similar to the classification model, involving input transform, feature extraction, max pooling for the global feature, and concatenation of global features with local features. Segmentation layers use 1 × 1 convolutions to predict semantic labels for each point, focusing on the semantic meaning of different regions.
PointNet’s advantages include preserving geometric details through direct processing, ensuring permutation invariance with symmetric functions, and robustness to noise. However, it initially struggled with capturing local geometric features, which was addressed later by PointNet++, and requires significant computational resources for large point clouds. Alternatives like PointNet++ improve local context capture through hierarchical feature learning. Graph Convolution Networks (GCNs) [107] perform convolutions based on dynamic point graphs, incorporating the neighboring point context but with higher computational complexity. Post-PointNet segmentation methods like Density-Based Spatial Clustering of Applications with Noise (DBSCAN) and Ordering points to identify the clustering structure (OPTICS) clustering refine and segment individual parts and are useful in complex scenarios like varying numbers of leaves on plants. Effective preprocessing is crucial for preparing point cloud data for PointNet, involving normalization, standardization, data augmentation, and downsampling.
Normalization centers point clouds at the origin and scales them to fit within a unit sphere, ensuring a consistent scale and uniform positioning, which is essential for geometric consistency. Standardization adjusts the point clouds so that each feature (x, y, and z coordinates) has a mean of zero and a standard deviation of one, stabilizing the learning process by reducing the impact of varying scales and distributions.
Data augmentation applies techniques like random rotation, scaling, and jittering to increase the diversity of the training data, helping to prevent overfitting and improving the model’s generalization ability. Downsampling reduces point clouds to a fixed number of points to lower computational complexity and ensure uniform input size. Voxel downsampling, where the 3D space is divided into a grid and points within each voxel are averaged, is a common method. The impact of preprocessing on PointNet’s performance is evident when comparing data distributions before and after normalization and standardization. The histograms in Figure 11 highlight the initial state of data consistency and distribution across training, validation, and test sets. In comparison, Figure 12 demonstrates significant improvements in these areas, which are crucial for ensuring that the model learns meaningful features and generalizes well to unseen data. PointNet represents a significant advancement in processing point cloud data directly, preserving raw geometric details, and achieving high performance. Its effectiveness relies on preprocessing steps like normalization, standardization, and augmentation, which ensure consistent and representative data, leading to improved performance and generalization. Future advancements like PointNet++ and Graph Convolution Networks (GCNs) build on these foundations, addressing PointNet’s limitations and pushing point cloud processing further.
In this approach, because plants have stems and leaves, training for the segmentation of every individual leaf is complex. We simplified this by dividing it into two classes: leaf and stem. Further individual segmentation of leaves is performed using clustering methods, which highlight unique characteristics and reveal the data structure. As the plant structure becomes more complex, the individual leaf segmentation is not performed properly (Figure 13). However, by integrating preprocessing techniques and leveraging PointNet for the initial segmentation followed by clustering methods for detailed segmentation, this approach is effective for complex scenarios like plant phenotyping, where accurately distinguishing individual parts is crucial. The training loss and accuracy of the leaf and stem classification via PointNet are shown in Figure 14.

6. Discussion

This paper presents a comprehensive review and demonstration of plant feature extraction and segmentation techniques utilizing multi-scale remote and proximal sensing platforms, ranging from canopy-level to organ-level resolution. The methods discussed emphasize significant technical advancements in phenotyping and agricultural monitoring, focusing on the integration of diverse sensing modalities to enhance the accuracy and efficiency of plant characterization. The integration of satellite, UAV, and ground-based sensors has enabled detailed analyses of plant traits, ranging from large-scale canopy assessments to precise organ-level phenotyping. Techniques such as multispectral and hyperspectral imaging, LiDAR, and thermal imaging have been particularly effective in providing multi-modal data for characterizing crop health, growth patterns, and the physiological status. However, despite these advancements, several challenges remain. For instance, while satellite-based systems offer extensive coverage and frequent data collection, they suffer from lower spatial resolution and are limited by cloud cover and weather conditions. On the other hand, UAV platforms provide high-resolution and flexible data collection capabilities but are constrained by the flight duration, battery life, and coverage area. Ground-based systems offer the most detailed phenotyping, but they are often labor-intensive and impractical for large-scale monitoring.
A critical gap in the current literature is the limited focus on integrating these diverse sensing platforms to create a unified phenotyping framework that can capture multi-scale plant traits across different spatial and temporal scales. While some studies have explored the combination of UAV and satellite data for canopy-level assessments, the integration of these methods with proximal sensing technologies, such as 3D point cloud-based segmentation for an organ-level analysis, is relatively underdeveloped.
The practical deployment of these advanced sensing and segmentation methods in real-world agricultural settings faces several challenges. Data processing, computation, and storage requirements are major constraints, particularly when dealing with high-resolution 3D data or dense point clouds from LiDAR and photogrammetry. High-throughput phenotyping generates vast amounts of data that require sophisticated storage solutions and a powerful computing infrastructure for a real-time analysis. Additionally, variability in environmental conditions, such as varying lighting, wind interference, and occlusions from overlapping plant organs, can introduce noise and reduce the accuracy of segmentation algorithms. Developing robust models that can adapt to these real-world conditions without compromising efficiency remains an open challenge.

7. Conclusions and Future Outlook

This literature review has presented a comprehensive analysis of plant feature extraction and segmentation techniques using 2D image-based and 3D point cloud-based data modalities, highlighting their contributions, strengths, and limitations. The study has shown how the use of advanced remote sensing technologies, such as multispectral, hyperspectral, LiDAR, and 3D imaging, enables high-throughput, multi-scale phenotyping from the canopy level down to the organ level. Techniques like image segmentation, shape analysis, and machine learning-based methods for 2D images, as well as direct 3D point cloud processing using neural networks, have made significant strides in precision agriculture. However, the practical implementation of these methods in real-world settings still faces numerous challenges, particularly in terms of data variability, computational requirements, and the integration of multi-modal datasets.
In addition to reviewing advancements in remote and proximal sensing for plant phenotyping, this paper demonstrates various plant feature extraction techniques, covering the entire process from data collection to feature extraction and segmentation at different levels of remote sensing studies. The insights gained from each experiment are expected to contribute to ongoing research in the field, inform future studies, and support the development of next-generation phenotyping platforms that integrate the strengths of both remote and proximal sensing technologies.
The significance of creating robust phenotyping models extends beyond improving crop yield and health monitoring. By linking high-resolution phenotypic traits to genomic data, researchers can better understand how different traits influence plant growth and resilience, ultimately leading to the development of new crop varieties that are more productive and adaptable to changing environmental conditions. Future research should also explore ways to optimize data processing and storage solutions for large-scale phenotyping studies, ensuring that these advanced methods can be deployed efficiently in diverse agricultural landscapes. Reducing the labor intensity of data collection in the field will also be essential to streamline the entire phenotyping process, enhancing efficiency and scalability in agricultural research.
Looking forward, the future of plant phenotyping should focus on creating integrated phenotyping platforms that combine the strengths of satellite, UAV, and ground-based systems with advanced data analytics and machine learning models. The use of deep learning frameworks, such as transformer-based architectures and graph neural networks for visual grounding, holds promise for achieving higher accuracy in complex phenotyping scenarios, such as dense canopies or intertwined plant organs. Furthermore, the integration of remote sensing data with in-depth genomic information will allow researchers to build predictive models that can link phenotypic traits to specific genetic markers, ultimately enhancing crop breeding efforts. By leveraging these interdisciplinary approaches, the field can move towards a more holistic and scalable phenotyping solution that supports both large-scale agricultural monitoring and detailed organ-level analysis for precision breeding and crop management.
One potential direction is the development of hybrid models that combine the strengths of 2D image-based approaches with 3D point cloud techniques to capture more comprehensive plant phenotypes. Developing hybrid approaches that integrate multi-modal data from diverse sources will enable more comprehensive and scalable phenotyping frameworks. Moreover, improving the segmentation performance for minority classes can be achieved through techniques such as synthetic data generation and adjusting class weights during training.
By overcoming these challenges, the field can move towards more scalable and comprehensive phenotyping frameworks that not only capture detailed plant traits but also contribute to sustainable agricultural management and breeding practices.

Author Contributions

Conceptualization: P.N. and D.U.; methodology: P.N.; software: P.N., N.V. and O.F.M.; validation: D.U., K.L. and M.B.; formal analysis and investigation: P.N., N.V. and O.F.M.; data curation: P.N., N.V. and O.F.M.; writing—original draft preparation: D.U.; writing—review and editing: P.N.; visualization: P.N., N.V. and O.F.M.; supervision: D.U., K.L. and M.B.; project administration and funding acquisition: D.U., K.L. and M.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the United States Department of Agriculture, National Institute of Food and Agriculture, under grant number USDA-NRCS-UAIP-22-NOFO0001178.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sishodia, R.P.; Ray, R.L.; Singh, S.K. Applications of Remote Sensing in Precision Agriculture: A Review. Remote Sens. 2020, 12, 3136. [Google Scholar] [CrossRef]
  2. Liu, J.; Zhou, Z.; Li, B. Editorial: Remote Sensing for Field-Based Crop Phenotyping. Front. Plant Sci. 2024, 15, 1368694. [Google Scholar] [CrossRef]
  3. Araus, J.L.; Cairns, J.E. Field high-throughput phenotyping: The new crop breeding frontier. Trends Plant Sci. 2014, 19, 52–61. [Google Scholar] [CrossRef] [PubMed]
  4. Zero, E.; Sacile, R.; Trinchero, D.; Fossa, M. Smart Sensors and Smart Data for Precision Agriculture: A Review. Sensors 2024, 24, 2647. [Google Scholar] [CrossRef] [PubMed]
  5. Ghahremani, M.; Williams, K. Direct and accurate feature extraction from 3D point clouds of plants using RANSAC. Comput. Electron. Agric. 2022, 184, 40–47. [Google Scholar] [CrossRef]
  6. Huu, T.N.; Lee, S. Development of volumetric image descriptor for urban object classification using 3D LiDAR based on convolutional neural network. In Proceedings of the 2022 22nd International Conference on Control, Automation, and Systems, Seoul, Republic of Korea, 27–30 November 2022; pp. 1442–1448. [Google Scholar]
  7. Weyler, J.; Milioto, A.; Falck, T.; Behley, J.; Stachniss, C. Joint plant instance detection and leaf count estimation for in-field plant phenotyping. IEEE Robot. Autom. Lett. 2021, 6, 314–321. [Google Scholar] [CrossRef]
  8. Xin, B.; Sun, J.; Bartholomeus, H.; Kootstra, G. 3D data-augmentation methods for semantic segmentation of tomato plant parts. Front. Plant Sci. 2023, 14, 1045545. [Google Scholar] [CrossRef]
  9. Munisami, T.; Ramsurn, M.; Kishnah, S.; Pudaruth, S.C. Plant leaf recognition using shape features and colour histogram with K-nearest neighbour classifiers. Procedia Comput. Sci. 2015, 58, 740–747. [Google Scholar] [CrossRef]
  10. Giang TT, H.; Ryoo, Y.-J. Sweet pepper leaf area estimation using semantic 3D point clouds based on semantic segmentation neural network. AgriEngineering 2024, 6, 645–656. [Google Scholar] [CrossRef]
  11. Kalyoncu, C.; Toygar, Ö. GTCLC: Leaf classification method using multiple descriptors. IET Comput. Vis. 2016, 10, 700–708. [Google Scholar] [CrossRef]
  12. Fuentes-Peñailillo, F.; Gutter, K.; Vega, R.; Carrasco Silva, G. Transformative Technologies in Digital Agriculture: Leveraging IoT, Remote Sensing, and AI for Smart Crop Management. J. Sens. Actuator Netw. 2024, 13, 39. [Google Scholar] [CrossRef]
  13. Xu, Y.; Shrestha, V.; Piasecki, C.; Wolfe, B.; Hamilton, L.; Millwood, R.J.; Mazarei, M.; Stewart, C.N. UAV-Based Remote Sensing for Automated Phenotyping of Field-Grown Switchgrass: Modeling Sustainability Traits. Plants 2024, 10, 2726. [Google Scholar] [CrossRef]
  14. Borah, K.; Das, H.S.; Seth, S. A review on advancements in feature selection and feature extraction for high-dimensional NGS data analysis. Funct. Integr. Genom. 2024, 24, 139. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, C.; Liu, B.; Liu, L. A review of deep learning used in the hyperspectral image analysis for agriculture. Artif. Intell. Rev. 2021, 54, 5205–5253. [Google Scholar] [CrossRef]
  16. Pineda, M.; Barón, M.; Pérez-Bueno, M.-L. Thermal Imaging for Plant Stress Detection and Phenotyping. Remote Sens. 2021, 13, 68. [Google Scholar] [CrossRef]
  17. Dong, L.; Jiao, N.; Zhang, T.; Liu, F.; You, H. GPU Accelerated Processing Method for Feature Point Extraction and Matching in Satellite SAR Images. Appl. Sci. 2024, 14, 1528. [Google Scholar] [CrossRef]
  18. Farhan, S.M.; Yin, J.; Chen, Z.; Memon, M.S. A Comprehensive Review of LiDAR Applications in Crop Management for Precision Agriculture. Sensors 2024, 24, 5409. [Google Scholar] [CrossRef]
  19. Gano, B.; Bhadra, S.; Vilbig, J.M.; Ahmed, N.; Sagan, V.; Shakoor, N. Drone-based imaging sensors, techniques, and applications in plant phenotyping for crop breeding: A comprehensive review. Plant Phenome J. 2024, 7, e20100. [Google Scholar] [CrossRef]
  20. Wen, T.; Li, J.-H.; Wang, Q.; Gao, Y.; Hao, G.-F.; Song, B.-A. Thermal imaging: The digital eye facilitates high-throughput phenotyping traits of plant growth and stress responses. Sci. Total Environ. 2023, 899, 165626. [Google Scholar] [CrossRef]
  21. Yuan, H.; Bennett, R.S.; Wang, N.; Chamberlin, K. Development of a Peanut Canopy Measurement System Using a Ground-Based LiDAR Sensor. Front. Plant Sci. 2019, 10, 203. [Google Scholar] [CrossRef]
  22. Zhou, J.; Li, F.; Wang, X.; Yin, H.; Zhang, W.; Du, J.; Pu, H. Hyperspectral and Fluorescence Imaging Approaches for Nondestructive Detection of Rice Chlorophyll. Plants 2024, 13, 1270. [Google Scholar] [CrossRef] [PubMed]
  23. Ijaz, S.; Ul Haq, I.; Babar, M. Fluorescent Imaging System-Based Plant Phenotyping for Disease Recognition. In Trends in Plant Disease Assessment; Ul Haq, I., Ijaz, S., Eds.; Springer: Singapore, 2022; pp. 189–206. [Google Scholar] [CrossRef]
  24. Rouse, J.W.; Haas, R.H.; Schell, J.A. Monitoring vegetation systems in the Great Plains with ERTS. NASA Spec. Publ. 1974, 351, 309. [Google Scholar]
  25. Vekerdy, Z.; Laake, P.E.; Timmermans, J.; Dost, R. Canopy Structural Modeling Based on Laser Scanning; International Institute for Geo-Information Science and Earth Observation: Enschede, The Netherlands, 2007. [Google Scholar]
  26. Gutierrez, M.; Reynolds, M.P.; Klatt, A.R. Association of water spectral indices with plant and soil water relations in contrasting wheat genotypes. J. Exp. Bot. 2010, 61, 3291–3303. [Google Scholar] [CrossRef] [PubMed]
  27. Musungu, K.; Dube, T.; Smit, J.; Shoko, M. Using UAV multispectral photography to discriminate plant species in a seep wetland of the Fynbos Biome. Wetl. Ecol. Manag. 2024, 32, 207–227. [Google Scholar] [CrossRef]
  28. Jiang, Z.; Huete, A.; Didan, K.; Miura, T. Development of a two-band enhanced vegetation index without a blue band. Remote Sens. Environ. 2008, 112, 3833–3845. [Google Scholar] [CrossRef]
  29. Huete, A.R. A soil-adjusted vegetation index (SAVI). Remote Sens. Environ. 1988, 25, 295–309. [Google Scholar] [CrossRef]
  30. Sharma, L.K.; Bu, H.; Denton, A.; Franzen, D.W. Active-optical sensors using red NDVI compared to red edge NDVI for prediction of corn grain yield in North Dakota, U.S.A. Sensors 2015, 15, 27832–27853. [Google Scholar] [CrossRef]
  31. Peñuelas, J.; Garbulsky, M.F.; Filella, I. Photochemical reflectance index (PRI) and remote sensing of plant CO₂ uptake. New Phytol. 2011, 191, 596–599. [Google Scholar] [CrossRef]
  32. Coswosk, G.G.; Gonçalves, V.M.L.; de Lima, V.J.; de Souza, G.A.R.; Junior, A.T.D.A.; Pereira, M.G.; de Oliveira, E.C.; Leite, J.T.; Kamphorst, S.H.; de Oliveira, U.A.; et al. Utilizing Visible Band Vegetation Indices from Unmanned Aerial Vehicle Images for Maize Phenotyping. Remote Sens. 2024, 16, 3015. [Google Scholar] [CrossRef]
  33. McFeeters, S.K. The use of the normalized difference Water Index (NDWI) in the delineation of open water features. Int. J. Remote Sens. 1996, 17, 1425–1432. [Google Scholar] [CrossRef]
  34. Tucker, C.J. Red and photographic infrared linear combinations for monitoring vegetation. Remote Sens. Environ. 1979, 8, 127–150. [Google Scholar] [CrossRef]
  35. Rasmussen, J.; Ntakos, G.; Nielsen, J.; Svensgaard, J.; Poulsen, R.N.; Christensen, S. Are vegetation indices derived from consumer-grade cameras mounted on UAVs sufficiently reliable for assessing experimental plots? Eur. J. Agron. 2016, 74, 75–92. [Google Scholar] [CrossRef]
  36. Sripada, R.P.; Heiniger, R.W.; White, J.G.; Meijer, A.D. Aerial Color Infrared Photography for Determining Early In-Season Nitrogen Requirements in Corn. Agron. J. 2006, 98, 968–977. [Google Scholar] [CrossRef]
  37. Gitelson, A.A.; Viña, A.; Ciganda, V.; Rundquist, D.C.; Arkebauer, T.J. Remote estimation of canopy chlorophyll content in crops. Geophys. Res. Lett. 2005, 32, L08403. [Google Scholar] [CrossRef]
  38. Numbisi, F.N.; Coillie, F.V.; Wulf, R.R. Delineation of Cocoa Agroforests Using Multiseason Sentinel-1 SAR Images: A Low Grey Level Range Reduces Uncertainties in GLCM Texture-Based Mapping. ISPRS Int. J. Geo-Inf. 2019, 8, 179. [Google Scholar] [CrossRef]
  39. Lhermitte, E.; Hilal, M.; Furlong, R.; O’Brien, V.; Humeau-Heurtier, A. Deep Learning and Entropy-Based Texture Features for Color Image Classification. Entropy 2022, 24, 1577. [Google Scholar] [CrossRef]
  40. Javidan, S.M.; Banakar, A.; Rahnama, K.; Vakilian, K.A.; Ampatzidis, Y. Feature engineering to identify plant diseases using image processing and artificial intelligence: A comprehensive review. Smart Agric. Technol. 2024, 8, 100480. [Google Scholar] [CrossRef]
  41. Ashfaq, M.; Ahmad, W.; Shah, S.M.; Ilyas, M. Citrus Leaf Disease Detection Using Joint Scale Local Binary Pattern. Sindh Univ. J. Inf. Commun. Technol. 2021, 4, 199–206. Available online: https://sujo.usindh.edu.pk/index.php/USJICT/article/view/3292 (accessed on 21 April 2024).
  42. Lang, N.; Jetz, W.; Schindler, K.; Wegner, J.D. A high-resolution canopy height model of the Earth. Nat. Ecol. Evol. 2023, 7, 1778–1789. [Google Scholar] [CrossRef]
  43. Hasselquist, N.J.; Benegas, L.; Roupsard, O.; Malmer, A.; Ilstedt, U. Canopy cover effects on local soil water dynamics in a tropical agroforestry system: Evaporation drives soil water isotopic enrichment. Hydrol. Process. 2018, 32, 994–1004. [Google Scholar] [CrossRef]
  44. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Maimaitiyiming, M.; Hartling, S.; Peterson, K.T.; Maw, M.J.W.; Shakoor, N.; Mockler, T.; Fritschi, F.B. Vegetation Index Weighted Canopy Volume Model (CVMVI) for soybean biomass estimation from Unmanned Aerial System-based RGB imagery. ISPRS J. Photogramm. Remote Sens. 2019, 151, 27–41. [Google Scholar] [CrossRef]
  45. Adak, A.; Murray, S.C.; Washburn, J.D. Deciphering temporal growth patterns in maize: Integrative modeling of phenotype dynamics and underlying genomic variations. New Phytol. 2024, 242, 121–136. [Google Scholar] [CrossRef] [PubMed]
  46. Zeng, L.; Wardlow, B.D.; Xiang, D.; Hu, S.; Li, D. A review of vegetation phenological metrics extraction using time-series, multispectral satellite data. Remote Sens. Environ. 2020, 237, 111511. [Google Scholar] [CrossRef]
  47. Sabie, R.; Bawazir, A.S.; Buenemann, M.; Steele, C.; Fernald, A. Calculating Vegetation Index-Based Crop Coefficients for Alfalfa in the Mesilla Valley, New Mexico Using Harmonized Landsat Sentinel-2 (HLS) Data and Eddy Covariance Flux Tower Data. Remote Sens. 2024, 16, 2876. [Google Scholar] [CrossRef]
  48. Tang, B.; Xie, W.; Meng, Q.; Moorhead, R.J.; Feng, G. Soil Moisture Estimation Using Hyperspectral Imagery Based on Metric Learning. In Proceedings of the 21st IEEE International Conference on Machine Learning and Applications (ICMLA), Nassau, Bahamas, 12–14 December 2022; pp. 1392–1396. [Google Scholar]
  49. Guo, Z.; Cai, D.; Zhou, Y.; Xu, T.; Yu, F. Identifying rice field weeds from unmanned aerial vehicle remote sensing imagery using deep learning. Plant Methods 2024, 20, 105. [Google Scholar] [CrossRef] [PubMed]
  50. Selvaraj, M.G.; Valderrama, M.; Guzman, D.; Valencia, M.; Ruiz, H.; Acharjee, A. Machine learning for high-throughput field phenotyping and image processing provides insight into the association of above and below-ground traits in cassava (Manihot esculenta Crantz). Plant Methods 2020, 16, 87. [Google Scholar] [CrossRef]
  51. Bhugra, S.; Srivastava, S.; Kaushik, V.; Mukherjee, P.; Lall, B. Plant Data Generation with Generative AI: An Application to Plant Phenotyping. In Applications of Generative AI; Lyu, Z., Ed.; Springer: Cham, Switzerland, 2024. [Google Scholar] [CrossRef]
  52. Zhang, Y.; Wu, H.; Yang, W. Forests Growth Monitoring Based on Tree Canopy 3D Reconstruction Using UAV Aerial Photogrammetry. Forests 2019, 10, 1052. [Google Scholar] [CrossRef]
  53. Ni, Z.; Burks, T.F.; Lee, W.S. 3D Reconstruction of Plant/Tree Canopy Using Monocular and Binocular Vision. J. Imaging 2016, 2, 28. [Google Scholar] [CrossRef]
  54. Bhandari, M.; Baker, S.; Rudd, J.C.; Ibrahim, A.M.H.; Chang, A.; Xue, Q.; Jung, J.; Landivar, J.; Auvermann, B. Assessing the Effect of Drought on Winter Wheat Growth Using Unmanned Aerial System (UAS)-Based Phenotyping. Remote Sens. 2021, 13, 1144. [Google Scholar] [CrossRef]
  55. Wu, S.; Wen, W.; Xiao, B.; Guo, X.; Du, J.; Wang, C.; Wang, Y. An accurate skeleton extraction approach from 3D point clouds of maize plants. Front. Plant Sci. 2019, 10, 248. [Google Scholar] [CrossRef]
  56. Yu, S.; Fan, J.; Lu, X.; Wen, W.; Shao, S.; Guo, X.; Zhao, C. Hyperspectral Technique Combined with Deep Learning Algorithm for Prediction of Phenotyping Traits in Lettuce. Front. Plant Sci. 2022, 13, 927832. [Google Scholar] [CrossRef] [PubMed]
  57. Zhang, H.; Ge, Y.; Xie, X.; Atefi, A.; Wijewardane, N.K.; Thapa, S. High throughput analysis of leaf chlorophyll content in sorghum using RGB, hyperspectral, and fluorescence imaging and sensor fusion. Plant Methods 2022, 18, 60. [Google Scholar] [CrossRef] [PubMed]
  58. Luo, B.; Sun, H.; Zhang, L.; Chen, F.; Wu, K. Advances in the tea plants phenotyping using hyperspectral imaging technology. Front. Plant Sci. 2024, 15, 1442225. [Google Scholar] [CrossRef] [PubMed]
  59. Yuan, T.; Xu, C.G.; Ren, Y.X.; Feng, Q.C.; Tan, Y.Z.; Li, W. Detecting the information of cucumber in greenhouse for picking based on NIR image. Spectrosc. Spectr. Anal. 2009, 29, 2054–2058. [Google Scholar] [PubMed]
  60. Abboud, T.; Hedjam, R.; Noumeir, R.; Berinstain, A. Segmentation of imaged plants captured by a fluorescent imaging system. In Proceedings of the 25th IEEE Canadian Conference on Electrical and Computer Engineering (CCECE), Montreal, QC, Canada, 29 April–2 May 2012; pp. 1–4. [Google Scholar]
  61. Janssens, O.; Vylder, J.D.; Aelterman, J.; Verstockt, S.; Philips, W.; Straeten, D.V.; Hoecke, S.V.; Walle, R.V. Leaf segmentation and parallel phenotyping for the analysis of gene networks in plants. In Proceedings of the 21st European Signal Processing Conference (EUSIPCO 2013), Marrakech, Morocco, 9–13 September 2013; pp. 1–5. [Google Scholar]
  62. Wang, K.; Pu, X.; Li, B. Automated Phenotypic Trait Extraction for Rice Plant Using Terrestrial Laser Scanning Data. Sensors 2024, 24, 4322. [Google Scholar] [CrossRef] [PubMed]
  63. Andayani, U.; Sumantri, I.B.; Arisandy, B. Classification of Zingiber Plants Based on Stomate Microscopic Images Using Probabilistic Neural Network (PNN) Algorithm. J. Phys. Conf. Ser. 2020, 1566, 012122. [Google Scholar] [CrossRef]
  64. Uhrmann, F.; Hügel, C.; Paris, S.; Scholz, O.; Zollhöfer, M.; Greiner, G. A Model-based Approach to Extract Leaf Features from 3D Scans. In Proceedings of the 7th International Conference on Functional-Structural Plant Models, Saariselkä, Finland, 9–14 June 2013. [Google Scholar]
  65. Nguyen, T.T.; Slaughter, D.C.; Max, N.; Maloof, J.N.; Sinha, N. Structured Light-Based 3D Reconstruction System for Plants. Sensors 2015, 15, 18587–18612. [Google Scholar] [CrossRef]
  66. Son, H.; Kim, C.; Kim, C. Fully Automated As-Built 3D Pipeline Extraction Method from Laser-Scanned Data Based on Curvature Computation. J. Comput. Civ. Eng. 2015, 29, B4014003. [Google Scholar] [CrossRef]
  67. Bao, Y.; Tang, L.; Schnable, P.S.; Salas-Fernandez, M.G. Infield Biomass Sorghum Yield Component Traits Extraction Pipeline Using Stereo Vision. In Proceedings of the 2016 ASABE Annual International Meeting, Orlando, FL, USA, 17–20 July 2016. [Google Scholar]
  68. Paulus, S. Measuring Crops in 3D: Using Geometry for Plant Phenotyping. Plant Methods 2019, 15, 103. [Google Scholar] [CrossRef]
  69. Li, D.; Li, J.; Xiang, S.; Pan, A. PSegNet: Simultaneous Semantic and Instance Segmentation for Point Clouds of Plants. Plant Phenomics 2022, 2022, 9787643. [Google Scholar] [CrossRef]
  70. Heiwolt, K.; Öztireli, C.; Cielniak, G. Statistical shape representations for temporal registration of plant components in 3D. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 9587–9593. [Google Scholar]
  71. Peng, Y.; Lin, S.; Wu, H.; Cao, G. Point Cloud Registration Based on Fast Point Feature Histogram Descriptors for 3D Reconstruction of Trees. Remote Sens. 2023, 15, 3775. [Google Scholar] [CrossRef]
  72. Li, J.; Li, Q.; Qiao, J.; Li, L.; Yao, J.; Tu, J. Organ-Level Instance Segmentation of Oilseed Rape at Seedling Stage Based on 3D Point Cloud. Appl. Eng. Agric. 2024, 40, 151–164. [Google Scholar] [CrossRef]
  73. Akhtar, M.S.; Zafar, Z.; Nawaz, R.; Fraz, M.M. Unlocking Plant Secrets: A Systematic Review of 3D Imaging in Plant Phenotyping Techniques. Comput. Electron. Agric. 2024, 222, 109033. [Google Scholar] [CrossRef]
  74. Kim, J.Y.; Abdel-Haleem, H. Open-Source Electronics for Plant Phenotyping and Irrigation in Controlled Environment. Smart Agric. Technol. 2022, 3, 100093. [Google Scholar] [CrossRef]
  75. Clarke, J.L.; Qiu, Y.; Schnable, J.C. Experimental Design for Controlled Environment High-Throughput Plant Phenotyping. Methods Mol. Biol. 2022, 2539, 57–68. [Google Scholar] [CrossRef]
  76. Patil, M.; Soma, S. Identification of Growth Rate of Plant Based on Leaf Features Using Digital Image Processing Techniques. Int. J. Emerg. Technol. Adv. Eng. 2013, 3, 2250–2259. [Google Scholar]
  77. Khirade, S.D.; Patil, A.B. Plant Disease Detection Using Image Processing. In Proceedings of the 2015 International Conference on Computing Communication Control and Automation, Pune, India, 26–27 February 2015; pp. 768–771. [Google Scholar]
  78. Sabrol, H.; Satish, K.V. Tomato Plant Disease Classification in Digital Images Using Classification Tree. In Proceedings of the 2016 International Conference on Communication and Signal Processing (ICCSP), Lonere, India, 26–27 December 2016; pp. 1242–1246. [Google Scholar]
  79. Prakash, R.M.; Saraswathy, G.; Ramalakshmi, G.; Mangaleswari, K.; Kaviya, T. Detection of Leaf Diseases and Classification Using Digital Image Processing. In Proceedings of the 2017 International Conference on Innovations in Information, Embedded and Communication Systems (ICIIECS), Coimbatore, India, 17–18 March 2017; pp. 1–4. [Google Scholar]
  80. Bhujade, V.G.; Sambhe, V.; Banerjee, B. Digital Image Noise Removal Towards Soybean and Cotton Plant Disease Using Image Processing Filters. Expert Syst. Appl. 2024, 246, 123031. [Google Scholar] [CrossRef]
  81. Jagtap, S.B.; Hambarde, S.M. Agricultural Plant Leaf Disease Detection and Diagnosis Using Image Processing Based on Morphological Feature Extraction. IOSR J. VLSI Signal Process. 2014, 4, 24–30. [Google Scholar] [CrossRef]
  82. Meirista, E.; Mukhlash, I.; Setiyono, B.; Suryani, D.R.; Nurvitasari, E. Classification of Plants and Weeds in Multi-Leaf Image Using Support Vector Machine Based on Leaf Shape and Texture Features. In Proceedings of the International Conference on Science and Technology (ICST 2018), Yogyakarta, Indonesia, 7–8 August 2018. [Google Scholar]
  83. Liang, F.; Chen, H.; Cui, S.; Yang, L.; Wu, X. Detection Method of Vegetable Maturity Based on Neural Network and Bayesian Information Fusions. In Proceedings of the Sixth International Conference on Electronics and Information Engineering, Fukuoka, Japan, 24–26 March 2023. [Google Scholar]
  84. Thomkaew, J.; Intakosum, S. Plant Species Classification Using Leaf Edge Feature Combination with Morphological Transformations and SIFT Key Point. J. Image Graph. 2023, 2023, 91–97. [Google Scholar] [CrossRef]
  85. Labellapansa, A.; Yulianti, A.; Yuliani, A.T. Segmentation of Palm Oil Leaf Disease Using Zoning Feature Extraction. In Proceedings of the Proceedings of the Second International Conference on Science, Engineering and Technology, Oxford, UK, 8–10 November 2019. [Google Scholar]
  86. Prasad, S.; Kumar, P.; Hazra, R.; Kumar, A. Plant Leaf Disease Detection Using Gabor Wavelet Transform. In Swarm, Evolutionary, and Memetic Computing (SEMCCO 2012); Panigrahi, B.K., Das, S., Suganthan, P.N., Nanda, P.K., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7677, pp. 435–443. [Google Scholar] [CrossRef]
  87. Ahmed, A.A.N.; Haque, H.M.F.; Rahman, A.; Shatabda, S. Wavelet and Pyramid Histogram Features for Image-Based Leaf Detection. In Emerging Technologies in Data Mining and Information Security; Abraham, A., Dutta, P., Mandal, J., Bhattacharya, A., Dutta, S., Eds.; Springer: Berlin/Heidelberg, Germany, 2019; Volume 814, pp. 245–251. [Google Scholar] [CrossRef]
  88. Kiran, S.M.; Chandrappa, N.D. Plant Disease Identification Using Discrete Wavelet Transforms and SVM. J. Univ. Shanghai Sci. Technol. 2021, 23, 108–112. [Google Scholar]
  89. Bhardwaj, A.; Kaur, M.; Kumar, A. Recognition of Plants by Leaf Image Using Moment Invariant and Texture Analysis. Int. J. Innov. Appl. Stud. 2013, 3, 237–248. [Google Scholar]
  90. Adam, S.; Amir, A. Fruit Plant Leaf Identification Feature Extraction Using Zernike Moment Invariant (ZMI) and Methods Backpropagation. In Proceedings of the 2019 International Conference on Informatics, Multimedia, Cyber and Information System (ICIMCIS), Jakarta, Indonesia, 24–25 October 2019; pp. 225–230. [Google Scholar]
  91. Sulc, M.; Matas, J. Texture-Based Leaf Identification. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 185–200. [Google Scholar]
  92. Chaki, J.; Parekh, R.; Bhattacharya, S. Plant Leaf Recognition Using Texture and Shape Features with Neural Classifiers. Pattern Recognit. Lett. 2015, 58, 61–68. [Google Scholar] [CrossRef]
  93. Mohan, A.; Peeples, J. Lacunarity Pooling Layers for Plant Image Classification using Texture Analysis. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 16–22 June 2024; pp. 5384–5392. [Google Scholar]
  94. Sangeetha, M.; Kannan, S.R.; Boopathi, S.; Ramya, J.; Ishrat, M.; Sabarinathan, G. Prediction of Fruit Texture Features Using Deep Learning Techniques. In Proceedings of the 2023 4th International Conference on Smart Electronics and Communication (ICOSEC), Trichy, India, 20–22 September 2023; pp. 762–768. [Google Scholar] [CrossRef]
  95. Anggraini, R.A.; Wati, F.F.; Shidiq, M.J.; Suryadi, A.; Fatah, H.; Kholifah, D.N. Identification of herbal plant based on leaf image using GLCM feature and K-Means. Techno Nusa Mandiri 2020, 17, 71–78. [Google Scholar] [CrossRef]
  96. Saha, D.; Hamer, G.; Lee, J.Y. Development of Inter-Leaves Weed and Plant Regions Identification Algorithm using Histogram of Oriented Gradient and K-Means Clustering. In Proceedings of the International Conference on Research in Adaptive and Convergent Systems, Krakow, Poland, 20–23 September 2017. [Google Scholar]
  97. Jafari, A.; Mohtasebi, S.S.; Jahromi, H.E.; Omid, M. Color Feature Extraction by Means of Discriminant Analysis for Weed Segmentation. In Proceedings of the 2004 ASAE Annual Meeting, Ottawa, ON, Canada, 1–4 August 2004. [Google Scholar]
  98. Shigang, C.; Hongdou, C.; Fan, L.; Lili, Y.; Xing-li, W. Veins feature extraction for LED plant growth cabinet. In Proceedings of the 2016 35th Chinese Control Conference (CCC), Chengdu, China, 27–29 July 2016; pp. 4917–4920. [Google Scholar]
  99. Ambarwari, A.; Adrian, Q.J.; Herdiyeni, Y.; Hermadi, I. Plant species identification based on leaf venation features using SVM. TELKOMNIKA Telecommun. Comput. Electron. Control 2020, 18, 726–732. [Google Scholar] [CrossRef]
  100. Chen, G.; Meng, Y.; Lu, J.; Wang, D. Research on Color and Shape Recognition of Maize Diseases Based on HSV and OTSU Method. In Proceedings of the Conference on Control Technology and Applications, Buenos Aires, Argentina, 19–22 September 2016. [Google Scholar]
  101. Rojanarungruengporn, K.; Pumrin, S. Early Stress Detection in Plant Phenotyping Using CNN and LSTM Architecture. In Proceedings of the 2021 9th International Electrical Engineering Congress (iEECON), Pattaya, Thailand, 10–12 March 2021; pp. 389–392. [Google Scholar]
  102. Giuffrida, M.V.; Doerner, P.; Tsaftaris, S.A. Pheno-Deep Counter: A unified and versatile deep learning architecture for leaf counting. Plant J. 2018, 96, 880–890. [Google Scholar] [CrossRef]
  103. Ullah, S.; Panzarová, K.; Trtílek, M.; Lexa, M.; Máčala, V.; Neumann, K.; Altmann, T.; Hejátko, J.; Pernisová, M.; Gladilin, E. High-Throughput Spike Detection in Greenhouse Cultivated Grain Crops with Attention Mechanisms-Based Deep Learning Models. Plant Phenomics 2024, 6, 0155. [Google Scholar] [CrossRef] [PubMed]
  104. Li, B.; Guo, C. MASPC_Transform: A Plant Point Cloud Segmentation Network Based on Multi-Head Attention Separation and Position Code. Sensors 2022, 22, 9225. [Google Scholar] [CrossRef]
  105. Qi, C.; Chen, K.; Gao, J. A Vision Transformer-Based Robotic Perception for Early Tea Chrysanthemum Flower Counting in Field Environments. J. Field Robot. 2024. [Google Scholar] [CrossRef]
  106. Alajas, O.J.; Concepcion, R.S.; Bandala, A.A.; Sybingco, E.; Dadios, E.P.; Mendigoria, C.H.; Aquino, H.L. Grape Phaeomoniella chlamydospora Leaf Blotch Recognition and Infected Area Approximation Using Hybrid Linear Discriminant Analysis and Genetic Programming. In Proceedings of the 2022 IEEE 14th International Conference on Humanoid, Nanotechnology, Information Technology, Communication and Control, Environment, and Management (HNICEM), Boracay, Philippines, 1–4 December 2022; pp. 1–6. [Google Scholar]
  107. Wang, L.; Huang, Y.; Hou, Y.; Shan, J.; Zhang, S. Graph Attention Convolution for Point Cloud Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 10296–10305. [Google Scholar]
  108. SK, P.K.; Sumithra, M.G.; Saranya, N. Particle Swarm Optimization (PSO) with Fuzzy C-Means (PSO-FCM)–Based Segmentation and Machine Learning Classifier for Leaf Diseases Prediction. Concurr. Comput. Pract. Exp. 2021, 33, e5312. [Google Scholar] [CrossRef]
  109. Muthukannan, K.; Latha, P. A PSO Model for Disease Pattern Detection on Leaf Surfaces. Image Anal. Stereol. 2015, 34, 209–216. [Google Scholar] [CrossRef]
  110. Cristin, R.; Kumar, B.S.; Priya, C.; Kanagarathinam, K. Deep Neural Network-Based Rider-Cuckoo Search Algorithm for Plant Disease Detection. Artif. Intell. Rev. 2020, 53, 4993–5018. [Google Scholar] [CrossRef]
  111. Patel, S.P.; Rungta, R. Automatic Detection of Plant Leaf Disease Using K-Means Clustering and Segmentation. Int. J. Sci. Res. Eng. Technol. 2017, 6, 774–779. [Google Scholar]
  112. Kabir, R.; Jahan, S.; Islam, M.R.; Rahman, N.; Islam, M.R. Discriminant Feature Extraction Using Disease Segmentation for Automatic Leaf Disease Diagnosis. In Proceedings of the International Conference on Computing Advancements, Dhaka, Bangladesh, 10–12 January 2020. [Google Scholar]
  113. Rajani, S.; Veena, M.N. Medicinal Plants Segmentation Using Thresholding and Edge-Based Techniques. Int. J. Innov. Technol. Explor. Eng. 2019, 8, 24–30. [Google Scholar]
  114. Khan, M.A.; Akram, T.; Sharif, M.; Javed, K.; Raza, M.; Saba, T. An Automated System for Cucumber Leaf Diseased Spot Detection and Classification Using Improved Saliency Method and Deep Features Selection. Multimed. Tools Appl. 2020, 79, 18627–18656. [Google Scholar] [CrossRef]
  115. Harrap, M.J.M.; Rands, S.A.; Hempel de Ibarra, N.; Whitney, H.M. The Diversity of Floral Temperature Patterns, and Their Use by Pollinators. eLife 2017, 6, e31262. [Google Scholar] [CrossRef]
  116. Subramani, K.; Periyasamy, S.; Theagarajan, P. Double Line Clustering-Based Colour Image Segmentation Technique for Plant Disease Detection. Curr. Med. Imaging Rev. 2019, 15, 769–776. [Google Scholar] [CrossRef]
  117. Senthilkumar, C.; Kamarasan, M. An Optimal Weighted Segmentation with Hough Transform Based Feature Extraction and Classification Model for Citrus Disease. In Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–28 February 2020; pp. 215–220. [Google Scholar]
  118. Xia, C.; Lee, J.; Li, Y.; Chung, B.; Chon, T. In Situ Detection of Small-Size Insect Pests Sampled on Traps Using Multifractal Analysis. Opt. Eng. 2012, 51, 027001. [Google Scholar] [CrossRef]
  119. Deepika, K.C.; Ruth, I.; Keerthana, S.; Sathya Bama, B.; Avvailakshmi, S.; Vidhya, A. Robust Plant Recognition Using Graph Cut Based Flower Segmentation and PHOG Based Feature Extraction. In Proceedings of the 2012 International Conference on Machine Vision and Image Processing (MVIP), Tamil Nadu, India, 14–15 December 2012; pp. 44–47. [Google Scholar]
  120. Glezakos, T.J.; Tsiligiridis, T.A.; Yialouris, C.P. Piecewise Evolutionary Segmentation for Feature Extraction in Time Series Models. Neural Comput. Appl. 2012, 24, 243–257. [Google Scholar] [CrossRef]
  121. Koenderink, J.J.; van Doorn, A.J. Surface Shape and Curvature Scales. Image Vis. Comput. 1992, 10, 557–565. [Google Scholar] [CrossRef]
  122. Boissonnat, J.D.; Cazals, F. Smooth Surface Reconstruction via Natural Neighbour Interpolation of Distance Functions. In Proceedings of the Symposium on Computational Geometry (SCG ‘02), Barcelona, Spain, 5–7 June 2002; pp. 223–232. [Google Scholar] [CrossRef]
  123. Belongie, S.; Malik, J.; Puzicha, J. Shape Matching and Object Recognition Using Shape Contexts. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar] [CrossRef]
  124. Rusu, R.B.; Marton, Z.C.; Blodow, N.; Beetz, M. Learning Informative Point Classes for the Acquisition of Object Model Maps. In Proceedings of the 10th International Conference on Control Automation Robotics and Vision (ICARCV), Hanoi, Vietnam, 17–20 December 2008. [Google Scholar]
  125. Rusu, R.B.; Blodow, N.; Marton, Z.C.; Beetz, M. Aligning Point Cloud Views Using Persistent Feature Histograms. In Proceedings of the 21st IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Nice, France, 22–26 September 2008. [Google Scholar]
  126. Rusu, R.B.; Blodow, N.; Beetz, M. Fast Point Feature Histograms (FPFH) for 3D Registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation (ICRA), Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  127. Johnson, A.E.; Hebert, M. Using Spin Images for Efficient Object Recognition in Cluttered 3D Scenes. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 433–449. [Google Scholar] [CrossRef]
  128. Benenson, R.; Omran, M.; Hosang, J.; Schiele, B. Volumetric and Multi-View CNNs for Object Classification on 3D Data. arXiv 2016, arXiv:1604.03265. [Google Scholar] [CrossRef]
  129. Su, H.; Maji, S.; Kalogerakis, E.; Learned-Miller, E. Multi-view Convolutional Neural Networks for 3D Shape Recognition. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 945–953. [Google Scholar] [CrossRef]
  130. Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral Networks and Locally Connected Networks on Graphs. arXiv 2013, arXiv:1312.6203. [Google Scholar]
  131. Leng, L.; Zhong, Z.; Wang, R.; Yang, L. Feature-Based Deep Neural Networks for 3D Data Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 507–516. [Google Scholar] [CrossRef]
  132. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  133. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. Adv. Neural Inf. Process. Syst. 2017, 30, 1–10. [Google Scholar] [CrossRef]
  134. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 146. [Google Scholar] [CrossRef]
  135. Xu, M.; Ding, R.; Zhao, H.; Qi, X. Paconv: Position Adaptive Convolution with Dynamic Kernel Assembling on Point Clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 3173–3182. [Google Scholar]
  136. Xiang, T.; Zhang, C.; Song, Y.; Yu, J.; Cai, W. Walk in the Cloud: Learning Curves for Point Clouds Shape Analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 915–924. [Google Scholar]
  137. Zhao, H.; Jiang, L.; Jia, J.; Torr PH, S.; Koltun, V. Point Transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 16259–16268. [Google Scholar]
  138. Lai, X.; Liu, J.; Jiang, L.; Wang, L.; Zhao, H.; Liu, S.; Qi, X.; Jia, J. Stratified Transformer for 3D Point Cloud Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 8500–8509. [Google Scholar]
  139. Hou, J.; Dai, X.; He, Z.; Dai, A.; Nießner, M. Mask3D: Pre-Training 2D Vision Transformers by Learning Masked 3D Priors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 13510–13519. [Google Scholar]
  140. Mertoğlu, K.; Şalk, Y.; Sarıkaya, S.K.; Turgut, K.; Evrenesoğlu, Y.; Çevikalp, H.; Gerek, Ö.N.; Dutağacı, H.; Rousseau, D. PLANesT-3D: A New Annotated Dataset for Segmentation of 3D Plant Point Clouds. arXiv 2024, arXiv:2407.21150. [Google Scholar]
  141. Liu, Z.; Tang, H.; Lin, Y.; Han, S. Point-Voxel CNN for Efficient 3D Deep Learning. arXiv 2019, arXiv:1907.03739. [Google Scholar]
  142. Yan, X.; Zheng, C.; Li, Z.; Wang, S.; Cui, S. PointASNL: Robust Point Clouds Processing Using Nonlocal Neural Networks with Adaptive Sampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattlepp, WA, USA, 13–19 June 2020; pp. p.5589–5598. [Google Scholar]
  143. Kaul, C.; Pears, N.; Manandhar, S. FatNet: A Feature-Attentive Network for 3D Point Cloud Processing. In Proceedings of the 2020 International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 7211–7218. [Google Scholar]
  144. Xu, S.; Li, Y.; Zhao, J.; Zhang, B.; Guo, G. POEM: 1-bit Point-Wise Operations Based on Expectation-Maximization for Efficient Point Cloud Processing. arXiv 2021, arXiv:2111.13386. [Google Scholar]
  145. Rosu, R.A.; Schutt, P.; Quenzel, J.; Behnke, S. LatticeNet: Fast Point Cloud Segmentation Using Permutohedral Lattices. In Proceedings of the IEEE Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022. [Google Scholar]
  146. Zhang, J.; Chen, T.; Ding, D.; Ma, Z. G-PCC++: Enhanced Geometry-Based Point Cloud Compression. In Proceedings of the 31st ACM International Conference on Multimedia (MM ‘23), Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1352–1363. [Google Scholar] [CrossRef]
  147. Zeng, B.; Liu, B.; Li, H.; Liu, X.; Liu, J.; Chen, D.; Peng, W.; Zhang, B. FNeVR: Neural Volume Rendering for Face Animation. Adv. Neural Inf. Process. Syst. 2022, 35, 22451–22462. [Google Scholar]
  148. Poux, F.; Ponciano, J.-J. Self-Learning Ontology for Instance Segmentation of 3D Indoor Point Cloud. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2020, XLIII-B2, 309–316. [CrossRef]
  149. Van Natijne, A. GeoTiles.nl: Readymade Geodata with a Focus on the Netherlands. 2020. Available online: https://geotiles.nl/ (accessed on 18 March 2024).
  150. Xiao, Z.; Gao, J.; Lanyu, Z. Voxel Grid Downsampling for 3D Point Cloud Recognition. Modul. Mach. Tool Autom. Manuf. Technol. 2022, 11, 43–47. [Google Scholar]
  151. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2017, arXiv:1708.02002. [Google Scholar] [CrossRef]
  152. Chawla, N.; Bowyer, K.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-Sampling Technique. arXiv 2002, arXiv:1106.1813. [Google Scholar] [CrossRef]
  153. Velastegui, R.; Yang, L.; Han, D. The Importance of Color Spaces for Image Classification Using Artificial Neural Networks: A Review. In Proceedings of the International Conference on Computational Science and Its Applications, Cagliari, Italy, 13–16 September 2021; Springer International Publishing: Cham, Switzerland; pp. 70–83. [Google Scholar]
  154. Al-Mashhadani, Z.; Chandrasekaran, B. Autonomous ripeness detection using image processing for an agricultural robotic system. In Proceedings of the 2020 11th IEEE Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON), New York, NY, USA, 28–31 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 743–748. [Google Scholar] [CrossRef]
  155. Rao, U.S.N. Design of automatic cotton picking robot with machine vision using image processing algorithms. In Proceedings of the 2013 International Conference on Control, Automation, Robotics and Embedded Systems (CARE), Jabalpur, India, 16–18 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 1–5. [Google Scholar] [CrossRef]
  156. Ganesan, P.; Sathish, B.S.; Vasanth, K.; Sivakumar, V.G.; Vadivel, M.; Ravi, C.N. A Comprehensive Review of the Impact of Color Space on Image Segmentation. In Proceedings of the 2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS), Coimbatore, India, 15–16 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 962–967. [Google Scholar]
  157. Dewi, T.; Mulya, Z.; Risma, P.; Oktarina, Y. BLOB Analysis of an Automatic Vision-Guided System for a Fruit Picking and Placing Robot. Int. J. Comput. Vis. Robot. 2021, 11, 315–327. [Google Scholar] [CrossRef]
  158. Dewi, T.; Rusdianasari, R.; Kusumanto, R.D.; Siproni, S. Image Processing Application on Automatic Fruit Detection for Agriculture Industry. In Proceedings of the 5th FIRST T1 T2 2021 International Conference (FIRST-T1-T2 2021), Palembang, Indonesia, 20–21 October 2021; Atlantis Press: Dordrecht, The Netherlands, 2022; pp. 47–53. [Google Scholar]
  159. Bulanon, D.M.; Kataoka, T. Fruit Detection System and an End Effector for Robotic Harvesting of Fuji Apples. Agric. Eng. Int. CIGR J. 2010, 12, 203–210. [Google Scholar]
  160. Fernandes, L.; Shivakumar, B.R. Identification and Sorting of Objects Based on Shape and Colour Using Robotic Arm. In Proceedings of the 2020 Fourth International Conference on Inventive Systems and Control (ICISC), Coimbatore, India, 8–10 January 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 866–871. [Google Scholar]
  161. Zhang, Y.; Wu, L. Classification of Fruits Using Computer Vision and a Multiclass Support Vector Machine. Sensors 2012, 12, 12489–12505. [Google Scholar] [CrossRef]
  162. Ge, Y.; Xiong, Y.; Tenorio, G.L.; From, P.J. Fruit Localization and Environment Perception for Strawberry Harvesting Robots. IEEE Access 2019, 7, 147642–147652. [Google Scholar] [CrossRef]
  163. Kodagali, J.A.; Balaji, S. Computer Vision and Image Analysis Based Techniques for Automatic Characterization of Fruits: A Review. Int. J. Comput. Appl. 2012, 50, 6–12. [Google Scholar]
  164. Vasconcelos, G.J.Q.; Costa, G.S.R.; Spina, T.V.; Pedrini, H. Low-Cost Robot for Agricultural Image Data Acquisition. Agriculture 2023, 13, 413. [Google Scholar] [CrossRef]
  165. Barth, R.; IJsselmuiden, J.; Hemming, J.; Van Henten, E.J. Data Synthesis Methods for Semantic Segmentation in Agriculture: A Capsicum Annuum Dataset. Comput. Electron. Agric. 2018, 144, 284–296. [Google Scholar] [CrossRef]
  166. Jiao, Y.; Luo, R.; Li, Q.; Deng, X.; Yin, X.; Ruan, C.; Jia, W. Detection and Localization of Overlapped Fruits Application in an Apple Harvesting Robot. Electronics 2020, 9, 1023. [Google Scholar] [CrossRef]
  167. Wang, C.; Wang, H.; Han, Q.; Zhang, Z.; Kong, D.; Zou, X. Strawberry Detection and Ripeness Classification Using YOLOv8+ Model and Image Processing Method. Agriculture 2024, 14, 751. [Google Scholar] [CrossRef]
  168. Bu, L.; Chen, C.; Hu, G.; Sugirbay, A.; Sun, H.; Chen, J. Design and Evaluation of a Robotic Apple Harvester Using Optimized Picking Patterns. Comput. Electron. Agric. 2022, 198, 107092. [Google Scholar] [CrossRef]
  169. Miao, Z.; Yu, X.; Li, N.; Zhang, Z.; He, C.; Li, Z.; Deng, C.; Sun, T. Efficient Tomato Harvesting Robot Based on Image Processing and Deep Learning. Precis. Agric. 2023, 24, 254–287. [Google Scholar] [CrossRef]
  170. Yin, H.; Sun, Q.; Ren, X.; Guo, J.; Yang, Y.; Wei, Y.; Huang, B.; Chai, X.; Zhong, M. Development, Integration, and Field Evaluation of an Autonomous Citrus-Harvesting Robot. J. Field Robot. 2023, 40, 1363–1387. [Google Scholar] [CrossRef]
  171. Onishi, Y.; Yoshida, T.; Kurita, H.; Fukao, T.; Arihara, H.; Iwai, A. An Automated Fruit Harvesting Robot by Using Deep Learning. Robomech J. 2019, 6, 13. [Google Scholar] [CrossRef]
  172. Fujinaga, T. Strawberries Recognition and Cutting Point Detection for Fruit Harvesting and Truss Pruning. Precis. Agric. 2024, 25, 1–22. [Google Scholar] [CrossRef]
  173. Peng, H.; Xue, C.; Shao, Y.; Chen, K.; Xiong, J.; Xie, Z.; Zhang, L. Semantic Segmentation of Litchi Branches Using DeepLabV3+ Model. IEEE Access 2020, 8, 164546–164555. [Google Scholar] [CrossRef]
  174. Fujinaga, T.; Nakanishi, T. Semantic Segmentation of Strawberry Plants Using DeepLabV3+ for Small Agricultural Robot. In Proceedings of the 2023 IEEE/SICE International Symposium on System Integration (SII), Atlanta, GA, USA, 17–20 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
  175. Giang, T.T.H.; Khai, T.Q.; Im, D.Y.; Ryoo, Y.J. Fast Detection of Tomato Sucker Using Semantic Segmentation Neural Networks Based on RGB-D Images. Sensors 2022, 22, 5140. [Google Scholar] [CrossRef]
  176. Kalampokas, T.; Tziridis, K.; Nikolaou, A.; Vrochidou, E.; Papakostas, G.A.; Pachidis, T.; Kaburlasos, V.G. Semantic Segmentation of Vineyard Images Using Convolutional Neural Networks. In Proceedings of the 21st EANN (Engineering Applications of Neural Networks) 2020 Conference, Halkidiki, Greece, 5–7 June 2020; Springer International Publishing: Cham, Switzerland, 2020; pp. 292–303. [Google Scholar]
  177. Lin, G.; Tang, Y.; Zou, X.; Xiong, J.; Li, J. Guava Detection and Pose Estimation Using a Low-Cost RGB-D Sensor in the Field. Sensors 2019, 19, 428. [Google Scholar] [CrossRef]
  178. Hussain, M.; He, L.; Schupp, J.; Lyons, D.; Heinemann, P. Green Fruit Segmentation and Orientation Estimation for Robotic Green Fruit Thinning of Apples. Comput. Electron. Agric. 2023, 207, 107734. [Google Scholar] [CrossRef]
  179. Hoppe, H.; De Rose, T.; Duchamp, T.; McDonald, J.; Stuetzle, W. Surface Reconstruction from Unorganized Points. SIGGRAPH Comput. Graph. 1992, 26, 71–78. [Google Scholar] [CrossRef]
  180. Harandi, N.; Vandenberghe, B.; Vankerschaver, J.; Depuydt, S.; Van Messem, A. How to Make Sense of 3D Representations for Plant Phenotyping: A Compendium of Processing and Analysis Techniques. Plant Methods 2023, 19, 60. [Google Scholar] [CrossRef] [PubMed]
  181. Um, D. Multiple Intensity Differentiation Based 3D Surface Reconstruction with Photometric Stereo Compensation. IEEE Sens. J. 2014, 14, 1453–1458. [Google Scholar] [CrossRef]
  182. Beltran, D.; Basañez, L. A Comparison Between Active and Passive 3D Vision Sensors: BumblebeeXB3 and Microsoft Kinect. In Proceedings of the ROBOT2013: First Iberian Robotics Conference, Madrid, Spain, 28–29 November 2014; Springer: Cham, Switzerland, 2014; Volume 252. [Google Scholar] [CrossRef]
  183. Panjvani, K.; Dinh, A.V.; Wahid, K.A. LiDARPheno: A Low-Cost LiDAR-Based 3D Scanning System for Leaf Morphological Trait Extraction. Front. Plant Sci. 2019, 10, 147. [Google Scholar] [CrossRef] [PubMed]
  184. Patel, K.; Park, E.-S.; Lee, H.; Priya, G.G.L.; Kim, H.; Joshi, R.; Arief, M.A.A.; Kim, M.S.; Baek, I.; Cho, B.-K. Deep Learning-Based Plant Organ Segmentation and Phenotyping of Sorghum Plants Using LiDAR Point Cloud. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 8492–8507. [Google Scholar] [CrossRef]
  185. Zhu, Y.; Sun, G.; Ding, G.; Zhou, J.; Wen, M.; Jin, S.; Zhao, Q.; Colmer, J.; Ding, Y.; Ober, E.S.; et al. Large-scale field phenotyping using backpack LiDAR and CropQuant-3D to measure structural variation in wheat. Plant Physiol. 2021, 187, 716–738. [Google Scholar] [CrossRef]
  186. Han, B.; Li, Y.; Bie, Z.; Peng, C.; Huang, Y.; Xu, S. MIX-NET: Deep Learning-Based Point Cloud Processing Method for Segmentation and Occlusion Leaf Restoration of Seedlings. Plants 2022, 11, 3342. [Google Scholar] [CrossRef]
  187. Syed, T.N.; Jizhan, L.; Xin, Z.; Shengyi, Z.; Yan, Y.; Mohamed, S.H.A.; Lakhiar, I.A. Seedling-Lump Integrated Non-Destructive Monitoring for Automatic Transplanting with Intel RealSense Depth Camera. Artif. Intell. Agric. 2019, 3, 18–32. [Google Scholar] [CrossRef]
  188. Eltner, A.; Sofia, G. Structure from Motion Photogrammetric Technique. In Developments in Earth Surface Processes; Elsevier: Amsterdam, The Netherlands, 2020; Volume 23, pp. 1–24. [Google Scholar] [CrossRef]
  189. Iglhaut, J.; Cabo, C.; Puliti, S.; Piermattei, L.; O’Connor, J.; Rosette, J. Structure from Motion Photogrammetry in Forestry: A Review. Curr. For. Rep. 2019, 5, 155–168. [Google Scholar] [CrossRef]
  190. Paulus, S.; Dupuis, J.; Mahlein, A.K.; Kuhlmann, H. Surface Feature-Based Classification of Plant Organs from 3D Laser-Scanned Point Clouds for Plant Phenotyping. BMC Bioinform. 2013, 14, 238. [Google Scholar] [CrossRef]
  191. Rossi, R.; Leolini, C.; Costafreda-Aumedes, S.; Leolini, L.; Bindi, M.; Zaldei, A.; Moriondo, M. Performances Evaluation of a Low-Cost Platform for High-Resolution Plant Phenotyping. Sensors 2020, 20, 3150. [Google Scholar] [CrossRef] [PubMed]
  192. Rossi, R.; Costafreda-Aumedes, S.; Leolini, L.; Bindi, M.; Moriondo, M. Implementation of an Algorithm for Automated Phenotyping through Plant 3D-Modeling: A Practical Application on the Early Detection of Water Stress. Comput. Electron. Agric. 2022, 197, 106937. [Google Scholar] [CrossRef]
  193. Gao, T.; Zhu, F.; Paul, P.; Sandhu, J.; Doku, H.A.; Sun, J.; Pan, Y.; Staswick, P.; Walia, H.; Yu, H. Novel 3D Imaging Systems for High-Throughput Phenotyping of Plants. Remote Sens. 2021, 13, 2113. [Google Scholar] [CrossRef]
  194. Tanabata, T.; Hayashi, A.; Kochi, N.; Isobe, S. Development of a Semi-Automatic 3D Modeling System for Phenotyping Morphological Traits in Plants. In Proceedings of the 2018 IECON—44th Annual Conference of the IEEE Industrial Electronics Society, Washington, DC, USA, 21–23 October 2018; pp. 5439–5444. [Google Scholar] [CrossRef]
  195. Pongpiyapaiboon, S.; Tanaka, H.; Hashiguchi, M.; Hashiguchi, T.; Hayashi, A.; Tanabata, T.; Akashi, R. Development of a Digital Phenotyping System Using 3D Model Reconstruction for Zoysiagrass. Plant Phenome J. 2023, 6, e20076. [Google Scholar] [CrossRef]
  196. Gomathi, N.; Rajathi, K.; Mahdal, M.; Elangovan, M. Point Sampling Net: Revolutionizing Instance Segmentation in Point Cloud Data. IEEE Access 2023, 11, 2731–2740. [Google Scholar] [CrossRef]
  197. Roggiolani, G.; Magistri, F.; Guadagnino, T.; Behley, J.; Stachniss, C. Unsupervised Pre-Training for 3D Leaf Instance Segmentation. IEEE Robot. Autom. Lett. 2023, 8, 2212–2219. [Google Scholar] [CrossRef]
  198. Li, D.; Shi, G.; Li, J.; Chen, Y.; Zhang, S.; Xiang, S.; Jin, S. PlantNet: A Dual-Function Point Cloud Segmentation Network for Multiple Plant Species. ISPRS J. Photogramm. Remote Sens. 2022, 184, 243–263. [Google Scholar] [CrossRef]
  199. Masuda, T. Leaf Area Estimation by Semantic Segmentation of Point Cloud of Tomato Plants. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Montreal, BC, Canada, 11–17 October 2021; pp. 1923–1931. [Google Scholar]
  200. Guo, R.; Xie, J.; Zhu, J.; Cheng, R.; Zhang, Y.; Zhang, X.; Gong, X.; Zhang, R.; Wang, H.; Meng, F. Improved 3D Point Cloud Segmentation for Accurate Phenotypic Analysis of Cabbage Plants Using Deep Learning and Clustering Algorithms. Comput. Electron. Agric. 2023, 211, 108014. [Google Scholar] [CrossRef]
  201. Wang, Y.; Liu, Q.; Yang, J.; Ren, G.; Wang, W.; Zhang, W.; Li, F. A Method for Tomato Plant Stem and Leaf Segmentation and Phenotypic Extraction Based on Skeleton Extraction and Supervoxel Clustering. Agronomy 2024, 14, 198. [Google Scholar] [CrossRef]
  202. Gu, W.; Wen, W.; Wu, S.; Zheng, C.; Lu, X.; Chang, W.; Xiao, P.; Guo, X. 3D Reconstruction of Wheat Plants by Integrating Point Cloud Data and Virtual Design Optimization. Agriculture 2024, 14, 391. [Google Scholar] [CrossRef]
  203. Imabuchi, T.; Kawabata, K. Semantic and Volumetric 3D Plant Structures Modeling Using Projected Image of 3D Point Cloud. In Proceedings of the 2024 IEEE/SICE International Symposium on System Integration (SII), Ha Long, Vietnam, 9–11 January 2024; pp. 141–146. [Google Scholar]
  204. Priestnall, G.; Jaafar, J.; Duncan, A. Extracting Urban Features from LiDAR Digital Surface Models. Comput. Environ. Urban Syst. 2000, 24, 65–78. [Google Scholar] [CrossRef]
  205. Zhao, L.; Um, D.; Nowka, K.; Landivar-Scott, J.L.; Landivar, J.; Bhandari, M. Cotton Yield Prediction Utilizing Unmanned Aerial Vehicles (UAV) and Bayesian Neural Networks. Comput. Electron. Agric. 2024, 226, 109415. [Google Scholar] [CrossRef]
  206. Available online: https://alexelvisbadillo.weebly.com/gvals-resources.html (accessed on 2 May 2024).
  207. Schunck, D.; Magistri, F.; Rosu, R.A.; Cornelißen, A.; Chebrolu, N.; Paulus, S.; Léon, J.; Behnke, S.; Stachniss, C.; Kuhlmann, H.; et al. Pheno4D: A Spatio-Temporal Dataset of Maize and Tomato Plant Point Clouds for Phenotyping and Advanced Plant Analysis. PLoS ONE 2021, 16, e0256340. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Trend for publications on feature extraction of plants using remote sensing.
Figure 1. Trend for publications on feature extraction of plants using remote sensing.
Remotesensing 16 04370 g001
Figure 2. Overview of remote sensing techniques applied to canopy-, plant-, and organ-level phenotyping.
Figure 2. Overview of remote sensing techniques applied to canopy-, plant-, and organ-level phenotyping.
Remotesensing 16 04370 g002
Figure 3. Support Vector Machine (SVM) segmentation pipeline.
Figure 3. Support Vector Machine (SVM) segmentation pipeline.
Remotesensing 16 04370 g003
Figure 4. Comparison between PointNet and SVM segmentation.
Figure 4. Comparison between PointNet and SVM segmentation.
Remotesensing 16 04370 g004aRemotesensing 16 04370 g004b
Figure 5. Comparison of PointNet vs. SVM performance.
Figure 5. Comparison of PointNet vs. SVM performance.
Remotesensing 16 04370 g005
Figure 6. PointNet loss and accuracy plots for training and validation.
Figure 6. PointNet loss and accuracy plots for training and validation.
Remotesensing 16 04370 g006
Figure 7. Unmanned aerial vehicle data processing [205].
Figure 7. Unmanned aerial vehicle data processing [205].
Remotesensing 16 04370 g007
Figure 8. Three-dimensional point clouds by remote sensing.
Figure 8. Three-dimensional point clouds by remote sensing.
Remotesensing 16 04370 g008
Figure 9. Three-dimensional data collection system for tomato plants.
Figure 9. Three-dimensional data collection system for tomato plants.
Remotesensing 16 04370 g009
Figure 10. (a) Zoomed image with annotation. (b) F1–confidence curve.
Figure 10. (a) Zoomed image with annotation. (b) F1–confidence curve.
Remotesensing 16 04370 g010
Figure 11. Data distribution before preprocessing.
Figure 11. Data distribution before preprocessing.
Remotesensing 16 04370 g011
Figure 12. Data distribution after preprocessing.
Figure 12. Data distribution after preprocessing.
Remotesensing 16 04370 g012
Figure 13. Leaf and stem classification and segmentation results.
Figure 13. Leaf and stem classification and segmentation results.
Remotesensing 16 04370 g013
Figure 14. Training loss and training accuracy.
Figure 14. Training loss and training accuracy.
Remotesensing 16 04370 g014
Table 1. Techniques used for plant phenotyping in remote sensing: contributions and limitations.
Table 1. Techniques used for plant phenotyping in remote sensing: contributions and limitations.
YearAuthorsContributionsLimitations
2020[1]Comprehensive review of remote sensing applications in precision agriculture; highlights the integration of satellite imagery and UAVs for optimizing inputs like water and nutrients.Focuses more on existing technologies rather than proposing new methodologies; limited discussion on future innovations.
2024[2]Overview of advancements in high-throughput phenotyping platforms using remote sensing; emphasizes the integration of diverse sensors like hyperspectral and 3D imaging.Limited by regional infrastructure availability, particularly in developing regions.
2014 [3]Reviews the state-of-the-art UAV remote sensing technologies for field-based crop phenotyping, emphasizing the ability to measure a wide range of phenotypic traits.Discusses existing technologies, lacks an exploration of emerging trends and future applications.
2024[4]Explores the transformative impacts of smart sensors, IoT, and AI in modern agricultural practices; discusses the integration of these technologies with remote sensing for enhanced precision.Primarily theoretical; requires further empirical validation and field tests.
2024[12]Compares UAV, satellite, and ground-based remote sensing approaches for field-based crop phenotyping; finds UAVs superior in resolution and efficiency for large-scale breeding programs.Limited to specific physiological traits like the canopy temperature and NDVI, does not cover broader phenotypic traits.
2024 [13]The first study using UAV-based remote sensing to model sustainability traits in crops like switchgrass; highlights UAVs’ potential in high-throughput phenotyping.Focused on a specific crop (switchgrass), which may limit generalizability to other crop types.
Table 2. Phenotyping feature extraction in different plant levels.
Table 2. Phenotyping feature extraction in different plant levels.
CriteriaCanopy-Level PhenotypingPlant-Level PhenotypingPlant-Organ-Level Phenotyping
Imaging TechniquesMultispectral
Hyperspectral
LiDAR
Thermal
SAR
Photogrammetry
High-Resolution RGB
Multispectral
Hyperspectral
LiDAR
Thermal
3D Imaging
Fluorescence
Hyperspectral
Data ResolutionMedium to low (depends on the altitude and sensor, satellite and UAV systems typically have lower resolution at higher altitudes)High due to proximity (due to closer proximity in UAV and controlled environment setups (CESs))Very nigh (detailed 3D reconstruction,
organ-specific data acquisition)
Phenotyping FocusWhole canopy cover, vegetation health, biomass estimationIndividual plant shapes, growth rates, nutrient stressLeaf area, stem diameter,
fruit/flower detection,
photosynthetic activity
Common Sensors UsedRGB cameras, multispectral/hyperspectral sensors, LiDAR, thermalRGB cameras, LiDAR, thermal, hyperspectral, multispectralFluorescence sensors, 3D scanners, high-resolution RGB
Feature ExtractionNDVI, EVI,
3D canopy structure from LiDAR,
thermal stress indices
Detailed 3D plant models, growth tracking,
nutrient stress detection
Leaf area, stem diameter, fruit/flower detection, photosynthetic activity (fluorescence)
Data Analysis TechniquesSpectral analysis,
image segmentation,
3D modeling,
time series analysis
Advanced segmentation, 3D reconstruction, thermal and spectral analysisDeep learning for organ-level segmentation, photosynthetic efficiency analysis
Data Collection FrequencyPeriodic (depending on the satellite or UAV revisit time)Continuous or high frequency in CES and UAV settingsOn-demand or continuous for high-resolution monitoring
AdvantagesLarge area coverage, whole crop monitoring, temporal analysisDetailed individual plant monitoring,
flexible for research needs
Highly precise phenotyping of organs (leaf, stem, and fruit), detailed trait extraction
Table 3. Features for canopy level phenotyping.
Table 3. Features for canopy level phenotyping.
Feature CategoryFeature NameDescriptionUse in Phenotyping and Crop ModelingAuthors
Spectral FeaturesNormalized Difference Vegetation Index (NDVI)Measures the photosynthetic capacity and vegetation health by comparing near-infrared (NIR) and red-light reflectance.Estimates biomass, crop vigor, and health status. Used in yield prediction and growth monitoring.[24,25]
Green NDVI (GNDVI)Similar to NDVI, focuses on green light, and assesses the chlorophyll content and leaf area index.Assesses the chlorophyll concentration, which is an indicator of the nitrogen content and crop stress.[26,27]
Enhanced Vegetation Index (EVI)Corrects for atmospheric and soil background noise, useful in areas with a high biomass.Enhances the signal in dense vegetation for better biomass and health estimations.[28]
Soil-Adjusted Vegetation Index (SAVI)Adjusts for soil brightness in areas with sparse vegetation.Improves the vegetation signal in areas where bare soil is exposed.[29]
Red Edge NDVIUses the red-edge band to detect subtle changes in the chlorophyll content and high biomass conditions.Early detection of plant stress and monitoring of senescence stages. Strongly correlated with the plant biomass due to its sensitivity to chlorophyll variations.[30]
Photochemical Reflectance Index (PRI)Measures light use efficiency (LUE) and CO2 uptake using reflectance changes at 531 nm and 570 nm due to xanthophyll cycle activity.Estimates gross primary productivity (GPP) and photosynthetic performance. Useful for detecting plant stress and quantifying CO2 fixation under varying environmental conditions.[31]
Visible Atmospherically Resistant Index (VARI)Assesses the physiological status of vegetation using the green, red, and blue bands.Detects plant stress and is significantly correlated with the grain yield, flowering time, plant height, and anthocyanin concentration. Shows strong utility in maize phenotyping.[32]
Normalized Difference Water Index (NDWI)Measures the water content in vegetation.Detects drought stress and water retention in crops.[33]
Normalized Green–Red Difference Index (NGRDI)Compares green and red bands to highlight vegetation health.Effective for monitoring crop health and early stress detection.[34]
Enhanced NDVI index (ENDVI)Combines the NIR and green bands to enhance vegetation signals.Useful for precision agriculture and assessing the nitrogen status.[35]
Green Ratio Vegetation Index (GRVI)Ratio of NIR and green band reflectance to assess crop health.Indicates chlorophyll activity and the nitrogen status.[36]
Chlorophyll Index Red
Edge (CIRE)
Uses the red edge band to estimate the chlorophyll concentration.Measures subtle changes in vegetation health and productivity.[37]
Log RedLogarithmic transformation of red reflectance to reduce saturation effects.Improves sensitivity to changes in vegetation cover.[27]
Red, Green Vegetation
Index (RG)
Weighted combination of red and green bands to highlight vegetation features.Identifies areas of crop stress and biomass estimation.[27]
Textural FeaturesGray-Level Co-occurrence Matrix (GLCM) FeaturesCaptures texture via contrast, correlation, energy, and homogeneity.Useful for detecting crop type variations and stress levels in vegetation.[38]
EntropyQuantitative measure of randomness in texture. Higher values indicate diverse texture elements.Differentiates healthy from stressed crops based on the uniformity of the canopy structure.[39,40]
Local Binary Patterns (LBPs)Describes spatial texture patterns in the canopy.Useful in differentiating crop species and identifying stress patterns.[41]
Structural FeaturesCanopy Height Model (CHM)Derived from DSM and DTM, represents plant height.Biomass estimation, growth rate monitoring, and yield potential prediction.[42]
Canopy CoverMeasures the proportion of ground covered by the canopy using thresholding techniques.Essential for yield prediction, irrigation management, and monitoring crop density.[43]
Canopy VolumeA 3D representation of the canopy structure.Biomass estimation and modeling crop architecture.[44]
Leaf Area Index (LAI)Measures the leaf area relative to the ground area.Estimates the photosynthetic capacity, crop growth, and potential yield.[27]
Temporal FeaturesGrowth RateMeasures changes in the canopy height, volume, and spectral indices over time.Useful for assessing crop growth, phenological changes, and stress responses.[45]
Phenological MetricsKey transitions like greening, flowering, and senescence stages. Combines structural and spectral features to estimate the above-ground biomass.Tracking phenological development and predicting yield. Critical for yield prediction and crop health monitoring.[46]
Environmental FeaturesEvapotranspiration (ET)Measures evaporation and plant transpiration, derived from UAV and spectral data.Used in irrigation management and water balance models.[47]
Soil Moisture ContentDerived from NDWI and thermal imagery.Estimates the soil water content and crop water stress.[48]
Machine Learning FeaturesDeep Learning-Based FeaturesExtracts complex patterns from high-dimensional datasets. Uses CNNs, graph-based models and transformers to learn patterns and generate high-resolution crop imagery.Improves feature extraction and phenotyping accuracy, and simulates realistic crop growth patterns.[49,50]
GAN-based FeaturesTrains GANs to generate synthetic images mimicking real-world growth patterns.Creates realistic synthetic images for simulations or data augmentation.[51]
Ordinary Differential Equations (ODEs) and Partial Differential Equation Methods (PDEs)Models the dynamics and temporal changes of complex systems. Useful for understanding and predicting physiological interactions in crops.Facilitates the study of dynamic growth patterns, water use efficiency, and stress responses under varying environmental conditions.[52]
3D Reconstruction Point Cloud Density and DistributionAnalyzing 3D point clouds for canopy structure and biomass estimation.Provides detailed structural information for accurate phenotyping and yield estimation.[53]
Table 4. Organ-level and plant-level feature extraction methods.
Table 4. Organ-level and plant-level feature extraction methods.
Methods/TechniquesDescriptionAuthors
Hyperspectral ImagingImaging technique capturing spectral information across wavelengths for a detailed analysis.[56,57,58]
NIR (Near-Infrared) ImagingUses near-infrared light to penetrate plant tissues, revealing internal structures.[59]
Fluorescence ImagingExploits natural fluorescence from chlorophyll to assess plant health and segmentation.[60,61]
Terrestrial Laser Scanning/Laser TechniquesUtilizes laser scanning to create detailed 3D models of plants for precise measurements, and includes Terrestrial Laser Scanning (TLS).[62]
Microscopic ApproachInvolves high-magnification imaging for detailed cellular and tissue analyses. Specialized microscopy focusing on stomata for physiological studies.[63]
Point Cloud3D representations of plants and their organs are obtained via scanning techniques like structured light, laser scanning, photogrammetry, and multiview stereo. Point cloud data are crucial for capturing precise structural information of plants and their individual organs. Extraction techniques focus on various methods such as shape modeling, statistical feature descriptors, and segmentation algorithms for analyzing morphological traits and growth patterns at the organ and plant levels.[64,65,66,67,68,69,70,71,72,73]
For Controlled Environments/ChambersEnclosed environments designed to maintain stable, regulated conditions for consistent phenotyping and experimentation. Used to control variables such as temperature, humidity, light intensity, and irrigation schedules, ensuring precise environmental monitoring and data consistency.[74,75]
Digital Image ProcessingApplication of image processing techniques to enhance, segment, and classify digital images for plant feature extraction, disease detection, and growth rate analyses. Techniques often include filtering, noise removal, segmentation, and classification for plant health assessments and monitoring.[76,77,78,79,80]
Morphological Feature ExtractionExtracts structural features like the shape, size, leaf edges, and structures of plant organs using morphological transformations, edge feature analysis, and key point detection methods. Applied for plant disease detection, species classification, and structural phenotyping. It can also be used to determine the ripeness or maturity level of vegetables.[81,82,83,84]
Zoning Feature ExtractionDivides images into zones to extract localized features.[85]
WaveletsAnalyzes images at multiple resolutions using wavelet transformations to extract features like edges, textures, and shapes. Applied for plant disease detection, species classification, and leaf identification by capturing fine details and patterns in plant images.[86,87,88]
Moment InvariantMathematical descriptors that are invariant to image transformations like scaling and rotation.[89]
Zernike Movement Invariant (ZMI)Uses Zernike polynomials for rotation-invariant feature extraction.[90]
Textural Feature AnalysisAnalyzes the texture of images to extract features like smoothness or coarseness.[91,92,93]
GLCM FeaturesGray-Level Co-occurrence Matrix (GLCM) for a texture analysis. This is widely used for texture classification and segmentation in medical imaging, agriculture, and other fields requiring fine-grained texture analysis by computing contrast, dissimilarity, homogeneity, energy, and correlation of the matrix.[94,95]
Histogram-Oriented Gradient MethodCaptures edge or gradient structures in images for feature extraction.[71,96]
Color Feature ExtractionExtracts features based on color information in images.[97]
Veins/Leaf Venation DensityAnalyzes the venation patterns in leaves for species identification or health assessments.[98,99]
Shape Recognition—HSV and OTSU MethodsUses the Hue, Saturation, Value (HSV) color space and Otsu’s thresholding for shape-based segmentation.[100]
CNN (Convolutional Neural Network)Convolution layers capture local spatial features by applying different filters. These features are then flattened into a feature vector, excluding fully connected layers, and the VGG16 architecture is used as a feature extractor to capture visual patterns from the images.[101]
The model first extracts meaningful features from plant images using ResNet50, processes these features through multiple Dense layers, and then sums the resulting values to predict the leaf count, making it a regression-based leaf counter.[102]
LSTM (Long Short-Term Memory)LSTM is used to process time series data, making it suitable for capturing dynamic changes over time. LSTM helps detect subtle early stress indicators that may not be evident in a single image, thereby improving early-stage stress detection.[101]
Attention MechanismAttention mechanisms are designed to selectively emphasize the most relevant features of an object while suppressing irrelevant or noisy features. Neural networks that focus on important parts of input data improve performance on tasks like classification.[103]
TransformersAdvanced deep learning models using self-attention mechanisms, effective in sequence modeling.[104,105]
Genetic ProgrammingEvolutionary algorithm that evolves programs or models to perform specific tasks.[106,107]
Particle Swarm OptimizationThe Particle Swarm Optimization (PSO) algorithm enhances the accuracy and efficiency of leaf disease classification by optimizing the segmentation process. PSO aids in finding optimal cluster centers, ensuring that the best characteristics of the features are applied for separation of diseased regions.[108,109]
Rider Cuckoo Search AlgorithmThe Rider Cuckoo Search Algorithm is a hybrid optimization technique integrating the Rider Optimization Algorithm (ROA) and Cuckoo Search (CS) to improve performance. It optimizes the training of Deep Belief Networks (DBNs) by balancing exploration and exploitation, leading to better classification results.[110]
Clustering and SegmentationGroups data points based on similarity, used in the general process of partitioning an image into meaningful regions like segmentation and classification.[111]
Discriminant AnalysisStatistical method used to find a combination of features that separates classes of objects.[112]
Thresholding and Edge-Based TechniquesThese techniques use intensity thresholds and edge detection for segmenting images into meaningful regions. Thresholding converts gray-scale images into binary images based on defined threshold values, while edge detection identifies boundaries between objects using first-order (e.g., Sobel and Canny) and second-order derivatives. These methods are widely used for medicinal plant image segmentation.[113]
Improved Saliency MethodEnhances the saliency (prominence) of objects for better segmentation by fusing the Sharif saliency-based (SHSB) method with active contour segmentation, improving the clarity of infected regions in cucumber leaves. This step aids in accurate feature extraction.[114]
Thermal imaging from flower temperature patternsCaptures the floral surface temperature distribution using infrared imaging, revealing contrasting temperature patterns. Analyzes distinct thermal structures within a flower to identify temperature contrasts between petals and reproductive parts.[115]
Double Line ClusteringA clustering method that segments diseased plant regions by analyzing pairs of lines for more precise image segmentation. It is used to identify and segment diseased leaf areas in crops like tomatoes, grapes, and cucumbers.[116]
Hough TransformIdentifies geometric shapes (such as lines, circles, and ellipses) within digital images by mapping the image space into a parameter space. Uses a voting mechanism to extract shape details based on defined parameters, making it highly resilient to noise and incomplete boundaries. This method excels at detecting intricate shapes and patterns in noisy datasets and handles partial or broken edges effectively, ensuring reliable extraction of irregular object shapes. Ideal for shape analyses.[117]
Adaptive Gamma CorrectionAdjusts the image brightness adaptively to enhance features.[114,117]
Multi-Fractal AnalysisAnalyzes complex patterns that exhibit fractal properties at multiple scales using fractal dimensions to capture both the local and global characteristics of an object. The key steps include calculating singularities for each image pixel using Hölder exponents and extracting multifractal spectra such as the Hausdorff dimension for global characterization. This method is particularly effective at identifying small objects, such as insect pests, under challenging conditions like variable lighting. It involves box-counting for estimating the fractal dimension and uses techniques such as regional minima and morphological operations to isolate target regions.[118]
Graph-Based Methods like Pyramidal Histogram of Oriented Gradients (PHOGs)Graph Cut segmentation is used to segment flowers from complex backgrounds by treating the image as a graph. PHOGs capture the shape of flowers using histograms of edge orientations at multiple pyramid levels, effectively representing local and global shape features for image matching.[119]
Piecewise Evolutionary SegmentationThe method divides time series data into smaller segments using evolutionary algorithms to find the optimal segmentation pattern, reducing dimensionality and retaining important features. This segmentation adapts dynamically using genetic algorithms to enhance classification and regression models by finding the best segmentation pattern for each problem.[120]
Table 5. Conventional 3D point cloud techniques.
Table 5. Conventional 3D point cloud techniques.
TechniqueContributionLimitationAuthors
Curvature-basedEstimates curvature values at each point for shape analysis and feature extractionCan be sensitive to noise and high curvature variations[121]
Normal basedEstimates surface normals at each point for shape analysis and feature extractionSensitive to noisy data and missing points[107]
Spin ImagesEncodes the spatial distribution of points using 2D histograms for feature description and object recognitionSensitive to occlusions and requires precise point alignment[127]
Voxel-basedDiscretizes the 3D space into voxels for efficient processingMay lose accuracy due to discretization[122]
PFHEncodes the spatial distribution of points in a local neighborhood for feature description and registrationComputationally expensive and sensitive to noise[124,125]
FPFHAn improvement over the PFH with reduced computational complexity while maintaining accurate feature descriptionLimited to small neighborhoods for fast computation[126]
Shape ContextUses 2D histograms to encode the spatial distribution, successful in cluttered environmentsMay require more computational resources for complex point clouds[123]
Table 6. Conventional deep learning approaches.
Table 6. Conventional deep learning approaches.
TechniqueContributionsLimitationsAuthors
Spectral CNNsSpectral analysis for shape recognition on meshes. Robust to non-isometric deformations.Restricted to manifold meshes like organic objects. Not easily extendable to non-isometric shapes like furniture.[130]
Feature-based DNNs Converts 3D data into vectors for classification. Fast and efficient processing.Limited by the representation power of extracted features. Requires domain-specific feature engineering.[131]
Volumetric CNNsThree-dimensional convolutional neural networks for shape recognition. Real-time object recognition.Constrained by resolution due to data sparsity and computation costs. Limited to small-scale 3D data.[128]
Multiview CNNsTwo-dimensional convolutional neural networks for shape classification and retrieval. Efficient processing of large-scale 3D data.Limited to shape classification and retrieval tasks. Requires multiple views of the 3D object.[129]
Table 7. Contributions and limitations of advanced deep learning approaches.
Table 7. Contributions and limitations of advanced deep learning approaches.
TechniqueContributionsLimitationsReference
PointNetNovel approach for point-wise feature extraction for 3D point cloud segmentationLacks local context information, making it less effective for complex plant structures and dense datasets[132]
PointNet++Extended PointNet by incorporating local neighborhood features through hierarchical groupingStruggles with highly complex inter-point dependencies in large-scale datasets[133]
Point–voxel CNNImproves execution efficiency by pooling the advantages of voxels and pointsMay need additional fine-tuning for optimal performance[141]
PointASNLUses nonlocal neural networks for robust point cloud processingCan be computationally intensive and requires extensive memory[142]
PAConvUtilizes a dynamic kernel construction to adapt convolution weights based on the local geometryHigh complexity and increased training times[135]
DGCNNUses a graph-based approach with dynamic EdgeConv operations for local neighborhood relationship modelingComputationally intensive and sensitive to noise in low-density regions[134]
CurveNetLeverages curve-based features to capture connectivity and spatial context in curved plant structuresVulnerable to noise, reducing its effectiveness in real-world settings[136]
FatNetFeature-attentive network for 3D point cloud processingCan be sensitive to overfitting due to its high focus on features[143]
POEMReduces storage and computing costs with a 1-bit fully connected layer (Bi-FC)Potential loss of detail due to compression[144]
LatticeNetNovel approach for 3D semantic segmentation from raw point cloudsMay require significant computational resources for segmentation[145]
Point TransformerIntroduces self-attention for long-range feature aggregation, excelling in complex organ segmentationThe high computational cost limits real-time applications[137]
G-PCC++A KNN-based linear interpolation for geometry restoration, a KNN-based Gaussian distance weighted mapping for attribute enhancement.Computational complexity, which might limit resource-constrained environments[146]
Stratified TransformerExtends Mask R-CNN to 3D data using sparse convolutions, enabling robust instance segmentation of large-scale datasetsRelies on pre-computed features, increasing the preprocessing time and reducing flexibility[138]
Mask3DExtends Mask R-CNN to 3D data using sparse convolutions, enabling robust instance segmentation of large-scale datasetsRelies on pre-computed features, increasing the preprocessing time and reducing flexibility[139]
FNeVREnhances facial details for image rendering via neural volume renderingSpecific to facial applications and might not be generalizable[147]
SP-LSCnetCombines unsupervised clustering and an adaptive network for efficient segmentation using superpointsTwo-stage processing may lead to misclassification in geometrically ambiguous regions[140]
Table 8. Comparison of the processing time, frame rate, data size, and storage requirements for 3D point cloud methods.
Table 8. Comparison of the processing time, frame rate, data size, and storage requirements for 3D point cloud methods.
   MethodProcessing Time (ms)FrameRate (FPS)DataSize (MB)StorageRequirements (MB/min)
Conventional (before 2010)100–5001–5100–5001000–5000
PointNet (2017)10–3030–10010–50100–500
PointNet++ (2018)5–1560–2005–2050–200
Recent Neural Network-Based Methods (2020 and later)1–5200–5001–1010–50
Table 9. Comparison of deep learning (PointNet) and machine learning (SVM) models’ performance across class labels.
Table 9. Comparison of deep learning (PointNet) and machine learning (SVM) models’ performance across class labels.
  MetricBuildingsGroundVegetationWaterUnclassifiedOverall Accuracy
Precision (PointNet)PoorGoodGoodPoorPoor78%
Precision (SVM)PoorHighHighPoorPoor88%
Recall (PointNet)Very lowHighModerateVery low--
Recall (SVM)ModerateHighHigh---
F1 score (PointNet)WeaknessModerateModerate Weakness--
F1 score (SVM)WeaknessHighHigh---
Table 10. Contributions and limitations of 2D image-based feature extraction/segmentation methods.
Table 10. Contributions and limitations of 2D image-based feature extraction/segmentation methods.
  Technique   Contributions   LimitationsAuthors
Image segmentationDivides the image into background and foreground region of interests [158]
Edge detectionSobel edge detectionSlow process and results in a noisy image
Prewitt edge detection: faster processSuitable for a high-contrast and low-noise image
Canny edge detection: based on the Laplacian algorithm and is insensitive to noiseEncompasses a multistage algorithm, and hence results in longer processing time
Blob analysis (binary large object analysis)The position of the target can be detected.
Semantic SegmentationFour different deep learning segmentation methods; semi-automatic annotation processRequires large amount of annotated image datasets; exhaustive manual annotation process[164]
Color ConversionHSV color space: thresholding is utilized to determine ripe and turning fruit colors. Hue values for ripe tomatoes are 0–10, the saturation value is 170–256, and the hue value for turning fruit is 11–20 and the saturation value is 150–256.When ripe and turning fruits are grouped together, the technique can identify the fruit but cannot distinguish between both fruits. Hence, its counts the fruits as a single fruit than a couple of fruits (multiple fruits).

Focal bottleneck transformer network
(FBoT-Net) for green apple
[154]





[178]
Image SegmentationSegments the image based on HSV values
Deep learning based 2D image segmentation
Color ConversionIdentifies ripe strawberries based on the hue (HSV format). [179]
Single-stage deep learning (YOLOv8)Improves the performance by combining ECA with YOLO. Results in enhanced identification of ripe strawberries in complex environments (forelight, backlight, and occlusion).Limitation in Context understanding and struggle in detecting small objects compared to latest vrsions.
Color AnalysisThe required target has different color values compared to other parts of the plants.Requires external memory for image processing[155]
SSDHigh processing speed and accuracyDifficulty detecting fruits at the edge of the frame and farther away from the frame[171]
K-means SegmentationFaster fruit detectionLarger images required longer running time. Occlusions lead to inaccurate results.[166]
RGB-D Image SegmentationReal-time fruit detection (138 frames per second)Slightly noisy results. The neural network structure is complicated and the training process is longer.[177]
Low-cost sensorGuava fruits were easier to detect than the branch. Some of the unsuccessful results were due to sunlight.[176]
Table 11. Three-dimensional (3D) point cloud-based plant feature extraction: contributions and limitations.
Table 11. Three-dimensional (3D) point cloud-based plant feature extraction: contributions and limitations.
TechniqueContributionsLimitationsAuthors
Point Sampling MethodImproves instance segmentation in point clouds for plant growth phenotyping in agricultureMay require large datasets and computational resources[196]
RANSAC ExtensionExtends RANSAC for complex plant morphologies, achieving high accuracy in measuring plant attributesMay be sensitive to imaging noise and cluttered backgrounds[5]
Self-Supervised Pre-Training Reduces the labeling effort for 3D leaf instance segmentation in agricultural plant phenotypingMay not work as well in highly occluded environments[197]
LatticeNetData augmentation to enhance plant part segmentation, contributing to the plant architecture reconstructionMay require significant computational resources for semantic segmentation[145]
PlantNetDual-function deep learning network for semantic and instance segmentation in multiple plant speciesMay require additional validation for diverse plant species[198]
Joint Plant Instance DetectionMethod for joint plant instance detection and leaf count estimations in agricultural fieldsMay have limited scalability to other types of plants or environments[7]
Leaf Area EstimationEstimates the leaf area in tomato plants using RGB-D sensors and point cloud segmentationRelative error of about 20%, indicating possible accuracy issues[199]
Robust RANSACRobust method for the direct analysis of plant point cloud data using RANSAC, achieving high accuracyMay not work well in highly noisy or cluttered environments [5]
Virtual Design-Based ReconstructionA 3D reconstruction technique for wheat plants, integrating point cloud data with optimization algorithmsChallenges in precise reconstruction due to the lack of organ templates[202]
ASAP-PointNetMethod for analyzing cabbage plant phenotypes using 3D point cloud segmentationRequires complex preprocessing and training[200]
Semantic and Volumetric 3D ModelingMethod for semantic and volumetric 3D modeling from point cloud data for accurate radiation dose distributionsCan be computationally expensive and requires significant memory[203]
Tomato Plant SegmentationMethod for segmenting tomato plant stems and leaves, extracting phenotypic parametersChallenges with canopy adhesion and under-segmentation[201]
Sweet Pepper Leaf Area EstimationMethod to estimate the sweet pepper leaf area using semantic 3D point clouds generated from RGB-D imagesIssues in fully capturing tall plants and varying point cloud resolutions[10]
Table 12. Contributions and limitations of the 2D image-based plant feature extraction process.
Table 12. Contributions and limitations of the 2D image-based plant feature extraction process.
AspectContributionLimitation
Annotation ProcessEfficient segmentation of plant parts using a annotation tool (stems, branches, and suckers) in LabelStudioTime-consuming annotation process, approximately 2–4 min per image
Model TrainingSuccessful training of a segmentation model with the available datasetLimited using full plant images, resulting in an F1 confidence score of 0.65
Future DirectionsPlan to refine training by focusing on specific plant organs (branches, stems, and suckers)The current dataset and training approach are insufficient for higher accuracy
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nethala, P.; Um, D.; Vemula, N.; Montero, O.F.; Lee, K.; Bhandari, M. Techniques for Canopy to Organ Level Plant Feature Extraction via Remote and Proximal Sensing: A Survey and Experiments. Remote Sens. 2024, 16, 4370. https://doi.org/10.3390/rs16234370

AMA Style

Nethala P, Um D, Vemula N, Montero OF, Lee K, Bhandari M. Techniques for Canopy to Organ Level Plant Feature Extraction via Remote and Proximal Sensing: A Survey and Experiments. Remote Sensing. 2024; 16(23):4370. https://doi.org/10.3390/rs16234370

Chicago/Turabian Style

Nethala, Prasad, Dugan Um, Neha Vemula, Oscar Fernandez Montero, Kiju Lee, and Mahendra Bhandari. 2024. "Techniques for Canopy to Organ Level Plant Feature Extraction via Remote and Proximal Sensing: A Survey and Experiments" Remote Sensing 16, no. 23: 4370. https://doi.org/10.3390/rs16234370

APA Style

Nethala, P., Um, D., Vemula, N., Montero, O. F., Lee, K., & Bhandari, M. (2024). Techniques for Canopy to Organ Level Plant Feature Extraction via Remote and Proximal Sensing: A Survey and Experiments. Remote Sensing, 16(23), 4370. https://doi.org/10.3390/rs16234370

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop