Next Article in Journal
Comprehensive Analysis and Evaluation of Anomalous User Activity in Web Server Logs
Next Article in Special Issue
Playing Flappy Bird Based on Motion Recognition Using a Transformer Model and LIDAR Sensor
Previous Article in Journal
IMU-Based Fitness Activity Recognition Using CNNs for Time Series Classification
Previous Article in Special Issue
Enhancing Smart Building Surveillance Systems in Thin Walls: An Efficient Barrier Design
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Development of a Stereo Vision System to Study the Nutation Movement of Climbing Plants

by
Diego Rubén Ruiz-Melero
1,
Aditya Ponkshe
2,
Paco Calvo
2,3 and
Ginés García-Mateos
1,*
1
Computer Science and Systems Department, University of Murcia, 30100 Murcia, Spain
2
Minimal Intelligence Laboratory (MINT Lab), University of Murcia, 30100 Murcia, Spain
3
Department of Philosophy, University of Murcia, 30100 Murcia, Spain
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(3), 747; https://doi.org/10.3390/s24030747
Submission received: 6 December 2023 / Revised: 21 January 2024 / Accepted: 22 January 2024 / Published: 24 January 2024
(This article belongs to the Special Issue Robust Motion Recognition Based on Sensor Technology)

Abstract

:
Climbing plants, such as common beans (Phaseolus vulgaris L.), exhibit complex motion patterns that have long captivated researchers. In this study, we introduce a stereo vision machine system for the in-depth analysis of the movement of climbing plants, using image processing and computer vision. Our approach involves two synchronized cameras, one lateral to the plant and the other overhead, enabling the simultaneous 2D position tracking of the plant tip. These data are then leveraged to reconstruct the 3D position of the tip. Furthermore, we investigate the impact of external factors, particularly the presence of support structures, on plant movement dynamics. The proposed method is able to extract the position of the tip in 86–98% of cases, achieving an average reprojection error below 4 px, which means an approximate error in the 3D localization of about 0.5 cm. Our method makes it possible to analyze how the plant nutation responds to its environment, offering insights into the interplay between climbing plants and their surroundings.

1. Introduction

The nutational movement of climbing plants refers to the rhythmic, circular, or nodding motion exhibited by certain plant parts, such as stems, tendrils, or growing tips, as they explore and interact with their environment during the process of climbing [1]. This movement is often associated with the search for support structures, such as poles, trellises, or other plants, which the climbing plant can use for stability and upward growth [2]. It occurs as a result of differential growth rates on different sides of the plant organ. As the plant grows, the cells on one side of the organ elongate more rapidly than those on the opposite side, causing the organ to bend or curve [3]. This bending or curving allows the plant to explore its surroundings and find a suitable support for climbing.
Climbing plants, such as common beans (Phaseolus vulgaris L.), employ various mechanisms for climbing, including twining, where the plant winds around a support, and the use of tendrils, specialized structures that coil around objects for support [4].
The study of nutation movements in climbing plants is part of the broader field of plant tropisms, which involves the growth responses of plants to external stimuli such as light, gravity, and mechanical touch, first studied in the pioneering works of Charles Darwin [5]. This nutation movement allows climbing plants to optimize their growth in response to their environment, increasing their chances of successful climbing and their access to sunlight for photosynthesis. Moreover, it has been the inspiration for the development of new techniques and methods in robotics [6], computational intelligence [7,8], and other bioinspired innovations [9].
This current study of nutation and other plant movements requires the recording of time-lapse images of plants, due to the long time span in which the movements occur, and the use of artificial vision algorithms for the efficient analysis of the videos obtained. For example, Navarro et al. [10] developed a specific capture chamber to monitor the growth of plants at intervals from 30 s to hours or even days. The captured images were then analyzed with computer vision algorithms based on thresholding, morphological operators, and blobs detection. Using the parameters obtained from the blobs, features such as area, length, compactness, and perimeter were extracted for further analysis. Stolarz et al. [11] developed software called the “Circumnutation Tracker” to analyze the circular movement of the tip of the plants in zenithal time-lapse videos. The X and Y positions of the tip are manually selected by the user, and the system includes useful tools for the visualization and characterization of the plant movement, automatically extracting circumnutation parameters.
More recently, Tenzer and Clifford [12] proposed a technique to analyze the growing movement of hydroponic plants using different neural networks on grayscale images. U-net, Linknet, FPN, PSPNet, and a 34-layer ResNet architecture were used for comparison, obtaining a maximum area under the curve (AUC) for the validation set over 0.92. Another recent method based on deep learning was presented by Mao et al. [13], which proposed a U-net network architecture to segment the plant. This method is an improvement of their free software “Plant Tracer”, which analyzed plant movement using the classic computer vision techniques of object tracking and blob detection. Díaz-Galián et al. [14] presented a methodology to analyze the movement of plants using different reference points, which were obtained from image analysis. This methodology included coordinates transformation, curve fitting, and statistical analysis, allowing a comparison between different plant species. Plant movement has also been analyzed using other types of sensors. For example, Geldhof et al. [15] developed a digital inertial measurement unit (IMU) sensor to measure in real-time the movement of leaves, achieving a precision of pitch and roll angles under 0.5°. However, the exact 3D position of the leaves was not obtained.
The aim of this work is to develop a stereo system for the visual tracking of plant nutation movement, which enables obtaining the precise 3D location of the tip of the plant. This work was focused on the study of the common bean (Phaseolus vulgaris L.), since the ultimate purpose is to analyze the differences in the movement between using or not using a support pole for plants. Previous work from the MINT Lab research group suggests that the dynamic patterns of plant nutation are influenced by the presence of support to climb in their vicinity [2]. This fact has also been observed by other researchers. For example, Guerra et al. [16] analyzed the 3D kinematics of the movement of climbing plants, demonstrating that plants are not only able to perceive their environment but can also scale the movement of their tendrils depending on the size of the support.
The objective of this present study is not to analyze these kinematic changes under different environmental conditions but to develop computational tools for the accurate 3D measurement of plant tip movement that will be used in further studies. Consequently, for our analysis to be effective in a broader range of conditions, we obtained data on plants from different settings, i.e., with and without a support to climb onto. Since the proposed method is able to reconstruct the 3D position of the tip in both conditions, our method promises to offer a framework for future studies to analyze how plant nutation responds to different environments.

2. Materials and Methods

The proposed computer vision method to track the movement of the plant tip is shown in Figure 1. It consists of five main stages: calibration, data acquisition, plant tip detection in the top and side images, 3D position estimation, and visualization of the results.
In the following subsections, the characteristics of the experimental setup and the stereo imaging system are presented first. Then, a brief introduction to the mathematical foundations and artificial vision algorithms used is given. Finally, the procedure used to estimate the position of the plant tip at each of the moments captured in the time-lapse videos is described.

2.1. Experimental Setup

The experiments were carried out by the MINT Lab research group at the Scientific and Technical Research Area of the University of Murcia (Spain). These experiments were conducted under controlled conditions and consisted of two scenarios. In the first scenario, the climbing plant was placed near a pole, which served as a support for the plant to grip when it reached its position; in the second scenario, the plant was placed within a cabin with no pole in it. The pole was 90 cm in height and 1.8 cm in diameter and was placed at a distance of 30 cm from the plant center. Concerning the soil, a mixture of peat moss and perlite (70–30%) was used throughout all the experiments. Both cabins were equipped with white parabolic reflectors to provide symmetrical lighting. The temperature of the growth chamber was kept constant at 20 °C and the relative humidity at 85% ± 5%. We have to note that, although this is a high relative humidity, plants are able to accomplish their life cycles successfully with no harm, especially in controlled conditions in which heat stress can be prevented. A L16:D8 h photoperiod was provided via high-pressure sodium lamps (Lumatek pulse-start HPS Lamp 250 W; height: 150 cm and photon fluence rate: 430 ± 50 μmol m−2 s−1 at leaf level). During the 8 h of darkness, a dim phototropically inactive green safelight (fluence rate under 5 μmol m−2 s−1) was activated, which was enough to take pictures in the dark using a 4.2 μm high dynamic range (115 dB) image sensor.
Figure 2 shows an example of these two scenarios, with and without a supporting pole. Specifically, the plant species used for this experiment were the common bean (Phaseolus vulgaris L.). Seeds of the cultivar “Garrafal Oro Vega” variety were provided by Semillas Ramiro Arnedo S.A., Spain (https://www.ramiroarnedo.com). “Garrafal Oro Vega” is a variety with abundant foliage, having medium-sized leaves and containing 18–20 cm long pods. The plants went through multiple stages from germination to the start of the recording. The seeds took 48 h to germinate. Upon germination, the young seedlings were transferred to coconut fiber growing pellets and kept on propagation trays for further development. Once the first true leaves matured, which took around 8–12 days from the young seedling stage, healthy-looking, 25–30 cm tall plants were selected and transferred into the recording booths for video capture. The videos were recorded until the plant either grabbed the pole, or the tip came out of the experimental area (i.e., the area visible to the cameras), which usually happened after 3 or 4 days.
It can be observed in Figure 2 that a dark color was used at the bottom part of the booth. The purpose of this dark cover was to increase the contrast with the plant contour, thus enhancing the performance of the algorithm for extracting the position of the plant tip in the videos corresponding to the top view in time-lapse. We made sure that the white part remained disproportionately more than the black part. The proportion of the lower black part was calibrated so that the plants did not show any shade avoidance response, thereby altering the nutation movement.
Once the young seedlings were transferred to the coconut fiber growing pellets and kept on propagation trays for further development, 30 mL of water was initially added every day. Afterwards, when the plants started to mature, they were watered so that around 80% moisture was maintained throughout the growing stage. For recording purposes, the plants were transferred into big, black plastic pots. At the onset of the recording, 400 mL of water was added to each plant, which was found sufficient to keep the soil moist throughout the recording phase. No extra water was added during the four days of recording.
To estimate the nutational movement of the plant, a stereo pair of cameras was used, recording synchronously and in time-lapse. The target of the motion tracking was the plant tip. The configuration of the stereo vision system is shown in Figure 3. This imaging system consisted of two Brinno TLC200 Pro cameras (Brinno Ltd., Taipei, Taiwan) positioned at the top and side of the plant, forming a 90° angle. The technical specifications of the cameras are shown in Table 1. This top-side configuration of the cameras was chosen because it can offer high accuracy in detecting the tip of the plant, although it may not be the best choice for tracking other parts of the plant.
To synchronize the captured images, the recording was started simultaneously on both cameras. Each camera used its own internal clock. The images were captured at one-minute intervals for a continuous 24 h recording, lasting from 3 to 4 days. Because each camera had its own clock, there could be a slight offset in the images between the two cameras for some recordings. To determine the number of frames corresponding to this offset, the light–dark cycle used in the experiments to provide the L16:D8 photoperiod was used. For this purpose, the zenithal images were used as a reference, based on the photoperiod changes. During one of the lighting changes, the offset between the top and side view images was computed, measured as the number of frames of difference. This number was used in the further analysis of the sequence.

2.2. Mathematical Methods

In general, a camera maps 3D points in space to 2D points on the image plane. The cameras used in the experiments were modeled using a pinhole camera model. In a pinhole model, a ray from a point in space passes through a fixed point called the projection center. The intersection of this ray with a chosen plane in space, the image plane, is the projection of the point on the image plane [17]. The projective transformation given by a pinhole camera model is as follows:
s p = A R t P w
where:
  • s is an arbitrary scaling factor of the projective transformation.
  • p is the 2D pixel on the image plane.
  • P w is the 3D point expressed in the world coordinate system.
  • A is the matrix with the intrinsic parameters of the camera.
  • R and t are the rotation and translation, respectively, describing the coordinate change from the world to the camera coordinate system.
The matrix with the intrinsic parameters is composed of the focal lengths, f x and f y , expressed in pixel units, and the principal point of the camera, ( C x , C y ).
A = f x 0 C x 0 f y C y 0 0 1
This matrix with the intrinsic parameters does not depend on the scene, so it can be reused once estimated, as long as the focal length remains fixed. On the other hand, matrix [ R | t ] contains the extrinsic parameters of the camera. This matrix performs the homogeneous transformation and represents the change of basis from the world coordinate system to the camera coordinate system, as follows:
X c Y c Z c 1 = r 11 r 12 r 13 r 14 r 21 r 22 r 23 r 24 r 31 r 32 r 33 r 34 0 0 0 1 X w Y w Z w 1
Combining these two matrices yields the projective transformation, which maps the 3D points in the world to the 2D points on the image plane in normalized camera coordinates.
s u v 1 = f x 0 C x 0 f y C y 0 0 1   r 11 r 12 r 13 r 14 r 21 r 22 r 23 r 24 r 31 r 32 r 33 r 34 0 0 0 1 X w Y w Z w 1
The pinhole camera model does not consider the distortion caused by the lenses used in the cameras. To accurately represent a real camera, the camera model includes both radial and tangential lens distortion. Radial distortion occurs when light rays bend more at the edges of the lens than at the optical center. The radial distortion coefficients model this type of distortion as follows:
x d i s t o r t e d = x 1 + k 1 r 2 + k 2 r 4 + k 3 r 6 y d i s t o r t e d = y 1 + k 1 r 2 + k 2 r 4 + k 3 r 6
where:
  • x , y is the location of the pixel without distortion.
  • k 1   k 2 , and k 3 are the radial distortion coefficients of the lens.
  • r 2 = x 2 + y 2 .
On the other hand, tangential distortion occurs when the lenses and the image plane are not perfectly parallel. In this case, the tangential distortion coefficients model this type of distortion as follows:
x d i s t o r t e d = x + p 2 r 2 + 2 x 2 + 2 p 1 x y y d i s t o r t e d = y + p 1 r 2 + 2 y 2 + 2 p 2 x y
where:
  • x , y is the location of the pixel without distortion.
  • p 1 and p 2 are the tangential distortion coefficients of the camera.
  • r 2 = x 2 + y 2 .
Using this basis, the computer vision algorithms that have been used in the implementation of the pipeline for the method used to track the plant nutation movement are described in the following points.

2.2.1. Perspective-n-Point (PnP)

The pose computation problem consists of solving the rotation and translation of an object with respect to the camera, while minimizing the reprojection error for correspondences between the 3D points in the world and the 2D points on the image. Several methods have been proposed to estimate the pose of an object relative to the camera [18,19]. These methods can be used for calibration using a planar calibration object [20].

2.2.2. Calibration of a Stereo Pair System

The procedure outlined in the previous subsection enables the obtaining of the intrinsic parameters of the camera as well as the camera pose relative to the calibration object. If the calibration object is visible to both cameras, we have the pose ( R 1 , t 1 ) of the first camera relative to the calibration object and the pose ( R 2 , t 2 ) of the second camera. These two poses are related as follows:
R 2 = R × R 1   ;   t 2 = R × t 1 + t
Therefore, by using the previous equation, it is possible to obtain the pose of the first camera relative to the second camera. Once the pose of the first camera relative to the second one is obtained, the essential matrix can be constructed as follows:
E = 0 t 2 t 1 t 2 0 t 0 t 1 t 0 0 × R
where t i are the components of the translation vector t = [ t 0 , t 1 , t 2 ]. From this, the fundamental matrix can be computed as follows:
F = C a m e r a M a t r i x T × E × C a m e r a M a t r i x 1

2.2.3. Background Segmentation

The background segmentation problem involves detecting changes that occur in a sequence of images. Pixels in an image are classified either as part of the background or as part of an object of interest based on a background model. In real applications, depending on the characteristics of the problem, different methods have been proposed: models based on color thresholding, clustering, and fuzzy logic, and, more recently, neural network-based models [21] have been used. The usual steps in image segmentation are:
  • The background model is initialized with some of the video frames.
  • Once the background model is initialized, each pixel in the image is classified based on the background model.
  • The background model is updated with information from the last processed image.
In the algorithms developed to extract the position of the plant’s tip in time-lapse images, the Gaussian mixture method was used. This method uses multiple Gaussian functions per pixel to model the background of the scene. This method takes into account that a pixel can have different states over time. The implementation of this method in the OpenCV library is based on Zivkovic [22].

2.3. Motion Tracking and 3D Position Estimation

This section first describes the procedure used to calibrate the stereo system. Next, the steps in the process of extracting the position of the plant tip from the images captured by the stereo system cameras are outlined. Finally, it describes how the nutation motion of the plant was tracked based on the 2D position of the extracted plant tip from the images.

2.3.1. Camera Calibration

The stereo system calibration was carried out following the procedure described in the previous subsection. The calibration process was conducted in two phases. First, each of the stereo cameras was individually calibrated. A calibration object with a 10 × 7 chessboard pattern and dimensions of 59.4 × 84.1 cm was used for the calibration. These dimensions were chosen to cover a large part of the image plane, allowing the precise estimation of the distortion coefficients.
Once the individual calibration of the cameras was completed, the stereo system calibration was performed to estimate the projection matrices for each camera and the fundamental matrix. In this case, a calibration object with dimensions of 42.0 × 59.4 cm was used, visible in both cameras for a sufficient number of poses to carry out the stereo system calibration. Table 2 shows the results of the calibration, measured in terms of the reprojection error, that is, the distance from the calibration pattern key points detected in the images to a corresponding world point projected into the same images.
In Figure 4, the points used in the calibration of the stereo system can be observed, combining the positions of all the images used for calibration.

2.3.2. Detection of the Plant Tip in the Overhead View

The process consisted of two steps, as shown in Figure 5. First, for each frame, n points, where the plant tip could be located, were detected. Once the candidate points were obtained, a selection process for the position of the plant tip was performed for each frame, followed by interpolation for frames where no candidate points were detected. The steps of the first step can be seen in Figure 6.
The following are the highlights of the algorithm used in the first step of the process:
  • The segmentation of the plant contour was performed using the mixture of Gaussians algorithm.
  • After each change in lighting, the process reset the background model to reduce the stabilization time.
  • To detect when a lighting change occurred, the process checked the area of the selected contours, and, if the area exceeded a threshold specified by the lighting_change_area parameter, it was assumed that a lighting change had occurred.
  • After a lighting change, the process did not check again for a lighting change until after the frame period specified by the lighting_change_est_time parameter to prevent the repeated detection of a lighting change in every frame after the background model had been reset.
The process for selecting the plant tip position among the candidate points followed these two criteria:
  • If there was only one candidate point, it was selected.
  • If there was more than one candidate point, the point closest to the last point selected for previous frames was selected. In segments of frames where there was no candidate point, the position of the plant tip was estimated by performing linear interpolation using the plant tip position in the previous and subsequent frames of the segment.

2.3.3. Detection of the Plant Tip in the Lateral View

This process consisted of two stages, as in the case of the process used for extracting the position of the plant tip in the top view images. First, for each frame, the n points, where the plant tip could be located, were detected. Once the candidate points were obtained, a selection process for the plant tip position was executed for each frame, followed by interpolation for frames where no candidate points were detected. The steps of the first stage are depicted in Figure 7.
The highlights of this algorithm used in the first stage of the process are:
  • The segmentation of the plant contour was performed using the mixture of Gaussians algorithm, just as for the segmentation of the plant contour in the top view.
  • For detecting changes in lighting, the same procedure used in processing images from the top view was followed.
  • There were cases where multiple fragments of the plant contour are obtained when applying the background model to the image. In these cases, the use of the epipolar line helped to identify the fragment where the plant tip was located.
  • Once the contours near the epipolar line were selected, a minimum path algorithm was used to select the potential position of the plant tip within the contour.
The process for selecting the plant tip position among the candidate points followed the same criteria as indicated in the previous section.

2.3.4. Correcting Wrong Detections and Obtaining the 3D Positions of the Plant Tip

The algorithms described in the previous subsections were not always able to detect the plant tip position in all cases, as mentioned in each section. For this reason, a “Plant Tracker” tool was developed to allow the user to modify the automatically performed plant tip extractions. This tool consists of three screens, as shown in Figure 8. Once the user has completed the manual corrections, the application allows the extraction of a CSV file with the estimation of the 3D position for a selected range of frames.

3. Results and Discussion

In this section, the results obtained in the processing of the available videos are presented. The videos were processed in pairs, corresponding to the zenithal and side views of each same experiment. The nutation movement followed by the tip of the bean plant in three of the conducted experiments is shown in Figure 9.

3.1. Results of the Detection of the Tip of the Plant

Table 3 shows the cumulative results of the type of detection used to extract the position of the plant tip in the frames of the time-lapse videos corresponding to the experiments. The types of detection correspond to the following categories: Automatic: The algorithm for extracting the plant tip position found only one candidate point. Manual: The points that the user had to correct. Estimated: The algorithm for extracting the plant tip position found more than one candidate point and selected the one that represented the most probable position of the plant tip. Interpolated: The algorithm for extracting the plant tip position did not find a candidate point. The plant tip position in the frame was determined by interpolating the position from the plant position in the preceding and subsequent frames.
Figure 10 shows the reprojection error in pixels for each frame of the videos (side and top views) of the time-lapse within the range for which the 3D position estimation of the plant tip was performed.

3.2. Discussion of the Results

The plant tip detection algorithms were able to extract the correct position in a range of 86–98% of the cases, as observed in Table 3. The number of corrections made by the user was higher for the plant positions extracted in the lateral plant videos. This was mainly due to moments when the plant tip overlapped with the plant stem in the images. An example of this case is shown in Figure 11. This error is due to the use of a minimum path algorithm to find the plant tip in the segmented contour of the plant. When the plant tip overlapped with the stem, the position farthest from the plant’s base corresponded to the elevated part of the plant stem visible in the image.
Regarding the reprojection errors, the graphs shown in Figure 10 follow a sawtooth pattern across all processed videos, with the maximum reprojection errors aligned in both the top and side view videos of each time-lapse. Figure 12 shows the positions where the reprojection error is greater than 10 px.
The average reprojection error for each of the analyzed videos is shown in Table 4.
In general, it can be considered that the proposed method is able to achieve excellent results, obtaining an average reprojection error of only 3.7 px, which is below 0.3% of the width of the images. This value can be roughly translated into an estimated average error in the 3D localization of about 0.5 cm, although this measurement depends on other factors, such as the position of the plant tip. The total number of frames where there is a high error (greater than 10 px) is less than 5.4%. On the other hand, although the method required manual correction by an operator during some frames, as we have seen, this only occurred in 8.3% of the frames. In the remaining 91.7% of the frames, the automatic algorithm worked to locate the tip of the plants. Thus, the accuracy and robustness obtained by the proposed method in tracking the plant tips is suitable for practical use in circumnutation movement experimentation.
The direct comparison of our results with other methods is not feasible, since other works measure their results in terms of different parameters. For example, in the “Circumnutation Tracker” software of Stolarz et al. [11], manual annotation of the positions is required for all the frames of the videos. Tenzer and Clifford’s [12] plant monitoring method only segments the plants, without locating the 3D position of the tip, so the accuracy is given in terms of the area under the ROC curve for the classification problem. In the inertial tracker proposed by Geldhof et al. [15], the errors are expressed as the precision of pitch and roll angles, being under 0.5°. The recent method by Mao et al. [13] using deep learning is able to achieve an average error of 1.02 mm in the location of the apex of the plant, in videos of 640 × 480 px of resolution; however, only one video is analyzed per plant, so the 3D locations are not extracted.
The largest reprojection errors in our method are found in the area closest to the lateral camera. By overlaying the points used in the calibration onto the trace corresponding to the nutation movement of the plant tip, it is observed that not all the space travelled by the plant tip was covered. Furthermore, the greatest reprojection errors are located in the area where there are no calibration points, as shown in Figure 13.
To prevent this error in future experiments, it will be verified during the calibration process that there are enough calibration points across the entire space traveled by the plant tip in its nutation movement.

4. Conclusions

The study of the nutation movement of plants has aroused great interest in the research community in different areas, from philosophy to computer science, and has been a great source of inspiration for new algorithms and bioinspired systems. This paper has presented the development of a stereo vision system for studying the movement of climbing plants. The system consists of a pair of RGB cameras that synchronously record a side and top view of the plants in time-lapse. This study is focused on the common bean as a typical climbing plant model. Subsequently, the images are analyzed with a computer vision algorithm that obtains the tip of the plant and, using a previous calibration of the stereo pair, estimates the position in 3D coordinates. The method is able to extract the correct tip position in 86–98% of cases, depending on the video, with an average reprojection error below 4 px, which is translated to an approximate error in the 3D localization of about 0.5 cm. The proposed method allows researchers to know precisely and robustly the nutation movements of the plants and to compare their behavior under different situations, such as the use or absence of support structures for climbing.
In future work, it would be interesting to apply the latest deep learning methods to perform the accurate segmentation of the plants in the images, as well as subsequent matching between the stereo pair images for a complete estimation of the 3D position of the plant. In this way, it would be possible to know not only the position of the tip of the plant, but also of other parts such as the stems and leaves.

Author Contributions

Conceptualization, P.C.; methodology, A.P., P.C. and G.G.-M.; software, D.R.R.-M.; validation, D.R.R.-M. and G.G.-M.; investigation, D.R.R.-M., A.P., P.C. and G.G.-M.; resources, P.C. and A.P.; data curation, D.R.R.-M.; writing—original draft preparation, D.R.R.-M. and G.G.-M.; writing—review and editing, P.C.; visualization, D.R.R.-M.; supervision, P.C. and G.G.-M.; project administration, P.C.; and funding acquisition, P.C. and G.G.-M. All authors have read and agreed to the published version of this manuscript.

Funding

This research was partly funded by project 22130/PI/22 financed by the Region of Murcia (Spain) through the Regional Program for the Promotion of Scientific and Technical Research of Excellence (Action Plan 2022) of the Fundación Séneca-Agencia de Ciencia y Tecnología de la Región de Murcia and partly by the Office of Naval Research Global (Award # N62909-19-1-2015), awarded to P.C. and A.P.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data used in this study are available from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of this study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Mugnai, S.; Azzarello, E.; Masi, E.; Pandolfi, C.; Mancuso, S. Nutation in Plants. In Rhythms in Plants: Dynamic Responses in a Dynamic Environment; Springer: Cham, Switzerland, 2015; pp. 19–34. [Google Scholar] [CrossRef]
  2. Raja, V.; Silva, P.L.; Holghoomi, R.; Calvo, P. The dynamics of plant nutation. Sci. Rep. 2020, 10, 19465. [Google Scholar] [CrossRef] [PubMed]
  3. Gilroy, S. Plant tropisms. Curr. Biol. 2008, 18, R275–R277. [Google Scholar] [CrossRef] [PubMed]
  4. Sperotto, P.; Acevedo-Rodríguez, P.; Vasconcelos, T.N.; Roque, N. Towards a standardization of terminology of the climbing habit in plants. Bot. Rev. 2020, 86, 180–210. [Google Scholar] [CrossRef]
  5. Darwin, C.; Darwin, F. The Power of Movement in Plants; John Murray: London, UK, 1880. [Google Scholar]
  6. Gallentine, J.; Wooten, M.B.; Thielen, M.; Walker, I.D.; Speck, T.; Niklas, K. Searching and intertwining: Climbing plants and GrowBots. Front. Robot. AI 2020, 7, 118. [Google Scholar] [CrossRef]
  7. Meroz, Y. Plant tropisms as a window on plant computational processes. New Phytol. 2021, 229, 1911–1916. [Google Scholar] [CrossRef]
  8. Calvo, P.; Lawrence, N. Planta Sapiens: Unmasking Plant Intelligence; W. W. Norton & Company: Hachette, UK, 2022. [Google Scholar]
  9. Burris, J.N.; Lenaghan, S.C.; Stewart, C.N. Climbing plants: Attachment adaptations and bioinspired innovations. Plant Cell Rep. 2018, 37, 565–574. [Google Scholar] [CrossRef]
  10. Navarro, P.J.; Fernández, C.; Weiss, J.; Egea-Cortines, M. Development of a configurable growth chamber with a computer vision system to study circadian rhythm in plants. Sensors 2012, 12, 15356–15375. [Google Scholar] [CrossRef] [PubMed]
  11. Stolarz, M.; Żuk, M.; Król, E.; Dziubińska, H. Circumnutation Tracker: Novel software for investigation of circumnutation. Plant Methods 2014, 10, 24. [Google Scholar] [CrossRef]
  12. Tenzer, M.L.; Clifford, N.C. A digital green thumb: Neural networks to monitor hydroponic plant growth. In Proceedings of the 2020 Systems and Information Engineering Design Symposium (SIEDS), Charlottesville, VA, USA, 24 April 2020; pp. 1–6. [Google Scholar]
  13. Mao, Y.; Liu, H.; Wang, Y.; Brenner, E.D. A deep learning approach to track Arabidopsis seedlings’ circumnutation from time-lapse videos. Plant Methods 2023, 19, 18. [Google Scholar] [CrossRef] [PubMed]
  14. Díaz-Galián, M.V.; Perez-Sanz, F.; Sanchez-Pagán, J.D.; Weiss, J.; Egea-Cortines, M.; Navarro, P.J. A proposed methodology to analyze plant growth and movement from phenomics data. Remote Sens. 2019, 11, 2839. [Google Scholar] [CrossRef]
  15. Geldhof, B.; Pattyn, J.; Eyland, D.; Carpentier, S.; Van de Poel, B. A digital sensor to measure real-time leaf movements and detect abiotic stress in plants. Plant Physiol. 2021, 187, 1131–1148. [Google Scholar] [CrossRef] [PubMed]
  16. Guerra, S.; Peressotti, A.; Peressotti, F.; Bulgheroni, M.; Baccinelli, W.; D’Amico, E.; Gómez, A.; Massaccesi, S.; Ceccarini, F.; Castiello, U. Flexible control of movement in plants. Sci. Rep. 2019, 9, 16570. [Google Scholar] [CrossRef] [PubMed]
  17. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press eBooks: Cambridge, UK, 2003; Available online: https://digitalcollections.anu.edu.au/handle/1885/52332 (accessed on 21 January 2024 ).
  18. Zhang, Z.P. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  19. Lepetit, V.; Moreno-Noguer, F.; Fua, P. EPnP: An Accurate O(n) Solution to the PnP Problem. Int. J. Comput. Vis. 2008, 81, 155–166. [Google Scholar] [CrossRef]
  20. Hesch, J.A.; Roumeliotis, S.I. A Direct Least-Squares (DLS) method for PnP. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 383–390. [Google Scholar] [CrossRef]
  21. Garcia-Garcia, B.; Bouwmans, T.; Rosales, A. Background subtraction in real applications: Challenges, current models and future directions. Comput. Sci. Rev. 2020, 35, 100204. [Google Scholar] [CrossRef]
  22. Zivkovic, Z. Improved adaptive Gaussian mixture model for background subtraction. In Proceedings of the Pattern Recognition, ICPR 2004, of the 17th International Conference on Pattern Recognition, Cambridge, UK, 26 August 2004. [Google Scholar] [CrossRef]
Figure 1. Pipeline for the proposed computer vision method used to track the plant tip movement.
Figure 1. Pipeline for the proposed computer vision method used to track the plant tip movement.
Sensors 24 00747 g001
Figure 2. Images of the experiments conducted. At the top, zenithal and lateral images of the experiment in the no-pole condition; at the bottom, images of the pole condition.
Figure 2. Images of the experiments conducted. At the top, zenithal and lateral images of the experiment in the no-pole condition; at the bottom, images of the pole condition.
Sensors 24 00747 g002
Figure 3. Configuration of the stereo imaging system that was used. The recording cabin was a cylinder of 1 m height and 93 cm radius. The lateral camera was situated at 55 cm from the floor, the zenithal camera at 1.3 m, and the lamp at 1.5 m. The upper reflector had a height of 1 m.
Figure 3. Configuration of the stereo imaging system that was used. The recording cabin was a cylinder of 1 m height and 93 cm radius. The lateral camera was situated at 55 cm from the floor, the zenithal camera at 1.3 m, and the lamp at 1.5 m. The upper reflector had a height of 1 m.
Sensors 24 00747 g003
Figure 4. Points used in calibration of the stereo system (in green). Left: calibration scenario without pole. Right: calibration scenario with pole. The small red and blue arrows indicate the camera’s position in the scene.
Figure 4. Points used in calibration of the stereo system (in green). Left: calibration scenario without pole. Right: calibration scenario with pole. The small red and blue arrows indicate the camera’s position in the scene.
Sensors 24 00747 g004
Figure 5. Procedure designed for extracting the plant tip position.
Figure 5. Procedure designed for extracting the plant tip position.
Sensors 24 00747 g005
Figure 6. Algorithm for extracting possible positions of the plant tip in each frame of the top view.
Figure 6. Algorithm for extracting possible positions of the plant tip in each frame of the top view.
Sensors 24 00747 g006
Figure 7. Algorithm for extracting possible positions of the plant tip in each frame of the side view.
Figure 7. Algorithm for extracting possible positions of the plant tip in each frame of the side view.
Sensors 24 00747 g007
Figure 8. Sample view of the Plant Tracker tool developed to supervise the obtained locations. Top left: main window of the program; button “(1) Manual marking” opens the window labeled 1, and button “(2) Plant movement reconstruction” opens the window labeled 2.
Figure 8. Sample view of the Plant Tracker tool developed to supervise the obtained locations. Top left: main window of the program; button “(1) Manual marking” opens the window labeled 1, and button “(2) Plant movement reconstruction” opens the window labeled 2.
Sensors 24 00747 g008
Figure 9. Visualization of the obtained nutation movement of the bean plants. Left: position result for video pair 4 without pole. Middle: position result for video pair 7 without pole. Right: position result for video pair 10 without pole. Red and blue arrows indicate the camera’s position in the scene, and the green dot marks the position of the pot’s base.
Figure 9. Visualization of the obtained nutation movement of the bean plants. Left: position result for video pair 4 without pole. Middle: position result for video pair 7 without pole. Right: position result for video pair 10 without pole. Red and blue arrows indicate the camera’s position in the scene, and the green dot marks the position of the pot’s base.
Sensors 24 00747 g009
Figure 10. Visualization of reprojection errors of some videos along time. Top row: errors for video pair 4 without pole. Middle: errors for video pair 7 without pole. Bottom row: errors for video pair 10 without pole.
Figure 10. Visualization of reprojection errors of some videos along time. Top row: errors for video pair 4 without pole. Middle: errors for video pair 7 without pole. Bottom row: errors for video pair 10 without pole.
Sensors 24 00747 g010
Figure 11. A sample of an incorrect detection of the plant tip in the side view image.
Figure 11. A sample of an incorrect detection of the plant tip in the side view image.
Sensors 24 00747 g011
Figure 12. Visualization of the obtained 3D points in the analyzed videos, represented in a blue color. The points with a reprojection error greater than 10 px are marked in red. Red and blue arrows indicate the camera’s position in the scene, and the green dots mark the position of the pot’s base. Left: video pair 4 without pole. Middle: video pair 7 without pole. Right: video pair 10 without pole.
Figure 12. Visualization of the obtained 3D points in the analyzed videos, represented in a blue color. The points with a reprojection error greater than 10 px are marked in red. Red and blue arrows indicate the camera’s position in the scene, and the green dots mark the position of the pot’s base. Left: video pair 4 without pole. Middle: video pair 7 without pole. Right: video pair 10 without pole.
Sensors 24 00747 g012
Figure 13. Visualization of 3D points with a reprojection error greater than 10 px (marked in red) and calibrations points (marked in green). Left: video pair 4 without pole. Middle: video pair 7 without pole. Right: video pair 10 without pole. Red and blue arrows indicate the camera’s position in the scene, and the green dots mark the position of the pot’s base.
Figure 13. Visualization of 3D points with a reprojection error greater than 10 px (marked in red) and calibrations points (marked in green). Left: video pair 4 without pole. Middle: video pair 7 without pole. Right: video pair 10 without pole. Red and blue arrows indicate the camera’s position in the scene, and the green dots mark the position of the pot’s base.
Sensors 24 00747 g013
Table 1. Technical specifications of the cameras used in the stereo system.
Table 1. Technical specifications of the cameras used in the stereo system.
Characteristics
SensorType: 1/3″ HDR sensor
Dynamic range: 115 db
Resolution: 1.3 Mega pixel (1280 × 720 px)
Pixel size: 4.2 µm
Sensibility: 3650 mV/lux-sec
Optical lensType: CS Mount
Aperture: F 2.0
Field of view: 112°
Focal length: 19 mm
Table 2. Results of the stereo system calibration. The mean reprojection error (Mean RE) and the standard deviation of the reprojection error (Stdev RE) are shown for the top camera, the side camera, and the complete stereo system.
Table 2. Results of the stereo system calibration. The mean reprojection error (Mean RE) and the standard deviation of the reprojection error (Stdev RE) are shown for the top camera, the side camera, and the complete stereo system.
Reprojection Errors (in Pixels)
ScenarioMean RE Top CameraStdev RE
Top Camera
Mean RE Side CameraStdev RE Side CameraMean RE
Stereo System
Stdev RE
Stereo System
No Pole0.472020.345630.298310.274720.619300.55591
Pole0.305970.200530.284900.184050.910650.80017
Table 3. Results of the plant tip detection procedure in the images for the three video pairs. For each video pair, the condition (with or without a supporting pole), the frames that were analyzed, the camera (top or lateral), and the extraction of the tip location (automatic, manual, estimated or interpolated location) are indicated.
Table 3. Results of the plant tip detection procedure in the images for the three video pairs. For each video pair, the condition (with or without a supporting pole), the frames that were analyzed, the camera (top or lateral), and the extraction of the tip location (automatic, manual, estimated or interpolated location) are indicated.
Type of Detection
VideoPoleFrames 1ViewAutomaticManualEstimatedInterpolated
Pair 4No2483–4746Lateral158633129354
Top18787426547
Pair 7Yes3850–5798Lateral1607272246100
Top20632711322
Pair 7No750–2600Lateral128520828771
Top172453740
Pair 10No3850–5798Lateral1353251236109
Top157015120919
1 Range of frames that have been processed from the video.
Table 4. Results of the reprojection errors for the three video pairs. The mean reprojection error (Mean RE), the standard deviation of the reprojection error (Stdev RE), the total number of points analyzed, and the number of points with a reprojection error greater than 10 px (RE Points > 10 px) are shown for each video pair.
Table 4. Results of the reprojection errors for the three video pairs. The mean reprojection error (Mean RE), the standard deviation of the reprojection error (Stdev RE), the total number of points analyzed, and the number of points with a reprojection error greater than 10 px (RE Points > 10 px) are shown for each video pair.
Reprojection Errors (in Pixels)
VideoViewMean REStdev RETotal PointsRE Points > 10 px
Pair 4Lateral3.986792.84289226439
Top6.490045.743892264373
Pair 7Lateral2.054532.52104185133
Top2.663663.13255185147
Pair 10Lateral2.993912.62791194941
Top4.088863.317171949119
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ruiz-Melero, D.R.; Ponkshe, A.; Calvo, P.; García-Mateos, G. The Development of a Stereo Vision System to Study the Nutation Movement of Climbing Plants. Sensors 2024, 24, 747. https://doi.org/10.3390/s24030747

AMA Style

Ruiz-Melero DR, Ponkshe A, Calvo P, García-Mateos G. The Development of a Stereo Vision System to Study the Nutation Movement of Climbing Plants. Sensors. 2024; 24(3):747. https://doi.org/10.3390/s24030747

Chicago/Turabian Style

Ruiz-Melero, Diego Rubén, Aditya Ponkshe, Paco Calvo, and Ginés García-Mateos. 2024. "The Development of a Stereo Vision System to Study the Nutation Movement of Climbing Plants" Sensors 24, no. 3: 747. https://doi.org/10.3390/s24030747

APA Style

Ruiz-Melero, D. R., Ponkshe, A., Calvo, P., & García-Mateos, G. (2024). The Development of a Stereo Vision System to Study the Nutation Movement of Climbing Plants. Sensors, 24(3), 747. https://doi.org/10.3390/s24030747

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop