Next Article in Journal
A Review of Recent Developments in Driver Drowsiness Detection Systems
Previous Article in Journal
Real-Time Object Detection and Classification by UAV Equipped With SAR
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tree Trunk Recognition in Orchard Autonomous Operations under Different Light Conditions Using a Thermal Camera and Faster R-CNN

1
Graduate School of Science and Technology, University of Tsukuba, 1-1-1 Tennodai, Tsukuba 305-8577, Japan
2
Faculty of Life and Environmental Sciences, University of Tsukuba, 1-1-1 Tennodai, Tsukuba 305-8577, Japan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 2065; https://doi.org/10.3390/s22052065
Submission received: 7 December 2021 / Revised: 4 March 2022 / Accepted: 4 March 2022 / Published: 7 March 2022
(This article belongs to the Section Smart Agriculture)

Abstract

:
In an orchard automation process, a current challenge is to recognize natural landmarks and tree trunks to localize intelligent robots. To overcome low-light conditions and global navigation satellite system (GNSS) signal interruptions under a dense canopy, a thermal camera may be used to recognize tree trunks using a deep learning system. Therefore, the objective of this study was to use a thermal camera to detect tree trunks at different times of the day under low-light conditions using deep learning to allow robots to navigate. Thermal images were collected from the dense canopies of two types of orchards (conventional and joint training systems) under high-light (12–2 PM), low-light (5–6 PM), and no-light (7–8 PM) conditions in August and September 2021 (summertime) in Japan. The detection accuracy for a tree trunk was confirmed by the thermal camera, which observed an average error of 0.16 m for 5 m, 0.24 m for 15 m, and 0.3 m for 20 m distances under high-, low-, and no-light conditions, respectively, in different orientations of the thermal camera. Thermal imagery datasets were augmented to train, validate, and test using the Faster R-CNN deep learning model to detect tree trunks. A total of 12,876 images were used to train the model, 2318 images were used to validate the training process, and 1288 images were used to test the model. The mAP of the model was 0.8529 for validation and 0.8378 for the testing process. The average object detection time was 83 ms for images and 90 ms for videos with the thermal camera set at 11 FPS. The model was compared with the YOLO v3 with same number of datasets and training conditions. In the comparisons, Faster R-CNN achieved a higher accuracy than YOLO v3 in tree truck detection using the thermal camera. Therefore, the results showed that Faster R-CNN can be used to recognize objects using thermal images to enable robot navigation in orchards under different lighting conditions.

1. Introduction

Between 1995 and 2010, Japan’s agricultural labor force has gradually declined from 4.14 to 2.39 million, and its average age increased from 59.1 years to 65.8 years [1]. Therefore, agricultural robotics have the potential to support agricultural labor shortages and increase agricultural productivity in this critical stage of transformation [2]. In agricultural automation and robotics applications, vehicle navigation is important in outdoor environments, which are complex and uncertain compared to indoor conditions [3]. Open field navigation has had significant success using the Real Time Kinematic Global Navigation Satellite System (RTK-GNSS) with higher accuracy [4]. However, orchard navigation is the most complex, and interruption of RTK-GNSS signals due to high and dense canopies is frequently reported [5]. As Japanese orchards have net and dense branches, GNSS signals may be affected, and many farmland orchards do not have base stations set up to use GNSS directly. In addition, the performance of GNSS-based navigation depends highly on GNSS signal quality.
Therefore, orchard-based navigation systems remain a challenge for the development and application of agricultural robots. Light Detection and Ranging (LiDAR) is used to scan the surrounding environment in real time and returns accurate distance information using the time-of-flight principle. It can overcome the limitations of low light and interruption of RTK-GNSS signals. Laser sensors have been used for agricultural position detection and automatic coupling using artificial landmarks, concrete, and grass surface ground for navigation [6,7,8] or crop fields [9] by processing point clouds for crop and weed detection. However, LiDAR can only obtain distance and angular orientation information; it cannot provide information about the object’s type. To develop an intelligent robust system, LiDAR may not be sufficient for low-light conditions. Using a camera as a sensor is advantageous due to its low cost and ease of installation in agricultural robots. Camera sensors are widely used in navigation systems in various ways, such as using color information to segment paths in citrus groves for vehicle navigation [10], or focusing on color information of shadows and soil textures, to distinguish traversable areas (furrows) from impenetrable areas (ridges) for navigation [11]. However, common RGB cameras are easily affected by light, particularly in orchards, where the light differs between shaded and nonshaded areas, and the cameras typically experience exposure phenomena, making target identification unclear. Furthermore, low light is one of the most difficult aspects to create effective vision systems for agricultural robotics [12]. RGB cameras have high light requirements and work with low accuracy or are even unsuitable for low-light environments [13].
A thermal sensor measures the temperature of an object and the amount of heat it emits. Based on this imaging principle, a thermal camera was used to assess the health of tree trunks and detect damage in the interior [14]. Additionally, the thermal camera has a detection function that is not affected by visible light; thus, this study used a thermal camera to detect tree trunks in orchards for target detection. The target can be used to position a robot during navigation. In addition to thermal imagery, advancements in artificial intelligence and deep learning can be used for tree-trunk detection to enable vehicle navigation under different lighting conditions. Similar targets may have the same range of temperatures that emit and can easily localize the target using computer vision-based deep learning approaches [15]. Da Silva et al. compared the results of Yolo Only Look Once (YOLO) and Single Shot MultiBox Detector (SSD) for thermal images from forests, and the proposed use of thermal images allowed the execution of in-field forestry operations during the day and night [16].
In target detection, computer vision technology combines target localization and target classification; uses image processing, machine learning, and other techniques to identify the class of targets in the image; and returns the position of the target, generally using a rectangular box to indicate its position [17]. In general, the detection process is divided into two steps: (1) determine whether there is a target in the input image using target classification, and (2) perform labeling to search for the target using the bounding box [18]. This algorithm provides the accurate position of each target while accurately determining each target’s type. In the target-detection algorithm, images are stored as a pixel matrix, and the image features of the target class and position must be obtained before target detection [19]. The traditional method of machine learning is based on image processing, it lacks specificity for region selection, which typically causes sliding window redundancy and high time complexity, and in the process of dynamic change, the image contains a large amount of related noise, causes considerable interference for target detection, causes the target detection accuracy to be low, and easily misses detection and false detection [20].
After the introduction of a convolutional neural network (CNN) [21], deep learning has been more successful in target detection. Regions with CNN features (R-CNN) can use convolutional neural networks to solve target-detection problems, creating the framework of Region Proposal + CNN [22]. However, due to a series of problems such as low efficiency and long time spent, R-CNN has not been widely used. Compared with R-CNN, Fast R-CNN uses shared convolution to reduce the overall network consumption and improve training and testing speeds [23]. However, Fast R-CNN uses selective search, which requires 2~3 s to extract candidate regions and only 0.32 s to extract features for classification [24], which does not satisfy the demand of real-time applications. Faster R-CNN combines the fully convolutional network and replaces the original selective search with a region proposal network (RPN) [25], which markedly speeds up training and detection.
In recent years, the YOLO (You Only Look Once) algorithm [26] has defined the target detection problem as an image-space border regression and class prediction problem. YOLO contains only a single network without target candidate region extraction, which markedly reduces calculation complexity, and although the detection accuracy is not as good as that of Faster R-CNN, its detection speed is 10 times faster than that of Faster R-CNN [27]. YOLOv2 [28] and YOLOv3 [29] changed some structures due to low detection accuracies. YOLOv2 uses the anchor boxes to predict the location of objects in an image, batch normalization in the convolutional layers, and a high-resolution classifier to improve accuracy. In YOLOv3, the previous backbone network of YOLOv2 (Darknet-19) was replaced with Darknet-53, binary cross entropy was used in loss calculations, and logistic regression was used to predict the “objectness score” for each bounding box [30].
Both methods are used in research in various areas of agriculture, such as detecting cattle’s faces to manage livestock [31], the use of YOLO to estimate fruit load [32], to classify date fruits by color perception [33], and to detect ungulate to prevent them from damaging crops [34]. Comparing the results of Faster R-CNN and YOLOv3, Faster R-CNN achieves higher accuracies, although the accuracy of this algorithm comes at the cost of time complexity [35].
This study described a solution for low-light conditions in orchards that allows an effective vision system to be used for vehicle navigation, which requires accurate target detection with lower machine operation speed. Therefore, Faster R-CNN was used for object detection in this study. In addition, the Faster R-CNN has the potential to unlock thermal imagery in agricultural orchard applications under low- and no-light conditions to increase working efficiency. Thermal cameras use infrared information to detect targets in an orchard. Therefore, the objective of this study was to use a thermal camera for tree-trunk detection as target objects using Faster R-CNN to develop an autonomous speed sprayer navigation system. In the first part of this article, the thermal camera accuracy is analyzed to confirm the accuracy of object detection under the different lighting conditions while considering distances at different orientations. The details of Faster R-CNN and preparation of the training dataset are explained in this section. In the second part, the calibration, validation, and testing results are shown and analyzed. The strengths and potential of the research are discussed in the Section 3. Finally, the fourth part provides conclusions and proposes topics for future research on navigation systems for autonomous speed sprayers.

2. Materials and Methods

2.1. Calibration of Thermal Camera

A thermal camera (FLIR ADKTM, Teledyne FLIR LLC, Wilsonville, OR, USA) was used in each experiment, and calibration was performed to confirm the object detection performance under different lighting conditions while considering the orchard environment for different ranges (3, 5, 10, 15, 20, and 25 m) and orientations (0°, −30°, and 30° directions) of the thermal camera (Figure 1). The distance and orientations were measured using a laser distance meter (BOSCH GLM 250 VF).
Images were collected, and calibration was performed before each experiment to develop datasets from pear orchards. The experimental thermal camera was monocular and had a field of view (FOV) of 32°. The width and height of the target were required to be known to measure the distances. A 570 × 350 mm box was used as the target for confirmation of the accuracy of object detection by the thermal camera. According to the selected ranges and orientations, the returned distance and values were used to recognize the object in the image of the thermal camera.

2.2. Field Data Collection

The FLIR ADK® thermal camera has a resolution of 640 × 512, and images were collected from pear orchards at the Tsukuba-Plant Innovation Research Center, University of Tsukuba, Tsukuba, Ibaraki (36°06′56.8″ N, 140°05′37.7″ E). In this study, two types of pear orchards were used for data collection: (1) a conventional pear orchard, and (2) a joint-tree training pear orchard (Figure 2). Images were collected three times under high-light (12–2 PM), low-light (5–6 PM), and no-light (7–8 PM) conditions in both orchards (Table 1). Thermal images under different light conditions, such as no-, high-, and low-light conditions, were then analyzed (Figure 3).

2.3. Data Preparation

2.3.1. Image Frames from Videos

MATLAB® (Matrix Laboratory, Natick, MA, USA) was used to capture images every 15 frames and remove images that were blurred and that did not contain tree trunks. In total, 5313 images were analyzed, and the number of images of trunks in different periods was similar.

2.3.2. Labeling

LabelImg® was used to label datasets while preparing visual object class (VOC) datasets. When the position of the tree trunk on the image was selected, information such as the image name, object classification, and pixel coordinates were recorded in an .xml file.

2.3.3. Data Augmentation

To obtain a larger database and improve training accuracy, the original 5313 images were randomly flipped and rotated at random angles to obtain 7563 new images, and the labeling boxes of the images were also changed and saved as new .xml files.

2.3.4. Data Splitting

After labeling was completed, the data were divided into three sets: training, validation, and testing. Training was conducted to adjust the model parameters, and the classifier was adjusted to increase accuracy. Validation was performed to check the state and convergence of the model developed during training, and validation sets were used to adjust hyperparameters and determine which set of hyperparameters was most suitable. Testing was performed to evaluate the generalizability of the model, which was used for the validation set to determine the parameters.

2.4. Training Model Development

2.4.1. Faster R-CNN

In this study, the Faster R-CNN model was based on Fast R-CNN. Replacing the selective search in Fast R-CNN with RPN, to achieve more accurate regional proposals, reduce the redundancy of network computation, and improve detection speed. The structure of the Faster R-CNN model (Figure 4) and the network structure-selected images were kept for the datasets (Figure 5). The primary part of the model had convolutional layers, Regional Proposal Network (RPN), Region of Interest (ROI) pooling, and classification.
(1) Convolutional layers: As a CNN network target detection method, Faster R-CNN first used a set of basic convolutional + relu + pooling layers to extract feature maps of images for training datasets. The feature maps were shared for subsequent RPN layers and fully connected layers. The feature extraction process in Faster R-CNN is the same as that in CNN and can be performed using some common structures, such as the commonly used Visual Geometry Group (VGG) and ResNet [36]. In this study, the VGG16 model was used, which is a convolutional neural network model that was proposed by Simonyan and Zisserman [37]. VGG can be divided into 6 configurations (A, A-LRN, B, C, D, E) according to the size of the convolutional kernel and the number of convolutional layers. This study used the D configuration with 13 convolutional layers, 3 fully connected layers, and 5 pooling layers. The convolutional layer and fully connected layer had weight coefficients called weight layers (Figure 6).
The convolutional layer was used to extract features. A 3 × 3 convolutional kernel was used to slide over the image, and the output changed the image size because both the stride and padding were 1. As the linear model could not solve all problems, the ReLU layer was used to add nonlinear factors to simulate more subtle changes [38]. The pooling layer compressed the input feature map to make the feature map smaller and simplify the computational complexity of the network, and feature compression was used to extract the primary features. The pooling layer used a 2 × 2 convolutional kernel with a stride of 2 to perform the max pooling in the feature map. Therefore, the output feature map size was reduced by half. Finally, a fully connected layer connected all of the feature maps and transmitted the output to the softmax classifier.
In this VGGNet, there were 5 convolutional segments, all of which used 3 × 3 convolutional kernels because the 3 × 3 convolutional kernels can have fewer parameters and have more nonlinear transformations to increase the CNN’s ability to learn features [37].
(2) RPN: This network model is a new structure proposed by Faster R-CNN, which determines the approximate position of the target from the feature maps. RPN first generated many anchors (the candidate boxes) on the feature map. Then, the RPN network was divided into two lines: the upper line distinguished whether it was a tree trunk by softmax classification, and the lower line was used to calculate the bounding box regression offset for the anchors to obtain an accurate proposal [39]. The final proposal layer combined the anchors of the tree trunk and the corresponding bounding box regression offsets to obtain the proposals and eliminate the proposals that were too small or out of bounds.
(3) ROI pooling: After the RPN network had created the proposals, ROI pooling was applied to the proposals, and the feature maps were generated by the last layer in the VGG16 network. In addition, a fixed-size proposal feature map was obtained and sent to localization and recognition.
(4) Classification: Softmax was used to classify the target objects of tree trunk detection. The output classification contained the probability of the target being a tree trunk. Bounding box regression offset each proposal and was used to predict the target detection box more accurately.

2.4.2. Loss Function

The total loss function of Faster R-CNN consisted of classification loss and regression loss. The classification loss was calculated by softmax in the RPN classification layer, which was used to classify the anchors as positive and negative for training the network. Conversely, regression loss was calculated by the RPN bounding box regression layer and used to train the bounding box regression network. Therefore, the total loss function can be expressed as Equation (1):
L ( { p i } , { t i } = 1 N c l s i L c l s ( p i , p i * ) + λ 1 N r e g i p i * L r e g ( t i , t i * ) .  
where i is the anchor index, pi represents the positive probability of anchor I, p i * is the ground-truth label (if p i * is 1, the anchor is positive; if p i * is 0, the anchor is negative), t is the predicted bounding box, ti is the vector representing the 4 parameterized coordinates of the predicted bounding box, and t i * is a positive anchor associated with the ground-truth box.
Ncls is the trunk image during the training process, and Nreg is the number of anchor locations. As the difference between Ncls and Nreg was too large in the real process, the parameter λ was used to balance the two parameters. Therefore, the network total loss calculation process was considered for 2 types of losses: Lreg is the regression and Lcls is the classification loss function for the detection of trunks. The smooth function (L1) was used to estimate regression loss (Lreg), which is calculated and expressed as Equation (2):
L r e g ( t i , t i * ) = i { x , y , w , h } s m o o t h L 1 ( t i t i * ) .  
The smooth function is defined as Equation (3):
s o o m t h L 1 ( x ) = { 0.5 x 2 1 σ 2           | x |     1 σ 2 | x | 0.5           o t h e r w i s e .  
where x is the prediction error of the bounding box and the parameter σ is used to control the smoothing area.
For the bounding box regression, we used the 4 coordinates in the following parameterized expressions (4)–(11):
t x = ( x x a ) w a   .  
t y = ( y y a ) h a .  
t w = log ( w / w a ) .  
t h = log ( h / h a ) .  
t x * = ( x * x a ) w a .  
t y * = ( y * y a ) h a   .  
t w * = log ( w * / w a )   .  
t h * = log ( h * / h a ) .  
where x, y, w, and h denote the center coordinates of the box and its width and height, respectively. The variables x, xa, and x* are used for the predicted box, the anchor box, and the ground truth box, respectively, which can be considered a bounding box regression from the anchor box to the nearby real box.

2.5. Training Platform and Validation

The hardware environment used in this study included Windows 10, Intel(R) Core (TM) i7-10750H CPU @ 2.60 GHz, 32.0 GB of RAM, and an NVIDIA GeForce RTX 2060. The software environment included Anaconda 3, Python 3.5, and TensorFlow-GPU 1.13.1. In the experiment, 12,876 infrared images of orchard trunks were used for training, 9270 images in the training datasets, and 2318 images in the validation datasets.

2.6. Model Testing

To verify model reliability and stability, 1288 images were selected as the test set for validation after the model was trained. In this paper, the precision-recall curve was obtained by calculating precision and recall to highlight the trade-off between precision (P) and recall (R) of the classification model. The area under the curve was calculated to obtain the average precision (AP) of the model. As there was only one category in this experiment, AP was the same as mAP (mean average precision). The P, R, and AP can be expressed as expressions (12)–(14), respectively:
P   = T P T P + F P   .  
R   = T P T P + F N   .  
AP   = 0 1 P d R
where TP is the true positive value (i.e., the correct detection box); FP is the false-positive value (i.e., the false detection box that predicts the background as the target); and FN is the false negative value (i.e., the missed detection box). During testing, 1288 images were used as testing datasets in the Faster R-CNN model.

3. Results

3.1. Calibration Performance of Thermal Cameras in Different Lighting

Under three light conditions, images at different angles and distances from the thermal camera were analyzed with a laser distance meter to confirm the camera ranges for object detection (Figure 7). The horizontal coordinate represents the real distance of the box, the vertical coordinate represents the distance value measured by the thermal camera, and R2 represents the residual sum of squares, which represents the error between the real distance and the measured distance. The average errors for distance measurement were 0.16 m for 5 m, 0.24 m for 15 m, and 0.3 m for 20 m. Under low-light conditions with the thermal camera oriented at 30°, the highest error of 2.86 m was observed to locate the target point from the range of 25 m from the camera. This measurement error occurred when the range of the distance was increased to recognize orchard tree trunk detection. At 20 m, the thermal camera error markedly increased. Tree trunk detection by thermal cameras at distances below 20 m was more satisfactory to implement in the positioning system of autonomous vehicles and the detection of objects (Figure 8).

3.2. Model Training and Validation

The number of iterations were 40,000 (the training times are shown in Table 2), and the losses were reported as total loss, bounding box loss, classification loss, regional proposal network (RPN) classification loss, and RPN bounding box loss (Figure 9). According to Figure 9, the total loss function dropped to approximately 0.6 and oscillated near 0.6 when training reached 40,000 iterations, effectively reaching convergence. In addition, both the bounding box loss and classification loss converged; thus, training was stopped at 40,000 iterations. However, the results showed that the converged losses did not reach below 0.1, while other studies had less than 0.1 [40]; these results likely occurred because the two types of trees (conventional and joint) were investigated in this study, which differed markedly in shape. In addition, in the conventional orchard, the trees differed markedly in size, with some trees being 30 cm wide and some less than 5 cm, which led to a larger loss convergence result. In addition, randomly selected images were used to prepare the validation set. There were original images and randomly flipped and rotated images under the three lighting conditions (Figure 10). Compared with YOLOv3 using the same dataset and the same number of iterations, the mAP of YOLOv3 was 0.5024 and Faster R-CNN could reach 0.8529 (Figure 11). According to the validation results, the model worked properly to detect the tree trunk in conventional and joint orchards under high-, low-, and no-light conditions, and had a higher accuracy rate. The poles and shelves in the orchards could be distinguished from the fruit tree trunk. In the rotated images, the trunk could be identified accurately.

3.3. Model Testing

The mAP of the Faster R-CNN model was found to be 0.8378 (Figure 12). The images for testing were considered based on the training results conducted for the high-, low-, and no-light conditions (Figure 13). Two images were selected at different time periods, and the detection speed was related to the number of object proposals. In this study, we used 300 object proposals, and the detection speed was 83 ms in images and 90 ms in videos (Table 3). In the testing results in the conventional and joint orchards, this model was able to recognize tree trunks accurately. At longer distances of more than 20 m, image clarity decreased, and some errors in labeling the image occurred, which resulted in some omission and errors in the testing results for the recognition objects.

4. Discussion

Machine vision typically analyzes RGB imagery and thus cannot manage low-light conditions to recognize natural landmarks. Conversely, GNSS suffers signal interruptions due to dense canopies inside the orchards. LiDAR also has difficulties recognizing tree trunks because the scanning range decreases the chance of recognizing tree trunks [5]. To increase the robustness of these methods, this study used a thermal camera combined with deep learning to identify fruit tree trunks under high-, low-, and no-light conditions. The loss function shown in the Section 3 represents the difference between the predicted value of the model and the training samples. A smaller value indicates that the closer the predicted sample is to the real sample, the more robust the model is. In contrast, a larger value indicates that the difference between the predicted and real samples is larger [41]. The loss function in this research oscillated at approximately 0.6 due to the large variation in the shape and size of the trunks. According to the results, the accuracy of the Faster R-CNN testing was 83.8%. Compared with YOLO v3, Faster R-CNN had higher accuracy in the thermal image of this research. The model could distinguish poles and shelves of the same size, indicating that the thermal camera can be used for orchard navigation to detect tree trunks under different light conditions, and using this system can allow machinery to operate in orchards at any time. However, due to the unavoidable error in labeling, the recognition accuracies decreased to 20 m after the thermal camera and yielded some omissions and errors. In addition, for the thermal camera in monocular mode, distance measuring must be assisted by other sensors. The accuracy of the deep learning model can be improved with more experimental datasets that describe other lighting conditions and by improving the training process.
With comparison to our research outcomes, another contribution in tree trunk detection for forestry was conducted using different cameras (visible and thermal) to capture forest images [16]. Several deep learning-based nets were used to compare the accuracy and speed of detection through training, validation, and testing processes. The visible and thermal cameras can be used with mainly YOLO tiny and SSD for in-field forestry operations. In the results, visible imagery had a higher precision than thermal imagery. In forestry application, tree trunk detection can be helpful for mobile applications of robots; however, this research needs further extension to detect tree trunks at the different lighting and shadow stages, which was not discussed. In contrast, the presented research used one thermal camera to capture orchard images at the different times, demonstrated that thermal camera can be used in various light conditions through deep learning, and provided the basis for orchard navigation system. In the orchard navigation, more challenges exist with poles and the shape of trees. Our research considered different illumination times, poles, and included conventional pear orchard, and joint-tree training pear orchard together.
Therefore, in the joint tree system, the thermal camera has more flexibility to detect the tree trunk due to the uniform shape of the tree and growth in rows. However, natural orchards have different types, and the shapes of the field and canopy coverages can be markedly different. Traditionally, orchard growers in Japan control pruning and training according to their heights of operations, which makes using machinery inconvenient. Joint tree training systems have the advantage of uniform growing and machinery automation systems. Therefore, the thermal camera used in the research has a high application potential in joint tree training systems to use for orchard navigation in combination with other positional sensors such as LiDAR. In previous studies, LiDAR was able to detect artificial landmarks and cones for use in positional algorithms in orchard navigation [42,43]. In future research, a thermal camera is planned to be used to navigate speed sprayers in an orchard.

5. Conclusions

Thermal imagery has the potential to detect natural landmarks under different lighting conditions. This study proposed using the Faster R-CNN for thermal images to detect pear tree trunks to enable navigation under various lighting conditions in orchards. Conventional and joint tree training system orchards were considered for dataset collection. The accuracy and reliability of the model were verified by calibrating, training, and testing the thermal images, indicating that target detection could be performed under high-, low-, and no-light conditions using a thermal camera. In this study, all 640 × 512-pixel images and videos were used; 300 object proposals were generated for each image and frame; and this model achieved a mAP of 0.8378 in 83 ms with images and in 90 ms with videos. Faster R-CNN had higher accuracy compared to YOLO v3 using the same dataset and trained the same times, and speed sprayers typically use a lower speed (e.g., 3 m/s) in orchards to reduce the drift of pesticides and effective penetration of chemicals inside canopies. Therefore, the thermal camera can be used under high-, low-, or no-light conditions inside orchards to detect tree trunks using Faster R-CNN. In future research, an autonomous speed sprayer is planned to be used with a thermal camera and in combination with LiDAR for orchard navigation under different lighting conditions to provide solutions or orchard automation to increase productivity.

Author Contributions

Conceptualization, T.A.; Data curation, A.J.; Funding acquisition, T.A.; Investigation, A.J.; Methodology, A.J.; Resources, R.N.; Writing—review & editing, T.A. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by grants from the Japanese Society of Promotion Science JSPS Kiban C (Grant Number 21K05844).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset generated and analyzed during this study are available from the corresponding author upon reasonable request, but restrictions apply to the data reproducibility and commercially confident details.

Acknowledgments

The authors would like to thank the Tsukuba Plant Innovation Research Center (T-PIRC), University of Tsukuba, for providing facilities for conducting this research in its orchards.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, Y. Japan: Aging of the Agricultural Labor Force and its Solutions; World food Prize Foundation: Des Moines, IA, USA, 2018. [Google Scholar]
  2. Vadlamudi, S. How Artificial Intelligence Improves Agricultural Productivity and Sustainability: A Global Thematic Analysis. Asia Pac. J. Energy Environ. 2019, 6, 91–100. [Google Scholar]
  3. Bergerman, M.; Billingsley, J.; Reid, J.; Van Henten, E. Robotics in agriculture and forestry. In Springer Handbook of Robotics; Springer Science and Business Media LLC: Berlin/Heidelberg, Germany, 2016; pp. 1463–1492. [Google Scholar]
  4. Takai, R.; Barawid, O., Jr.; Ishii, K.; Noguchi, N. Development of Crawler-Type Robot Tractor based on GPS and IMU. IFAC Proc. Vol. 2010, 43, 151–156. [Google Scholar] [CrossRef]
  5. Li, M.; Imou, K.; Wakabayashi, K.; Yokoyama, S. Review of research on agricultural vehicle autonomous guidance. Int. J. Agric. Biol. Eng. 2009, 2, 1–16. [Google Scholar]
  6. Ahamed, T.; Takigawa, T.; Koike, M.; Honma, T.; Hasegawa, H.; Zhang, Q. Navigation using a laser range finder for autonomous tractor (part 1)—positioning of implement. J. Jpn. Soc. Agric. Mach. 2006, 68, 68–77. [Google Scholar]
  7. Ahamed, T.; Takigawa, T.; Koike, M.; Honma, T.; Hasegawa, H.; Zhang, Q. Navigation using a laser range finder for autonomous tractor (part 2)—Navigation for approach composed of multiple paths. J. Jpn. Soc. Agric. Mach. 2006, 68, 78–86. [Google Scholar]
  8. Ahamed, T.; Tian, L.; Takigawa, T.; Zhang, Y. Development of Auto-Hitching Navigation System for Farm Implements Using Laser Range Finder. Trans. ASABE 2009, 52, 1793–1803. [Google Scholar] [CrossRef]
  9. Malavazi, F.B.; Guyonneau, R.; Fasquel, J.-B.; Lagrange, S.; Mercier, F. LiDAR-only based navigation algorithm for an autonomous agricultural robot. Comput. Electron. Agric. 2018, 154, 71–79. [Google Scholar] [CrossRef]
  10. Subramanian, V.; Burks, T.F.; Arroyo, A.A. Development of machine vision and laser radar based autonomous vehicle guidance systems for citrus grove navigation. Comput. Electron. Agric. 2006, 53, 130–143. [Google Scholar] [CrossRef]
  11. Takagaki, A.; Masuda, R.; Michihisa, I.; Masahiko, S. Image Processing for Ridge/Furrow Discrimination for Autonomous Agricultural Vehicles Navigation. IFAC Proc. 2013, 46, 47–51. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Chen, H.; He, Y.; Ye, M.; Cai, X.; Zhang, D. Road segmentation for all-day outdoor robot navigation. Neurocomputing 2018, 314, 316–325. [Google Scholar] [CrossRef]
  13. Guo, Z.; Li, X.; Xu, Q.; Sun, Z. Robust semantic segmentation based on RGB-thermal in variable lighting scenes. Measurement 2021, 186, 110176. [Google Scholar] [CrossRef]
  14. Beyaz, A.; Özkaya, M.T. Canopy analysis and thermographic abnormalities determination possibilities of olive trees by using data mining algorithms. Not. Bot. Horti Agrobot. Cluj-Napoca 2021, 49, 12139. [Google Scholar] [CrossRef]
  15. Hespeler, S.C.; Nemati, H.; Ehsan, D.-N. Non-destructive thermal imaging for object detection via advanced deep learning for robotic inspection and harvesting of chili peppers. Artif. Intell. Agric. 2021, 5, 102–117. [Google Scholar] [CrossRef]
  16. da Silva, D.Q.; Dos Santos, F.N.; Sousa, A.J.; Filipe, V. Visible and Thermal Image-Based Trunk Detection with Deep Learning for Forestry Mobile Robotics. J. Imaging 2021, 7, 176. [Google Scholar] [CrossRef]
  17. Wang, K.; Meng, Z.; Wu, Z. Deep Learning-Based Ground Target Detection and Tracking for Aerial Photography from UAVs. Appl. Sci. 2021, 11, 8434. [Google Scholar] [CrossRef]
  18. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  19. Li, W.; Feng, X.S.; Zha, K.; Li, S.; Zhu, H.S. Summary of Target Detection Algorithms. J. Phys. Conf. Ser. 2021, 1757, 012003. [Google Scholar] [CrossRef]
  20. Qi, L.; Li, B.; Chen, L.; Wang, D.; Dong, L.; Jia, X.; Huang, J.; Ge, C.; Xue, G. Ship Target Detection Algorithm Based on Improved Faster R-CNN. Electronics 2019, 8, 959. [Google Scholar] [CrossRef] [Green Version]
  21. Alex, K.; Ilya, S.; Geoffrey, E.H. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Processing Syst. 2012, 25, 1097–1105. [Google Scholar]
  22. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 14–19 June 2020; 2014; pp. 580–587. [Google Scholar]
  23. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision, New York, NY, USA, 6 August 2002; 2015; pp. 1440–1448. [Google Scholar]
  24. Chahal, K.S.; Dey, K. A Survey of Modern Object Detection Literature using Deep Learning. arXiv 2018, arXiv:1808.07256. [Google Scholar]
  25. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 36, 1137–1149. [Google Scholar]
  26. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; 2016; pp. 779–788. [Google Scholar]
  27. Wen, H.; Dai, F.; Yuan, Y. A Study of YOLO Algorithm for Target Detection. J. Adv. Inn Artif. Life Robot. 2021, 2, 287–290. [Google Scholar] [CrossRef]
  28. Redmon, J.; Farhadi, A. YOL09000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 18–20 June 1996; 2016; pp. 7263–7271. [Google Scholar]
  29. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  30. Parico, A.I.B.; Ahamed, T. Real Time Pear Fruit Detection and Counting Using YOLOv4 Models and Deep SORT. Sensors 2021, 21, 4803. [Google Scholar] [CrossRef]
  31. Xu, B.; Wang, W.; Guo, L.; Chen, G.; Wang, Y.; Zhang, W.; Li, Y. Evaluation of Deep Learning for Automatic Multi-View Face Detection in Cattle. Agriculture 2021, 11, 1062. [Google Scholar] [CrossRef]
  32. Mirhaji, H.; Soleymani, M.; Asakereh, A.; Mehdizadeh, S.A. Fruit detection and load estimation of an orange orchard using the YOLO models through simple approaches in different imaging and illumination conditions. Comput. Electron. Agric. 2021, 191, 106533. [Google Scholar] [CrossRef]
  33. Beyaz, A.; Nourani, A. Date Fruit Varieties Classification Based on Dryness Levels by Using YOLOv3. Ama. Agric. Mech. Asia. Afr. Lat. Am. 2021, 51, 1193–1210. [Google Scholar]
  34. Adami, D.; Ojo, M.O.; Giordano, S. Design, Development and Evaluation of an Intelligent Animal Repelling System for Crop Protection Based on Embedded Edge-AI. IEEE Access 2021, 9, 132125–132139. [Google Scholar] [CrossRef]
  35. Srivastava, S.; Divekar, A.V.; Anilkumar, C.; Naik, I.; Kulkarni, V.; Pattabiraman, V. Comparative analysis of deep learning image detection algorithms. J. Big Data 2021, 8, 1–27. [Google Scholar] [CrossRef]
  36. Mahmud, S.; Zahid, A.; Das, A.K.; Muzammil, M.; Khan, M.U. A systematic literature review on deep learning applications for precision cattle farming. Comput. Electron. Agric. 2021, 187, 106313. [Google Scholar] [CrossRef]
  37. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  38. Oostwal, E.; Straat, M.; Biehl, M. Hidden unit specialization in layered neural networks: ReLU vs. sigmoidal activation. Phys. A Stat. Mech. Appl. 2021, 564, 125517. [Google Scholar] [CrossRef]
  39. Li, Z.; Li, Y.; Yang, Y.; Guo, R.; Yang, J.; Yue, J.; Wang, Y. A high-precision detection method of hydroponic lettuce seedlings status based on improved Faster R-CNN. Comput. Electron. Agric. 2021, 182, 106054. [Google Scholar] [CrossRef]
  40. Jiang, D.; Li, G.; Tan, C.; Huang, L.; Sun, Y.; Kong, J. Semantic segmentation for multiscale target based on object recognition using the improved Faster-RCNN model. Future Gener. Comput. Syst. 2021, 123, 94–104. [Google Scholar] [CrossRef]
  41. Zhang, H.; Tian, Y.; Wang, K.; Zhang, W.; Wang, F.-Y. Mask SSD: An effective single-stage approach to object instance segmentation. IEEE Trans. Image Process 2019, 29, 2078–2093. [Google Scholar] [CrossRef]
  42. Thanpattranon, P.; Ahamed, T.; Takigawa, T. Navigation of an Autonomous Tractor for a Row-Type Tree Plantation Using a Laser Range Finder—Development of a Point-to-Go Algorithm. Robotics 2015, 4, 341–364. [Google Scholar] [CrossRef] [Green Version]
  43. Thanpattranon, P.; Ahamed, T.; Takigawa, T. Navigation of autonomous tractor for orchards and plantations using a laser range finder: Automatic control of trailer position with tractor. Biosyst. Eng. 2016, 147, 90–103. [Google Scholar] [CrossRef]
Figure 1. Object detection under different lighting conditions at a range of 25 m from the thermal camera with orientations of 0°, −30°, and 30° directions (green indicates the measured distance by camera).
Figure 1. Object detection under different lighting conditions at a range of 25 m from the thermal camera with orientations of 0°, −30°, and 30° directions (green indicates the measured distance by camera).
Sensors 22 02065 g001
Figure 2. Aerial view of experimental pear orchards: (a) conventional planted pear orchard, and (b) joint-tree training pear orchard in the Tsukuba-Plant Innovation Research Center (T-PIRC), University of Tsukuba, Japan.
Figure 2. Aerial view of experimental pear orchards: (a) conventional planted pear orchard, and (b) joint-tree training pear orchard in the Tsukuba-Plant Innovation Research Center (T-PIRC), University of Tsukuba, Japan.
Sensors 22 02065 g002
Figure 3. Analysis of thermal images under different lighting conditions: (a,b) no-light conditions (7–8 PM), (c,d) high-light conditions (12–2 PM), and (e,f) low-light conditions (5–6 PM).
Figure 3. Analysis of thermal images under different lighting conditions: (a,b) no-light conditions (7–8 PM), (c,d) high-light conditions (12–2 PM), and (e,f) low-light conditions (5–6 PM).
Sensors 22 02065 g003
Figure 4. Faster R-CNN structure for tree truck detection used in this research.
Figure 4. Faster R-CNN structure for tree truck detection used in this research.
Sensors 22 02065 g004
Figure 5. Faster R-CNN network structure-focusing regional proposal network for feature map.
Figure 5. Faster R-CNN network structure-focusing regional proposal network for feature map.
Sensors 22 02065 g005
Figure 6. VGG16 for target of tree trunk detection.
Figure 6. VGG16 for target of tree trunk detection.
Sensors 22 02065 g006
Figure 7. Measured distance under different lighting and orientations. (ac) 0°, −30°, and 30° under high-light condition; (df) 0°, −30°, and 30° under low-light condition; (gi) 0°, −30°, and 30° under no-light condition.
Figure 7. Measured distance under different lighting and orientations. (ac) 0°, −30°, and 30° under high-light condition; (df) 0°, −30°, and 30° under low-light condition; (gi) 0°, −30°, and 30° under no-light condition.
Sensors 22 02065 g007aSensors 22 02065 g007b
Figure 8. Measurement errors of target objects from a distance of 0 to 25 m under different lighting conditions and orientations of the thermal camera.
Figure 8. Measurement errors of target objects from a distance of 0 to 25 m under different lighting conditions and orientations of the thermal camera.
Sensors 22 02065 g008
Figure 9. Faster R-CNN loss images. (a) Total loss. (b) Bounding box loss. (c) Classification loss. (d) Regional Proposal Network classification loss. (e) Regional Proposal Network bounding box loss.
Figure 9. Faster R-CNN loss images. (a) Total loss. (b) Bounding box loss. (c) Classification loss. (d) Regional Proposal Network classification loss. (e) Regional Proposal Network bounding box loss.
Sensors 22 02065 g009aSensors 22 02065 g009b
Figure 10. Validation results of Faster R-CNN. (af) Original images. (gl) Randomly flipped and rotated images.
Figure 10. Validation results of Faster R-CNN. (af) Original images. (gl) Randomly flipped and rotated images.
Sensors 22 02065 g010aSensors 22 02065 g010b
Figure 11. Precision-recall curve of Faster R-CNN and YOLOv3 validation.
Figure 11. Precision-recall curve of Faster R-CNN and YOLOv3 validation.
Sensors 22 02065 g011
Figure 12. Precision-recall curve of Faster R-CNN and YOLOv3 testing.
Figure 12. Precision-recall curve of Faster R-CNN and YOLOv3 testing.
Sensors 22 02065 g012
Figure 13. Image results of Faster R-CNN test: (ad) no-light conditions, (eh) high-light conditions, and (il) low-light conditions.
Figure 13. Image results of Faster R-CNN test: (ad) no-light conditions, (eh) high-light conditions, and (il) low-light conditions.
Sensors 22 02065 g013aSensors 22 02065 g013b
Table 1. Dataset collection times and light conditions in orchards.
Table 1. Dataset collection times and light conditions in orchards.
DateTimeLight Condition
24 August 202119:00–20:00No light
26 August 202113:00–14:00Strong light
6 September 202117:00–18:00Low light
Table 2. Iterations and training time of the deep learning model using Faster R-CNN.
Table 2. Iterations and training time of the deep learning model using Faster R-CNN.
Iterations Times10,00020,00030,00040,000
Training time3.05 h6.02 h8.98 h11.97 h
Table 3. Tree trunk detection time using the deep learning model of Faster R-CNN.
Table 3. Tree trunk detection time using the deep learning model of Faster R-CNN.
Objects ProposalsFPSDetection Time
(Images, ms)
Detection Time
(Videos, ms)
10013.97272
20012.38181
30011.18390
40010.69395
5009.995101
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, A.; Noguchi, R.; Ahamed, T. Tree Trunk Recognition in Orchard Autonomous Operations under Different Light Conditions Using a Thermal Camera and Faster R-CNN. Sensors 2022, 22, 2065. https://doi.org/10.3390/s22052065

AMA Style

Jiang A, Noguchi R, Ahamed T. Tree Trunk Recognition in Orchard Autonomous Operations under Different Light Conditions Using a Thermal Camera and Faster R-CNN. Sensors. 2022; 22(5):2065. https://doi.org/10.3390/s22052065

Chicago/Turabian Style

Jiang, Ailian, Ryozo Noguchi, and Tofael Ahamed. 2022. "Tree Trunk Recognition in Orchard Autonomous Operations under Different Light Conditions Using a Thermal Camera and Faster R-CNN" Sensors 22, no. 5: 2065. https://doi.org/10.3390/s22052065

APA Style

Jiang, A., Noguchi, R., & Ahamed, T. (2022). Tree Trunk Recognition in Orchard Autonomous Operations under Different Light Conditions Using a Thermal Camera and Faster R-CNN. Sensors, 22(5), 2065. https://doi.org/10.3390/s22052065

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop