Next Article in Journal
Performance Evaluation of a Wet Medium Made of Mangosteen Peels for a Direct Evaporative Cooling System
Previous Article in Journal
Artificial Neural Networks for Modeling and Optimizing Egg Cost in Second-Cycle Laying Hens Based on Dietary Intakes of Essential Amino Acids
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Segmentation of Leaves and Fruits of Tomato Plants by Color Dominance

by
Juan Pablo Guerra Ibarra
1,
Francisco Javier Cuevas de la Rosa
1,* and
Oziel Arellano Arzola
2
1
Centro de Investigaciones en Óptica A.C., Leon 37150, Guanajuato, Mexico
2
Instituto Tecnológico Nacional de México, Campus Zamora, Zamora 59720, Michoacan, Mexico
*
Author to whom correspondence should be addressed.
AgriEngineering 2023, 5(4), 1846-1864; https://doi.org/10.3390/agriengineering5040113
Submission received: 3 August 2023 / Revised: 26 August 2023 / Accepted: 28 September 2023 / Published: 12 October 2023

Abstract

:
The production of food generated by agriculture has been essential for civilizations throughout time. Tillage of fields has been supported by great technological advances in several areas of knowledge, which have increased the amount of food produced at lower costs. The use of technology applied to modern agriculture has generated a research area called precision agriculture, which has providing crops with resources in an exact amount at a precise moment as one of its most relevant objectives The data analysis process in precision agriculture systems begins with the filtering of the information available, which can come from sources such as images, videos, and spreadsheets. When the information source is digital images, the process is known as segmentation, which consists of assigning a category or label to each pixel of the analyzed image. In recent years, different algorithms of segmentation have been developed that make use of different pixel characteristics, such as color, texture, neighborhood, and superpixels. In this paper, a method to segment images of leaves and fruits of tomato plants is presented, which is carried out in two stages. The first stage is based on the dominance of one of the color channels over the other two, using the R G B color model. In the case of the segmentation of the leaves, the green channel dominance is used, whereas the dominance of red channel is used for the fruits. In the second stage, the false positives generated during the previous stage are eliminated by using thresholds calculated for each pixel that meets the condition of the first stage. The results are measured by applying performance metrics: Accuracy, Precision, Recall, F1-Score, and Intersection over Union. The results for segmentation of the fruit and leaves of the tomato plants with the highest metrics is Accuracy with 98.34% for fruits and Recall with 95.08% for leaves.

1. Introduction

Agriculture is a fundamental economic activity for the subsistence of human beings, and it has left its mark in history, making differences in the development of civilizations in different geographical locations and times [1,2]. Agricultural progress has been a fundamental basis for the growth of human populations, because the production of sufficient food in quantity and quality depends on it. By the year 2050, the human population will have grown to about 10 billion people, according to United Nations data [3], which implies great challenges for different food production activities.
In the last two decades, technological advances applied to the tillage of the land have generated what is called precision agriculture (PA). It encompasses a set of technologies that combine sensors, statistics, classical algorithms, and algorithms of artificial intelligence (AI) especially those of computer vision (CV) [1]. The increase in the amount of food from the cultivation of the fields and the optimization of resources used for this are some of the main goals of the PA. These can be achieved by monitoring certain characteristics of the crops, such as growth, irrigation, fertilization, detection of pests and diseases [4,5,6].
Currently, several CV methods have been developed to support agricultural activity in tasks such as the estimation of fruit quality [7,8,9], recognition of pests [10,11,12], improvement of irrigation systems [13] and nutrient deficiency detection [14].
Color attributes in digital images are used to segment different crop elements, such as leaves, fruits, and weeds, among others, from the rest of the elements present in it [15]. Rasmussen [16] evaluated the level of leaf development in fields free of weeds, highlighting the importance of the conditions under which these images are acquired for obtaining successful results such as camera angle and light conditions among others. Kirk [17] estimated the amount of vegetation or foliage present in images of cereal crops at early stages of phenological development. In [18], Story presented a work to determine overall plant growth and health status. In this method, RGB (red, green, blue) and HSL (hue, saturation, lightness) formats values as are used as color features. Wang [19] proposed a method to segment rice plants from the background of the image based on subtracting the value of the green channel from the value of the red channel for the pixels of the image. The results of the segmentation process are used to estimate the amount of nitrogen present in the leaves. Quemada [20] carried out a segmentation process in hyperspectral images for the estimation of nitrogen present in corn crops. Additionally, there are computational methods implemented to detect weeds in uncontrolled light conditions, as was reported by Jeon [21], in which images were acquired using an autonomous robot. Yadav [22] measured the amount of chlorophyll found in potato crops using CV algorithms. Philipp [23] performed a series of segmentation comparisons using different color representation models. In [24], Menesatti proposed the use of a rapid, non-destructive, cost-effective technique to predict the nutritional status of orange leaves utilizing a Vis–NIR (visible–near infrared) portable spectrophotometer. Fan [25] developed a method for segmenting apples combining local image features and color information through a pixel patch segmentation method based on a gray-centered R G B color model space to address this issue.
At a more specific level, there are papers that describe methods for the segmentation and analysis of different elements that are part of the plants in crops, which use different CV techniques. For example, Xu [26] reported a method for extracting color and textures characteristics of the leaves of tomato plants, which is based on histograms and Fourier transforms. Wan [27] proposed a procedure to measure the maturity of fresh supermarket tomatoes at three different levels through the development of a threshold segmentation algorithm based on the R G B color model, and classification is performed using a backpropagation neural network. Tian [28] used an improved k-means algorithm based on the adaptive clustering number for the segmentation of tomato leaf images. Castillo-Martínez [29] reported a color index-based thresholding method for background and foreground segmentation of plant images utilizing two color indexes which are modified to provide better information about the green color of the plants. Lin [30] proposed a detection algorithm based on color, depth, and shape information for detecting spherical or cylindrical fruits on plants. Lu [31] presented a method for automatic segmentation of plants from background in color images, which consists of the unconstrained optimization of a linear combination of R G B color model component images to enhance the contrast between plant and background regions.
In recent years, a new image processing technique called deep learning ( D L ) has been developed. It consists of several types of convolutional neural network (CNN) models [32], for example: LeNet [33], AlexNet [34], VGG-16 [35], and Inception [36]. The capabilities and applications of CNN models have increased as well as the number of trainable parameters in them. These are a function of the number of layers used; therefore, highly specialized hardware has been required for their training. In PA, the use of DL has been successfully used in different contexts, for example, for pest and disease detection [37,38,39], leaf identification [11,40], and estimation of nutrients present in plant leaves [41]. The development of CNN models for performing the separation of items of interest from items of non-interest has been explored in several papers. In [42]. Milioto used an RCNN model to segment sugar beet plants, weeds, and background using R G B images. Majjed [43] trained different CNN models based on SegNet and FNC to segment grapevine cordons and determine their trajectories. Kang [44] proposed several CNN models based on DaSNet and a ResNet50 backbone for real-time semantic apple detection and segmentation.
This paper details a segmentation method applied to images taken in greenhouses with tomato crops, classifying the pixels into three classes: leaves, fruits, and background. This is based on segmentation using the dominance of a color channel with respect to the others and, in a following stage, the determination of thresholds using the same color channel information. This method has the advantages of ease of implementation and low computational cost.
The reminder of this paper is organized as follows. The segmentation method developed to separate the leaves and fruits of the tomato plants is described in detail in Section 2. In Section 3, different images generated during the segmentation process of the leaves and fruits are shown. Additionally, tables are displayed with the metrics selected to measure performance. A comparison of results generated by the developed segmentation method against those of a CNN model is made in Section 4. In Section 5, a commentary on the performance of our method is provided. Lastly, the conclusions are presented in Section 6.

2. Method

The different methods of separating the elements present in images into portions that are easier to analyze are called segmentation methods [45,46,47]. These can be classified into the following categories:
  • Region-based methods. These methods are based on separating a group of pixels that are connected and share properties. This technique performs well on noisy images.
  • Edge-based methods. These algorithms are generally based on the discontinuity of the pixel intensities of the images to be segmented, which are manifested at the edges of the objects.
  • Feature-based clustering methods. These methods are based on looking for similarities between the objects present in the images; this allows for the creation of categories of interest for a particular objective.
  • Threshold methods. These methods are based on a comparison of the pixel intensity value against a threshold value T. There are two types of threshold segmentation methods depending on the value T; if it is constant, it is called global threshold segmentation, otherwise it is called local threshold segmentation.
The following sections detail the proposed method of using color dominance to segment all pixels of digital images acquired inside greenhouses into three different classes: leaves, tomato plant fruits, and background, which is based on the calculation of local thresholds for each pixel.

RGB Thresholding

The R G B color model is made up of three components: one for each primary color. These can have values ranging from 0 to 255, allowing any color in the visible spectrum to be represented.
The method uses a two-stage algorithm to classify each pixel of the image into one of three classes: leaves, fruits, and background. The first stage is based on the dominance of one of the color channels over the other two in the R G B color model; the green channel is used for the leaves, and the red channel is used for the fruits. The second stage aims to eliminate the false positives generated by the first stage. The segmentation of leaves and fruits of the tomato plant is based on the calculation of four thresholds for the differences of the dominant color channel against the other two, two to label the leaves and the remaining two for the fruits. In the calculation of the thresholds, statistical variables such as the standard deviation and maximum values of the dominant color channels are used.
An image can be mathematically represented as a two-dimensional function f ( x , y ) . If it is handled with the R G B color model, it is made up of three elements: f R ( x , y ) , f G ( x , y ) , and f B ( x , y ) , where the subscript refers to the primary colors red, green, and blue, respectively. x and y represent the spatial coordinates of a particular pixel within the image of dimension M x N , where M represents the number of rows and N is the number of columns in an image.
The first segmentation stage is based on the dominance of the green color channel over the other two channels. This is performed by applying:
h ( x , y ) = f ( x , y ) i f ( f G ( x , y ) f R ( x , y ) ) a n d ( f G ( x , y ) f B ( x , y ) ) 0 i n a n o t h e r c a s e ,
where h ( x , y ) contains the pixels that are filtered from f ( x , y ) with dominance of the green color channel over the other two for x = 0 , 1 , 2 , , M and y = 0 , 1 , 2 , , N .
The other group of pixels of interest to segment are those from the fruits of the tomato plants. The segmentation of the fruits is based on the dominance of the red color channel over the other two; it is performed by applying:
j ( x , y ) = f ( x , y ) i f ( f R ( x , y ) f G ( x , y ) ) a n d ( f R ( x , y ) f B ( x , y ) ) 0 i n a n o t h e r c a s e ,
where j ( x , y ) contains the pixels that are filtered from f ( x , y ) with dominance of the red color channel over the other two x = 0 , 1 , 2 , , M and y = 0 , 1 , 2 , , N .
The second stage of the segmentation process begins with the calculation of the differences between the dominant color channel and the non-dominant channels in h ( x , y ) and j ( x , y ) , which are calculated by applying:
Δ G R ( x , y ) = h G ( x , y ) h R ( x , y ) ,
Δ G B ( x , y ) = h G ( x , y ) h B ( x , y ) ,
Δ R G ( x , y ) = j R ( x , y ) j G ( x , y ) ,
Δ R B ( x , y ) = j R ( x , y ) j B ( x , y ) ,
where Δ G R ( x , y ) and Δ G B ( x , y ) are used to determine the dominance of the green color channel related to the pixels that form the leaves in h ( x , y ) . Δ R G ( x , y ) and Δ R B ( x , y ) are utilized to determine the dominance of the red color channel related to the pixels that form the fruit in j ( x , y ) . The subscript refers to the dominant color channel, whereas the superscript refers to one of the other two color channels.
Then, the thresholds are implemented to determine which pixels belong to leaves and fruits. These are calculated with:
U G R ( x , y ) = h G ( x , y ) M G σ G R α ,
U G B ( x , y ) = h G ( x , y ) M G σ G B α ,
U R G ( x , y ) = j R ( x , y ) M R σ R G α ,
U R B ( x , y ) = j R ( x , y ) M R σ R B α ,
where U G R ( x , y ) and U G B ( x , y ) are the thresholds used to detect leaves, and U R G ( x , y ) and U R B ( x , y ) are utilized to find the fruit region. M G and M R are the highest values of the green and red color channels in h ( x , y ) and j ( x , y ) , respectively. σ G R , σ G B , σ R G , and σ R B correspond to the standard deviations of the Δ G R , Δ G B , Δ R G , and Δ R B values, respectively. Finally, α is a factor utilized to control the thresholds with the objective of maximizing the result of a particular metric by experimenting with different values of α .
The image h F ( x , y ) with the pixels that make up the leaves filtered from h ( x , y ) x = 0 , 1 , 2 , , M and y = 0 , 1 , 2 , , N is obtained with:
h F ( x , y ) = h ( x , y ) i f ( Δ G R ( x , y ) > U G R ( x , y ) ) a n d ( Δ G B ( x , y ) ) > U G B ( x , y ) ) ) 0 i n a n o t h e r c a s e .
The image j F ( x , y ) with the pixels that make up the fruits filtered from j ( x , y ) , x = 0 , 1 , 2 , , M and y = 0 , 1 , 2 , , N is obtained with:
j F ( x , y ) = j ( x , y ) i f ( ( Δ G R ( x , y ) > U G R ( x , y ) ) a n d ( Δ G B ( x , y ) > U G B ( x , y ) ) ) 0 i n a n o t h e r c a s e .
Figure 1 describes the segmentation process of tomato leaves and fruits in a general way.

3. Experimentation

The characteristics of the computer, dataset images, and manual labeling process used to create the masks necessary for the evaluation of performance metrics and presentation of results of the segmentation of leaves and fruits of the tomato plant are described in the following sections.

3.1. Programming Language and Computer Characteristics

Python version 3.9 . 15 was the selected language for the programming of the proposed algorithm due to the number of libraries, of which OpenCV was used for image management.
A computer with the following characteristics was used for the experimental phase of the color dominance segmentation method:
  • Processor: Intel® Core™ i7-8550U CPU @ 1.80 GHz × 8.
  • RAM: 16 GB.
  • Video card: NVIDIA® GeForce® 150MX.
  • Operating system: Ubuntu 22.04.2 LTS 64 bits.

3.2. Dataset

The dataset named “Tomato Detection” consists of 850 images with shots of tomato plantations grown inside greenhouses, which are accessible from the web address https://www.kaggle.com/datasets/andrewmvd/tomato-detection, accessed on 13 February 2023.
In the experimental section, 100 images were used, representing a sample of 12 % of the dataset. The selected images were in the PNG file format and were 500 × 400 pixels in size resolution (see Figure 2).

Labeling of the Dataset

The proposed color dominance segmentation method does not require any prior labeling. To measure the performance of the results generated by segmenting the pixels that form the leaves, fruits, and background of the images, the selected images are labeled.
Image labeling was performed using the Computer Vision Annotation Tool (CVAT), available at the website https://www.cvat.ai/, accessed on 28 February 2023. CVAT is a free web tool with several features that allow different types of labeling for many uses such as recognition, detection, segmentation, and others. In the case of the segmentation of the leaves and fruits of tomato plants, it is necessary to create two labels in a CVAT labeling project task, one for each item of interest, because everything that is not manually labeled is considered background. In this case, the tool Draw new polygon was used and it was established to label where it will belong, as well as a number of points that are established according to the difficulty of the object to be labeled. It is important to mention that the labeling process must be performed on all images used in the experimental section; this requires a large amount of time to perform. Figure 3 shows the CVAT graphical interface in the labeling process of Figure 2.
Figure 4 shows the result of labeling the pixels of the leaves in green, the pixels of fruits in red, and the background in black.

3.3. Performance Metrics

To measure the effectiveness of the proposed segmentation method Accuracy, Precision, Recall, F1-Score [48], and Intersection over Union [49] are used; these are metrics regularly used when measuring the performance of some method or technique in image segmentation. In a binary classification, a pixel can be labeled as either positive or negative, where positive is belonging to a particular class and negative is not belonging to it. The decision is represented in a confusion matrix, which has four elements:
  • True positives ( T P ) are pixels labeled as positive in the real image and in the same way by the segmented method.
  • False positives ( F P ) are the pixels labeled by the segmentation method as positive, when in the real image they are not.
  • True negatives ( T N ) are the pixels marked as negative in real image and in the same way by the segmented method.
  • False negatives ( F N ) are pixels incorrectly labeled as negative by the segmentation method.
The Accuracy metric indicates the total number of pixels correctly classified in relation to the total number of classifications made by the segmentation method. It indicates what percentage of the classifications made by the model are correct and is recommended for use in problems in which the data are balanced. The metric is expressed as:
A c c u r a c y = T P + T N T P + T N + F P + F N .
The Precision metric is the percentage of the number of true positives that are actually positive compared to the total number of predicted positive by the segmentation method. This reflects the degree of proximity of the results of different measurements to each other. The metric is expressed as:
P r e c i s i o n = T P T P + F P .
The Recall metric represents the percentage difference between the number of true positives that the segmentation method has classified and the total number of predicted positive values. This metric is recommended when there exists a high cost associated with false negatives. The metric is expressed as:
R e c a l l = T P T P + F N .
The F1-Score metric represents the harmonic average of the precision and the recall. Its main advantage is that it summarizes both metrics in a single value. The metric is expressed as:
F 1 - S c o r e = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l .
The IoU metric refers to the similarity between the predicted image and the corresponding mask; it is an important metric when comparing the results of the method against manually created marks. The metric is expressed as:
I o U = T P T P + F P + F N .

3.4. Region Segmentation Process for the Leaves and Fruits of the Tomato Plants

The proposed method developed to separate the leaves, fruits, and background in tomato plant crops was applied to the dataset on several occasions, with the objective of maximizing the result of the IoU metric. The results are generated with an α = 3.5 , which maximizes the results for the interest classes (see Figure 5).
Figure 6 shows images resulting from applying the two stages of the segmentation method to five random images from the dataset to segment the pixels that belong to leaves of the tomato plant.
Furthermore, Figure 7 shows images resulting from applying the two stages of the segmentation method to the same five images in Figure 6 to segment the pixels that belong to fruits of the tomato plant.
The results when evaluating the images of Figure 6 and Figure 7 with the selected metrics are shown in Table 1 and Table 2 for leaves, whereas for the fruits, the results are in Table 3 and Table 4.
The average of the five metrics in Table 1 is 75.85 % , whereas for Table 2, the average is 94.30 % . Comparing the averages of the application of Equation (1) with those resulting from Equation (11), a general increase of 18.44 % is observed.
The average of the five metrics in Table 3 is 76.08 % , whereas for Table 4, the average is 91.96 % . Comparing the averages of the application of Equation (2) with those resulting from Equation (12), a general increase of 15.88 % is observed.
Table 5 shows the averages of the performance metrics for the segmentation of the leaves of tomato plants in the 100 images that belong to the test set; the lowest metric is I o U , with 82.62 % , whereas the highest is R e c a l l , with 95.08 % .
Figure 8 shows the distribution of the results of the segmentation process with the metrics for the leaves of the 100 images in the dataset. In all metrics, the mean is less than the median, which indicates that they have a distribution skewed to the left; two of them are in the first quartile, and the remaining three are in the second. Regarding the dispersion of the metric results, the greatest is in I o U followed by P r e c i s i o n , F1-Score, and A c c u r a c y , and the metric with the least dispersion is R e c a l l .
Table 6 shows the averages of the performance metrics in the segmentation of the fruits of tomato plants to the 100 images that belong to the test set. The lowest metric is I o U , with 80.60 % , whereas the highest is A c c u r a c y , with 98.34 % .
Figure 9 shows the distribution of the results of the segmentation process with the metrics for the fruits of the 100 images in the dataset; in all metrics, the mean is less than the median, which indicates that they have a distribution skewed to the left, and all are in the second quartile. Regarding the dispersion of the metric results, the greatest is in I o U followed by P r e c i s i o n , F1-Score, and R e c a l l . The metric with the least dispersion is A c c u r a c y .

4. Comparison of Results against PSPNet Model

A CNN PSPNet [50] model was trained with a ResNet50 [51] backbone to perform semantic segmentation of the leaves and fruits of tomato plants. Figure 10 shows the architecture of the implemented CNN PSPNet model.
The CNN model was created using the T e n s o r F l o w and O p e n C V libraries. For the training process, two sets of images of the initial dataset were created: the T r a i n i n g dataset with 180 items and the V a l i d a t i o n dataset with 80, where it was necessary to perform the same labeling as described in the corresponding section for both sets of images. The learning process was carried out for 70 epochs with a learning rate of 0.01 , lasting eight hours using the equipment described in corresponding section. The accuracy for the training set was about 0.95 , whereas for the validation set it was 0.90 .
Five images were processed with the color dominance segmentation method and compared with the segmentation using the CNN model. The qualitative and quantitative results are presented in Section 3.
A qualitative comparison of the segmentation on tomato leaves and fruits by the CNN model and color dominance method is shown in Figure 11 and Figure 12. The differences in the segmentation results in both cases are observed near the contours of the leaves and fruits.
Table 7 and Table 8 show the quantitative results of the five metrics used to measure the segmentation performance of leaves and fruits of tomato plants. In most cases, the performance of the color dominance segmentation method is superior to that of the CNN model.
The PSPNet model was used to perform the segmentation of 100 images in which the color dominance segmentation method was tested to obtain an overall quantitative comparison of the performance of both methods with the same dataset. The comparative averages of the performance metrics are shown in Table 9.
The averages of the color dominance segmentation method are higher than those of the CNN PSPNet model in all cases, both for leaves and fruits. The metric where the greatest difference is shown is I o U in both cases.
Calculating the average of the results of the five metrics, the color dominance segmentation method has a superior performance of 1.66 percentage points, whereas for fruits it has a performance advantage of 7.48 percentage points when compared to the CNN model.

5. Discussion

The images resulting from applying the two-stage segmentation method are shown in Figure 6 and Figure 7, which show a successful segmentation of the leaves and fruits of tomato plants.
As for the quantitative measurement, the results using the selected performance metrics from Table 5 and Table 6 show adequate performance of the color dominance segmentation method. Another aspect to highlight is that the processed images were taken in real growing environments without lighting control.
When comparing the results of the color dominance segmentation method with the semantic segmentation performed with the CNN PSPNet model in Table 9, the performance of the proposed method is superior, with the great advantage of not requiring a manual image labeling process nor a prior training process costly in time and computational power.
The performance of the color dominance segmentation method can be increased by adjusting the value of α to maximize results or by looking for a particular segmentation objective. For example, value adjustment can be performed by applying heuristic methods, such as simulated annealing or genetic algorithms. Another alternative is to use different values for α , one for the leaves and a different one for the fruits, which allows for an improvement in the results of the segmentation of the leaves and fruits of the tomato plant.
Challenges for the proposed color dominance segmentation method include testing the method on images of tomato crops in the field (outside greenhouses), where brown elements such as soil can affect segmentation performance, particularly in fruit segmentation; adaptation and testing on crops with similar colors, such as strawberries and raspberries; and utilizing other color dominances that occur naturally in other crops and make the necessary adaptations to take advantage of them when segmenting the elements of interest.
An aspect of research involves looking for a color dominance segmentation process that can be used in a different color model, such as H S V in addition to the R G B , which allows the establishment of other conditions based on the dominance of one of the characteristics of the color model used.
A clear disadvantage of the proposed method is that it cannot be applied to crops in which the fruits are green because they would be classified directly as leaves, for example, the cucumber. In these cases, it will be necessary to add a method or algorithm that allows for discrimination of the shape of the fruits from the leaves.
The results of the segmentation performed by the method facilitate the search for pests, diseases, and nutritional deficiencies that may manifest themselves in the leaves, fruits, and background of the segmented images. As part of the future work using the segmentation method implemented, it is intended to develop a system that supports a diagnosis of the state of crops using intelligent algorithms such as CNN, pattern recognition, and heuristic methods for the generation of plant-saving fertilization alternatives.
In addition, it is possible to measure the results of the segmentation performed by the presented algorithm more accurately, thus reducing manual labeling errors.

6. Conclusions

Image segmentation methods are a basic branch in the CV research area; these can be tested in a wide variety of contexts, including agriculture. The segmentation method presented in this paper proposes a color dominance method that takes advantage of the naturally occurring differences in shades between leaves and fruits of some crops. Among its features are its simple operation, fast response, and the fact that it does not require preprocessing of input data, special hardware such as video cards, or prior training to operate.
The method of segmentation by color dominance is based on the use of statistical variables such as standard deviation and the maximum value of the dominant channel to generate local thresholds. These are then used to perform a classification of all pixels that have a dominance of the green color channel for leaves and a dominance of the red color channel for fruits.

Author Contributions

J.P.G.I. and F.J.C.d.l.R. also participated in the conceptualization, methodology, validation and analysis of results. O.A.A. participated along with the other authors in the mathematical foundation. The development of the source code for the tests was done by J.P.G.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Acknowledgments

The authors wanted to thank Consejo Nacional de Humanidades Ciencia y Tecnologías (CONAHCYT), Centro de Investigaciones en Óptica A.C., Instituto Tecnológico de Estudios Superiores de Zamora, Opus Farms for giving us the opportunity to develop this research. To the teachers Luz Basurto from the Instituto Tecnologico de Estudios Superiores de Zamora for her contributions in the agricultural area. Mario Alberto Ruiz and Karla Maria Noriega from Investigaciones en Óptica A.C. for their support in the writing of the document.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Awasthi, Y. Press “a” for artificial intelligence in agriculture: A review. Int. J. Inform. Vis. 2020, 4, 112–116. [Google Scholar] [CrossRef]
  2. Ray, P.P. Internet of things for smart agriculture: Technologies, practices and future direction. J. Ambient Intell. Smart Environ. 2017, 9, 395–420. [Google Scholar] [CrossRef]
  3. FAO. Our approach|Food Systems|Food and Agriculture Organization of the United Nations. 2023. Available online: http://www.fao.org/food-systems/our-approach/en/ (accessed on 10 January 2023).
  4. Gebbers, R.; Adamchuk, V.I. Precision agriculture and food security. Science 2010, 327, 828–831. [Google Scholar] [CrossRef]
  5. Pierce, F.J.; Nowak, P. Aspects of precision agriculture. Adv. Agron. 1999, 67, 1–68. [Google Scholar]
  6. Patrício, D.I.; Rieder, R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef]
  7. Nyalala, I.; Okinda, C.; Nyalala, L.; Makange, N.; Chao, Q.; Chao, L.; Yousaf, K.; Chen, K. Tomato volume and mass estimation using computer vision and machine learning algorithms: Cherry tomato model. J. Food Eng. 2019, 263, 288–298. [Google Scholar] [CrossRef]
  8. Arakeri, M.P.; Lakshmana. Computer Vision Based Fruit Grading System for Quality Evaluation of Tomato in Agriculture industry. Procedia Comput. Sci. 2016, 79, 426–433. [Google Scholar] [CrossRef]
  9. Bhargava, A.; Bansal, A. Fruits and vegetables quality evaluation using computer vision: A review. J. King Saud Univ.-Comput. Inf. Sci. 2021, 33, 243–257. [Google Scholar] [CrossRef]
  10. Nanehkaran, Y.A.; Zhang, D.; Chen, J.; Tian, Y.; Al-Nabhan, N. Recognition of plant leaf diseases based on computer vision. J. Ambient Intell. Humaniz. Comput. 2020, 1–9. [Google Scholar] [CrossRef]
  11. Singh, V.; Misra, A.K. Detection of plant leaf diseases using image segmentation and soft computing techniques. Inf. Process. Agric. 2017, 4, 41–49. [Google Scholar] [CrossRef]
  12. Mukti, I.Z.; Biswas, D. Transfer Learning Based Plant Diseases Detection Using ResNet50. In Proceedings of the 2019 4th International Conference on Electrical Information and Communication Technology, EICT 2019, Khulna, Bangladesh, 20–22 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  13. Smith, R.; Baillie, J.; McCarthy, A.; Raine, S.; Baillie, C. Review of Precision Irrigation Technologies and Their Application; National Centre for Engineering in Agriculture Publication 1003017/1; USQ: Toowoomba, Australia, 2010; Volume 1. [Google Scholar]
  14. Kaur, G.; Engineering, C. Automated Nutrient Deficiency Detection in Plants: A Review. Palarch’s J. Archaeol. Egypt 2020, 17, 5894–5901. [Google Scholar]
  15. Hamuda, E.; Glavin, M.; Jones, E. A survey of image processing techniques for plant extraction and segmentation in the field. Comput. Electron. Agric. 2016, 125, 184–199. [Google Scholar] [CrossRef]
  16. Rasmussen, J.; Norremark, M.; Bibby, B.M. Assessment of leaf cover and crop soil cover in weed harrowing research using digital images. Weed Res. 2007, 47, 299–310. [Google Scholar] [CrossRef]
  17. Kirk, K.; Andersen, H.J.; Thomsen, A.G.; Jørgensen, J.R.; Jørgensen, R.N. Estimation of leaf area index in cereal crops using red-green images. Biosyst. Eng. 2009, 104, 308–317. [Google Scholar] [CrossRef]
  18. Story, D.; Kacira, M.; Kubota, C.; Akoglu, A.; An, L. Lettuce calcium deficiency detection with machine vision computed plant features in controlled environments. Comput. Electron. Agric. 2010, 74, 238–243. [Google Scholar] [CrossRef]
  19. Wang, Y.; Wang, D.; Zhang, G.; Wang, J. Estimating nitrogen status of rice using the image segmentation of G-R thresholding method. Field Crop. Res. 2013, 149, 33–39. [Google Scholar] [CrossRef]
  20. Quemada, M.; Gabriel, J.L.; Zarco-Tejada, P. Airborne hyperspectral images and ground-level optical sensors as assessment tools for maize nitrogen fertilization. Remote Sens. 2014, 6, 2940. [Google Scholar] [CrossRef]
  21. Jeon, H.Y.; Tian, L.F.; Zhu, H. Robust crop and weed segmentation under uncontrolled outdoor illumination. Sensors 2011, 11, 6270–6283. [Google Scholar] [CrossRef]
  22. Yadav, S.P.; Ibaraki, Y.; Gupta, S.D. Estimation of the chlorophyll content of micropropagated potato plants using RGB based image analysis. Plant Cell Tissue Organ Cult. 2010, 100, 183–188. [Google Scholar] [CrossRef]
  23. Philipp, I.; Rath, T. Improving plant discrimination in image processing by use of different colour space transformations. Comput. Electron. Agric. 2002, 35, 1–15. [Google Scholar] [CrossRef]
  24. Menesatti, P.; Antonucci, F.; Pallottino, F.; Roccuzzo, G.; Allegra, M.; Stagno, F.; Intrigliolo, F. Estimation of plant nutritional status by Vis-NIR spectrophotometric analysis on orange leaves [Citrus sinensis (L) Osbeck cv Tarocco]. Biosyst. Eng. 2010, 105, 448–454. [Google Scholar] [CrossRef]
  25. Fan, P.; Lang, G.; Yan, B.; Lei, X.; Guo, P.; Liu, Z.; Yang, F. A method of segmenting apples based on gray-centered rgb color space. Remote Sens. 2021, 13, 1211. [Google Scholar] [CrossRef]
  26. Xu, G.; Zhang, F.; Ghafoor Shah, S.; Ye, Y.; Mao, H.; Shah, S.G.; Ye, Y.; Mao, H. Use of leaf color images to identify nitrogen and potassium deficient tomatoes. Pattern Recognit. Lett. 2011, 32, 1584–1590. [Google Scholar] [CrossRef]
  27. Wan, P.; Toudeshki, A.; Tan, H.; Ehsani, R. A methodology for fresh tomato maturity detection using computer vision. Comput. Electron. Agric. 2018, 146, 43–50. [Google Scholar] [CrossRef]
  28. Tian, K.; Li, J.; Zeng, J.; Evans, A.; Zhang, L. Segmentation of tomato leaf images based on adaptive clustering number of K-means algorithm. Comput. Electron. Agric. 2019, 165, 104962. [Google Scholar] [CrossRef]
  29. Castillo-Martínez, M.; Gallegos-Funes, F.J.; Carvajal-Gámez, B.E.; Urriolagoitia-Sosa, G.; Rosales-Silva, A.J. Color index based thresholding method for background and foreground segmentation of plant images. Comput. Electron. Agric. 2020, 178, 105783. [Google Scholar] [CrossRef]
  30. Lin, G.; Tang, Y.; Zou, X.; Xiong, J.; Fang, Y. Color-, depth-, and shape-based 3D fruit detection. Precis. Agric. 2020, 21, 1–17. [Google Scholar] [CrossRef]
  31. Lu, Y.; Young, S.; Wang, H.; Wijewardane, N. Robust plant segmentation of color images based on image contrast optimization. Comput. Electron. Agric. 2022, 193, 106711. [Google Scholar] [CrossRef]
  32. Kamilaris, A.; Prenafeta-Boldú, F.X. Deep learning in agriculture: A survey. Comput. Electron. Agric. 2018, 147, 70–90. [Google Scholar] [CrossRef]
  33. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2323. [Google Scholar] [CrossRef]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  35. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  36. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
  37. Fuentes, A.; Yoon, S.; Kim, S.C.; Park, D.S. A robust deep-learning-based detector for real-time tomato plant diseases and pests recognition. Sensors 2017, 17, 2022. [Google Scholar] [CrossRef] [PubMed]
  38. Sladojevic, S.; Arsenovic, M.; Anderla, A.; Culibrk, D.; Stefanovic, D. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification. Comput. Intell. Neurosci. 2016, 2016, 3289801. [Google Scholar] [CrossRef] [PubMed]
  39. Mohanty, S.P.; Hughes, D.; Salathe, M. Using Deep Learning for Image-Based Plant Disease Detection. Front. Plant Sci. 2016, 7, 1419. [Google Scholar] [CrossRef] [PubMed]
  40. Hall, D.; McCool, C.; Dayoub, F.; Sünderhauf, N.; Upcroft, B. Evaluation of features for leaf classification in challenging conditions. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, WACV 2015, Waikoloa, HI, USA, 5–9 January 2015; pp. 797–804. [Google Scholar] [CrossRef]
  41. Roy, K.; Chaudhuri, S.S.; Pramanik, S.; Ponce, H.; Cevallos, C.; Espinosa, R.; Gutierrez, S. Estimation of Low Nutrients in Tomato Crops Through the Analysis of Leaf Images Using Machine Learning. J. Artif. Intell. Technol. 2021, 27, 131–137. [Google Scholar] [CrossRef]
  42. Anand, T.; Sinha, S.; Mandal, M.; Chamola, V.; Yu, F.R.; Milioto, A.; Lottes, P.; Stachniss, C.; Luo, Z.; Yang, W.; et al. AgriSegNet: Deep Aerial Semantic Segmentation Framework for IoT-Assisted Precision Agriculture. IEEE Sens. J. 2023, 21, 17581–17590. [Google Scholar] [CrossRef]
  43. Majeed, Y.; Karkee, M.; Zhang, Q.; Fu, L.; Whiting, M.D. Determining grapevine cordon shape for automated green shoot thinning using semantic segmentation-based deep learning networks. Comput. Electron. Agric. 2020, 171, 105308. [Google Scholar] [CrossRef]
  44. Kang, H.; Chen, C. Fruit detection and segmentation for apple harvesting using visual sensor in orchards. Sensors 2019, 19, 4599. [Google Scholar] [CrossRef]
  45. Gonzalez, R.; Richard, W. Digital Image Processing, 4th ed.; Pearson: London, UK, 1980. [Google Scholar]
  46. Jaiswal, S.; Pandey, M.K. A Review on Image Segmentation. Adv. Intell. Syst. Comput. 2021, 1187, 233–240. [Google Scholar] [CrossRef]
  47. Mardanisamani, S.; Eramian, M. Segmentation of vegetation and microplots in aerial agriculture images: A survey. Plant Phenome J. 2022, 5, e20042. [Google Scholar] [CrossRef]
  48. Taheri, M.; Lim, N.; Lederer, J. Balancing Statistical and Computational Precision and Applications to Penalized Linear Regression with Group Sparsity; Department of Computer Sciences and Department of Biostatistics and Medical Informatics: Winsconsin, WI, USA, 2016; pp. 233–240. [Google Scholar]
  49. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar] [CrossRef]
  50. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar] [CrossRef]
  51. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
Figure 1. Implementation methodology of the proposed segmentation algorithm by means of thresholds.
Figure 1. Implementation methodology of the proposed segmentation algorithm by means of thresholds.
Agriengineering 05 00113 g001
Figure 2. Example image from the tomato plantation dataset.
Figure 2. Example image from the tomato plantation dataset.
Agriengineering 05 00113 g002
Figure 3. CVAT labeling interface.
Figure 3. CVAT labeling interface.
Agriengineering 05 00113 g003
Figure 4. Labeling of Figure 2, with the leaves in green, fruits in red and background in black.
Figure 4. Labeling of Figure 2, with the leaves in green, fruits in red and background in black.
Agriengineering 05 00113 g004
Figure 5. Results of calibration sensitivity test for α .
Figure 5. Results of calibration sensitivity test for α .
Agriengineering 05 00113 g005
Figure 6. Images resulting from the segmentation process for the leaves. (a) Original image, (b) image mask, (c) result of applying Equation (1) to f ( x , y ) , (d) result of applying Equation (11) to h ( x , y ) .
Figure 6. Images resulting from the segmentation process for the leaves. (a) Original image, (b) image mask, (c) result of applying Equation (1) to f ( x , y ) , (d) result of applying Equation (11) to h ( x , y ) .
Agriengineering 05 00113 g006
Figure 7. Images resulting from the segmentation process for the fruits. (a) Original image, (b) image mask (c), result of applying Equation (2) to f ( x , y ) , (d) result of applying Equation (12) to j ( x , y ) .
Figure 7. Images resulting from the segmentation process for the fruits. (a) Original image, (b) image mask (c), result of applying Equation (2) to f ( x , y ) , (d) result of applying Equation (12) to j ( x , y ) .
Agriengineering 05 00113 g007
Figure 8. Box plot of the metrics for the segmentation of leaves.
Figure 8. Box plot of the metrics for the segmentation of leaves.
Agriengineering 05 00113 g008
Figure 9. Box plot of the metrics for the segmentation for fruits.
Figure 9. Box plot of the metrics for the segmentation for fruits.
Agriengineering 05 00113 g009
Figure 10. Architecture of the PSPNet model. (a) Input image, (b) backbone, (c) pyramid pooling module, (d) output image.
Figure 10. Architecture of the PSPNet model. (a) Input image, (b) backbone, (c) pyramid pooling module, (d) output image.
Agriengineering 05 00113 g010
Figure 11. Comparison of segmentation results for the leaves. (a) Original image, (b) image mask (c) segmentation by CNN PSPNet, (d) segmentation by color dominance.
Figure 11. Comparison of segmentation results for the leaves. (a) Original image, (b) image mask (c) segmentation by CNN PSPNet, (d) segmentation by color dominance.
Agriengineering 05 00113 g011
Figure 12. Comparison of segmentation results for the fruits. (a) Original image, (b) image mask, (c) segmentation by CNN PSPNet, (d) segmentation by color dominance.
Figure 12. Comparison of segmentation results for the fruits. (a) Original image, (b) image mask, (c) segmentation by CNN PSPNet, (d) segmentation by color dominance.
Agriengineering 05 00113 g012
Table 1. Performance metrics for the segmentation of the leaves of tomato plants with the five images of Figure 6c applying Equation (1).
Table 1. Performance metrics for the segmentation of the leaves of tomato plants with the five images of Figure 6c applying Equation (1).
Result Metrics of First Segmentation of Leaves
ImageAccuracyPrecisionRecallF1-ScoreIoU
161.90%41.95%99.98%59.11%41.95%
268.03%54.09%98.81%69.91%57.74%
381.92%79.48%99.99%88.56%79.48%
483.43%82.91%99.71%90.54%82.71%
567.15%64.44%99.89%78.34%64.39%
Averages72.48%64.57%99.67%77.29%62.25%
Table 2. Performance metrics for the segmentation of the leaves of tomato plants with the five images in Figure 6d applying Equation (11).
Table 2. Performance metrics for the segmentation of the leaves of tomato plants with the five images in Figure 6d applying Equation (11).
Result Metrics of Second Segmentation of Leaves
ImageAccuracyPrecisionRecallF1-ScoreIoU
195.96%93.25%91.99%92.61%86.24%
295.11%97.06%89.70%93.23%87.33%
395.19%95.08%98.21%96.62%93.46%
497.20%97.48%99.03%98.25%96.56%
593.52%91.45%98.30%94.75%90.02%
Averages95.39%94.86%95.44%95.09%90.72%
Table 3. Performance metrics for the segmentation of the fruits of tomato plants with the five images of Figure 7c applying Equation (2).
Table 3. Performance metrics for the segmentation of the fruits of tomato plants with the five images of Figure 7c applying Equation (2).
Result Metrics of First Segmentation of Fruits
ImageAccuracyPrecisionRecallF1-ScoreIoU
177.72%28.92%99.47%44.82%28.88%
297.83%85.15%99.93%91.95%85.10%
392.28%30.95%99.95%47.26%30.94%
498.77%69.53%99.29%81.79%69.19%
598.80%78.08%99.87%87.64%78.00%
Averages93.08%58.52%99.70%70.69%58.42%
Table 4. Performance metrics for the segmentation of the fruits of tomato plants with the five images of Figure 7d applying Equation (12).
Table 4. Performance metrics for the segmentation of the fruits of tomato plants with the five images of Figure 7d applying Equation (12).
Result Metrics of Second Segmentation of Fruits
ImageAccuracyPrecisionRecallF1-ScoreIoU
197.62%80.54%97.45%88.19%78.84%
298.89%92.26%99.40%95.68%91.72%
398.78%74.26%99.40%85.01%72.93%
499.71%95.24%94.22%94.72%89.98%
599.55%93.05%96.70%94.84%90.19%
Averages98.91%87.07%97.43%91.68%84.73%
Table 5. Averages of performance metrics for the segmentation methods over the 100 images of test set.
Table 5. Averages of performance metrics for the segmentation methods over the 100 images of test set.
Averages of Performances for Leaves
AccuracyPrecisionRecallF1-ScoreIoU
93.70%86.13%95.08%88.31%82.62%
Table 6. Averages of performance metrics for the segmentation methods over the 100 images of the test set.
Table 6. Averages of performance metrics for the segmentation methods over the 100 images of the test set.
Averages of Performance Metrics for Fruits
AccuracyPrecisionRecallF1-ScoreIoU
98.34%84.36%94.51%88.32%80.60%
Table 7. Comparison of five leaf image segmentation results.
Table 7. Comparison of five leaf image segmentation results.
Comparison of Leaf Segmentation Results
ImageBy Color DominancePSPNet
AccuracyPrecisionRecallF1-ScoreIoUAccuracyPrecisionRecallF1-ScoreIoU
195.96%93.25%91.99%92.61%86.24%95.81%90.88%94.26%92.54%86.12%
295.11%97.06%89.70%93.23%87.33%91.12%90.65%85.17%87.82%78.29%
395.19%95.08%98.21%96.62%93.46%93.24%92.41%98.44%95.33%91.07%
497.20%97.48%99.03%98.25%96.56%94.84%94.76%98.98%96.83%93.85%
593.52%91.45%98.30%94.75%90.02%88.07%74.22%93.81%82.87%70.76%
Table 8. Comparison of five fruit image segmentation results.
Table 8. Comparison of five fruit image segmentation results.
Comparison of Fruit Segmentation Results
ImageBy Color DominancePSPNet
AccuracyPrecisionRecallF1-ScoreIoUAccuracyPrecisionRecallF1-ScoreIoU
197.62%80.54%97.45%88.19%78.84%96.13%75.00%86.15%80.19%66.93%
298.89%92.26%99.40%95.68%91.72%98.64%94.24%94.82%94.53%89.63%
398.78%74.26%99.40%85.01%72.93%97.64%61.32%86.05%97.61%55.78%
499.71%95.24%94.22%94.72%89.98%99.27%93.86%78.87%85.71%75.00%
599.55%93.05%96.70%94.84%90.19%98.52%93.23%96.93%95.04%90.56%
Table 9. Comparisons of leaf and fruit segmentation averages in tomato plants.
Table 9. Comparisons of leaf and fruit segmentation averages in tomato plants.
Comparisons of Leaf and Fruit Segmentation Averages in Tomato Plants
LeavesFruits
AccuracyPrecisionRecallF1-ScoreIoUAccuracyPrecisionRecallF1-ScoreIoU
Color dominance93.70%86.13%95.08%88.31%82.62%98.34%84.36%94.51%88.32%80.60%
PSPNet91.46%85.09%93.19%87.59%80.19%95.95%79.21%82.84%80.13%68.56%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guerra Ibarra, J.P.; Cuevas de la Rosa, F.J.; Arellano Arzola, O. Segmentation of Leaves and Fruits of Tomato Plants by Color Dominance. AgriEngineering 2023, 5, 1846-1864. https://doi.org/10.3390/agriengineering5040113

AMA Style

Guerra Ibarra JP, Cuevas de la Rosa FJ, Arellano Arzola O. Segmentation of Leaves and Fruits of Tomato Plants by Color Dominance. AgriEngineering. 2023; 5(4):1846-1864. https://doi.org/10.3390/agriengineering5040113

Chicago/Turabian Style

Guerra Ibarra, Juan Pablo, Francisco Javier Cuevas de la Rosa, and Oziel Arellano Arzola. 2023. "Segmentation of Leaves and Fruits of Tomato Plants by Color Dominance" AgriEngineering 5, no. 4: 1846-1864. https://doi.org/10.3390/agriengineering5040113

APA Style

Guerra Ibarra, J. P., Cuevas de la Rosa, F. J., & Arellano Arzola, O. (2023). Segmentation of Leaves and Fruits of Tomato Plants by Color Dominance. AgriEngineering, 5(4), 1846-1864. https://doi.org/10.3390/agriengineering5040113

Article Metrics

Back to TopTop