Next Article in Journal
Application of the Reuterin System as Food Preservative or Health-Promoting Agent: A Critical Review
Next Article in Special Issue
Visual Detection of Water Content Range of Seabuckthorn Fruit Based on Transfer Deep Learning
Previous Article in Journal
Interaction between Kidney-Bean Polysaccharides and Duck Myofibrillar Protein as Affected by Ultrasonication: Effects on Gel Properties and Structure
Previous Article in Special Issue
Computer Vision System for Mango Fruit Defect Detection Using Deep Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DeepMDSCBA: An Improved Semantic Segmentation Model Based on DeepLabV3+ for Apple Images

1
College of Mathematics and Computer Science, Zhejiang A&F University, Hangzhou 311300, China
2
School of Information Engineering, Huzhou University, Huzhou 313000, China
*
Author to whom correspondence should be addressed.
Foods 2022, 11(24), 3999; https://doi.org/10.3390/foods11243999
Submission received: 8 November 2022 / Revised: 4 December 2022 / Accepted: 7 December 2022 / Published: 10 December 2022
(This article belongs to the Special Issue Vision-Based Sensors and Algorithms for Food Processing)

Abstract

:
The semantic segmentation of apples from images plays an important role in the automation of the apple industry. However, existing semantic segmentation methods such as FCN and UNet have the disadvantages of a low speed and accuracy for the segmentation of apple images with complex backgrounds or rotten parts. In view of these problems, a network segmentation model based on deep learning, DeepMDSCBA, is proposed in this paper. The model is based on the DeepLabV3+ structure, and a lightweight MobileNet module is used in the encoder for the extraction of features, which can reduce the amount of parameter calculations and the memory requirements. Instead of ordinary convolution, depthwise separable convolution is used in DeepMDSCBA to reduce the number of parameters to improve the calculation speed. In the feature extraction module and the cavity space pyramid pooling module of DeepMDSCBA, a Convolutional Block Attention module is added to filter background information in order to reduce the loss of the edge detail information of apples in images, improve the accuracy of feature extraction, and effectively reduce the loss of feature details and deep information. This paper also explored the effects of rot degree, rot position, apple variety, and background complexity on the semantic segmentation performance of apple images, and then it verified the robustness of the method. The experimental results showed that the PA of this model could reach 95.3% and the MIoU could reach 87.1%, which were improved by 3.4% and 3.1% compared with DeepLabV3+, respectively, and superior to those of other semantic segmentation networks such as UNet and PSPNet. In addition, the DeepMDSCBA model proposed in this paper was shown to have a better performance than the other considered methods under different factors such as the degree or position of rotten parts, apple varieties, and complex backgrounds.

1. Introduction

Computer vision is widely used in digital agriculture [1]. As a basic task in computer vision, semantic segmentation is applied in all aspects of agricultural automation. Semantic segmentation technology also provides strong technical support for the work of precision agricultural robots. The semantic segmentation of crop fruits can help robots detect and locate fruit positions and realize automatic picking, thereby reducing manual participation in agriculture, improving agricultural efficiency, and reducing production costs.
A widely used semantic segmentation network is the Fully Convolutional Network (FCN) proposed by Long et al. [2]. Wang et al. used the FCN to realize the recognition of wheat ear targets that is difficult to achieve with traditional methods [3]. The Pyramid Scene Parsing Network (PSPNet) [4] has made some improvements based on the FCN, adding an encoder–decoder structure that allows for more refined deconvolution results, improved classification accuracy, and improved overall efficiency [5]. Deng et al. [6] used PSPNet to segment a range of kiwifruit vines based on kiwifruit orchard images. The DeepLab semantic segmentation method uses Atrous Spatial Pyramid Pooling (ASPP), which can expand the receptive field without changing the resolution; at the same time, the features of different levels can be fused [7]. DeepLabV3+ as found to have better segmentation results than former segmentation methods such as FCN and PSPNet. Zhang et al. used DeepLabV3+ to segment the lodging area of different wheat growth stages, and their segmentation accuracy was high [8]. Using DeepLabV3+, Sharifzadeh et al. detected farm crops with low-resolution satellite images, and their edge segmentation effect was better [9].
However, the DeepLabV3+ method also has some shortcomings. Firstly, its computational complexity is relatively high because its feature extraction network Xception has a large number of network layers and a large number of parameters and the convolution method in the ASPP module is ordinary convolution, which increases the number of parameters. Secondly, its feature information extraction can be improved. In the process of feature extraction at the encoder, the spatial dimension of the input data is gradually reduced, resulting in the loss of useful information, and detail recovery cannot be achieved well during decoding. Finally, the target edge recognition accuracy is relatively low. Although the ASPP module can improve the method’s ability to extract the boundary of a target, it cannot fully simulate the relationship between the local features of a target, resulting in a reduction in the accuracy of the target segmentation and subsequent problems such as a low recognition accuracy and poor edge recognition.
In view of the above issues, in order to obtain more efficiently and accurately achieve apple fruit segmentation, DeepMDSCBA (a network segmentation model based on DeepLabV3+ with the MobileNet, Depthwise Separable Convolution, and Convolutional Block Attention modules) is proposed in this paper. DeepMDSCBA was constructed based on the DeepLabV3+ structure, and a more lightweight MobileNet network [10] was adopted as its backbone network other than the original Xception network to reduce the amount of parameter calculations and memory usage. In DeepMDSCBA, Depthwise Separable Convolution (DSC) [11] is used to replace the ordinary convolution in the ASPP module to improve the calculation speed of the method. In the feature extraction and ASPP modules of DeepMDSCBA, a Convolutional Block Attention Module (CBAM) [12] is added to filter the background information and reduce the loss of image edge detail information. Because of the improvements in the above aspects, the information processing efficiency and accuracy of DeepMDSCBA were found to be improved and the accuracy of the segmentation model was increased.
In addition, to verify the robustness of DeepMDSCBA, the influence of rot degree, rot position, apple variety, and complexity of background on the performance of apple image semantic segmentation were extensively studied in this paper.
The rest of this article is organized as follows. Section 2 describes the main idea of the DeepMDSCBA method. Section 3 describes the design and setup of our experiments. Section 4 analyzes the results of the comparative experiments. Finally, the last section summarizes the work.

2. DeepMDSCBA Segmentation Method

2.1. Main Ideas

The proposed DeepMDSCBA method, a network segmentation method based on deep learning, is demonstrated in Figure 1.
According to Figure 1, the main ideas of DeepMDSCBA are as follows:
(1)
Using MobileNet to replace Xception: Based on the DeepLabV3+ framework, the backbone Xception network for method feature extraction was replaced with a more lightweight MobileNet to greatly reduce the amount of parameter calculations, reduce memory usage, and improve the calculation speed of the method.
(2)
Using DSC to replace ordinary convolution: The ordinary convolution in the ASPP module was changed to DSC to further reduce the method’s parameters and improve the method’s calculation speed.
(3)
Adding CBAM: CBAM was added to the feature extraction module and the ASPP module to reduce accuracy loss in order to train a more accurate segmentation method and improve the segmentation accuracy of the proposed model for apple images.

2.2. Using MobileNet to Replace Xception

DeepLabV3+ is an improved version of DeepLabV3. It replaces the underlying network with a residual network and adds an encoding–decoding structure to restore spatial information to obtain clear object boundaries so that boundary segmentation is optimized [13,14]. At the encoder end, DeepLabV3+ uses the Xception network to extract the features of the input image, and it then uses the ASPP module to fuse image features to avoid information loss. Xception is a deep convolutional neural network with an input stream, an intermediate stream, and an output stream, and ASPP is a multi-scale pyramid feature extraction module. At the decoder end, the low-level features from the Xception module and the high-level features from the encoder are fused, and then bilinear interpolation upsampling is executed to output the segmentation results, which finally improves the accuracy of network segmentation.
In DeepMDSCBA, the method proposed in this paper, MobileNet (a more lightweight network) was adopted to replace Xception, the original backbone of DeepLabV3+. Compared with the Xception network, MobileNet has a shallower network layer, fewer parameters, a lower method complexity, a faster network training speed, and a faster convergence. This replacement solves the problems of large convolutional neural networks and insufficient hardware training in the process of image method training. In the version of the implementation of DeepMDSCBA in this paper, the second version of MobileNet, MobileNetV2, is used. This version mainly consists of three parts. The first part is a 3 × 3 convolution block for feature extraction, the second part is a network structure containing multiple DSC layers composed of multiple 1 × 1 and 3 × 3 convolution blocks, and the last part consists of two 1 × 1 convolution blocks and a 7 × 7 average pooling block [15]. Combined with the DeepLabV3+ network structure, DeepMDSCBA uses the front two parts of the network, and the specific network structure is shown in Table 1.

2.3. Using DSC to Replace Ordinary Convolution

The main function of a convolution layer is feature extraction. In the process of the convolution of the input feature map through the convolution kernel, the spatial features and channel features need to be learned at the same time while the deep separable convolution decouples the spatial correlation and channel correlation of the convolution layer, adds a transition layer in the standard convolution process, and decomposes it into depthwise convolution [16] and pointwise convolution [17] considering spatial correlation and channel correlation, respectively. Compared with standard convolution, DSC can greatly reduce the number of parameters and calculations while ensuring little loss of accuracy.
Suppose the input is a feature map of D F × D F × M and it is convoluted with a convolution kernel of D K × D K × M × N in size; the standard convolution operation process is shown in Figure 2a. Each input feature map is convoluted with M ×   D K × D K convolution kernels in the N class and then summed and biased to obtain an output. The final output is D F × D F × N in size. The depthwise separable operation process is shown in the following figure. Figure 2b shows the depthwise convolution process, and Figure 2c shows the pointwise convolution process.
In depthwise convolution, each input feature map is only biased by convolution with the corresponding convolution kernel, and the output size is D F × D F × M . In point-by-point convolution, the feature map of D F × D F × M is standardly convoluted with N   × 1 × 1 convolution kernels, and the number of channels is changed. Finally, the output feature map size is D F × D F × N .
The standard convolution process calculation is as follows.
D F × D F × A = π r 2 × N × D K × D K
In depthwise separable convolution, the calculation amount of the depthwise convolution process is as follows.
D F × D F × M × D K × D K
The calculation amount of pointwise convolution process is as follows.
D F × D F × M × N
The depthwise separable convolution is equivalent to n times the standard convolution. The expression of n is as follows.
D F × D F × M × D K × D K + D F × D F × M × N D F × D F × M × N × D K × D K = 1 N + 1 D K × D K
In above formulas, D F × D F is the size pixel of input feature map, M is the number of input channels, D K × D K is the size pixel of convolution kernel, and N is the number of output channels.

2.4. Adding CBAM

Attention plays an important role in human perception. Humans use a series of local observations and selectively focus on salient parts to better capture visual structures. In recent years, many researchers have improved the performance of convolutional neural networks in large-scale classification tasks by adding attention mechanisms. In this paper, CBAM was added to the feature extraction and ASPP modules. Firstly, the global dependencies between features are captured in the spatial and channel dimensions of features to capture the context feature information and enhance the expression ability of features, respectively. Then, the outputs of the two attention modules are added to further improve the feature representation; finally, more accurate segmentation results can be obtained [18]. The structure of CBAM is shown in Figure 3.
CBAM is a lightweight convolutional attention module that combines the channel attention mechanism and the spatial attention mechanism.
The channel attention mechanism selectively emphasizes interconnected channel maps by integrating relevant features in all channel maps [19]. Therefore, DeepMDSCBA explicitly models the interdependence between channels by adding a channel attention mechanism module. Assume an input feature map F R C × W × H (where C denotes the number of channels of the input feature map and W and H denote the width and height of the feature map, respectively). Firstly, the parallel global maximum pooling and global average pooling are used to compress the spatial dimension of the input feature map F to obtain the background description F m a x c and F a v g c , respectively. Then, the sigmoid function is added by the shared network calculation composed of multi-layer perceptron M L P to obtain the channel attention mechanism mapping feature map M C R C × 1 × 1 . Finally, the generated feature map is mapped to C × W × H , and the original input feature map is multiplied by elements to obtain the channel attention-weighted graph F C o u t R C × W × H . The specific calculation process is as follows:
M C = σ M L P F max c + M L P F a v g c
F C o u t = F F c M c
where σ is the sigmoid activation function, ⊗ is the multiplication between elements, and F c is the feature vector of C × W × H obtained by copying M c along the spatial dimension.
The spatial attention mechanism selectively aggregates the features of each location through the weighted sum of the features of all locations. Similar features are interrelated regardless of distance [20,21,22]. Therefore, in order to establish a richer context relationship between local features, the spatial attention mechanism module is introduced. Assume an input feature map F R C × W × H , where C denotes the number of channels of the input feature map and W and H denote the width and height of the feature map, respectively. Firstly, the background descriptions F m a x c and F a v g c are obtained via the channel compression of the input feature map F using parallel global maximum pooling and global average pooling, respectively. The two feature maps are merged and the dimension is reduced to 1 channel by a 7 × 7 convolution operation. Then, the mapping feature map M s R 1 × W × H of the spatial attention mechanism is obtained with the sigmoid function. Finally, the generated feature map is mapped to C × W × H , and the original input feature map is multiplied by elements to obtain a spatial attention-weighted graph F S o u t R C × W × H . The specific calculation process is as follows:
M s = σ f 7 × 7 F m a x s ; F a v g s
F S o u t = F F s M s
where σ represents the sigmoid activation function, f 7 × 7 represents the convolution operation, the convolution kernel size is 7 × 7 , ⊗ represents the multiplication between elements, and F s represents the C × W × H feature vector obtained by copying M s along the channel direction.

3. Experiments

This section introduces the design of experiments to test the performance of DeepMDSCBA, the method proposed in this paper. First, the hardware and software for the equipment configuration of the experiments is introduced, then the dataset production and preprocessing required for the experiments are described, and the network hyperparameters are established. The designs of the robust and ablation experiments are introduced last.

3.1. Hardware and Software Configuration

The experiments in this paper used the PyTorch deep learning framework to train and test the performance of the DeepMDSCBA method. The specific configuration of the experiments is shown in Table 2.

3.2. Data Acquisition and Preprocessing

The images in the datasets used for the experiments were mainly obtained from the internet, together with image synthesis, which was divided into training and test sets. The training and test sets had no duplicate images. The training and test sets were classified according to different apple varieties, different degrees of rot, different positions of rot, and background complexity. After enhancing the training set through the addition of random rotation, noise and mirroring and increasing the sample size to improve the model’s generalization ability, we trained the model. Furthermore, the apple images in the datasets were labeled with the graphical interface labeling software LabelMe [23] to generate JSON files. Although DeepLabV3+ does not limit image parameters such as the resolution of the images in the dataset, images in the datasets were uniformly converted into grayscale images with a resolution of 512 × 512 and a depth of 24 and then were stored in the PASCAL VOC [24] data format, which allowed the comparison of the performance of DeepMDSCBA and that of various other methods.

3.3. Dataset Partitioning

3.3.1. Dataset of Apple Images with Different Rot Degrees

This dataset consists of apple images with different rot degrees, where the degree is expressed as the proportion of rotten area. All images in the dataset are apples in the front view. Due to the small number of naturally rotten apple images and the difficulty in controlling the degree and position of rot, in addition to the naturally rotten apple images, image synthesis technology was used to synthesize the rotten parts of apple images so that apple images with different rotten areas could be obtained for experiments. According to the ratio of the rotten area in the image to the entire apple area, the proportions of the rotten area to the apple image in the dataset were divided into five sub-datasets of (0, 20%], (20%, 40%], (40%, 60%], (60%, 80%], and (80%, 100%]. Some samples are shown in Figure 4.

3.3.2. Dataset of Apple Images with Different Rot Positions

This dataset consists of three different views of apples. The views include the fruit stalk, fruit calyx, and neither fruit stalk nor fruit calyx. In order to unify the standard, an apple image with the proportion of rotten area between (0, 40%] was selected, and the dataset was divided into sub-datasets of the three abovementioned views to explore the influence of different rot positions on the apple semantic segmentation results. Some samples are shown in Figure 5.

3.3.3. Dataset of Apple Images of Different Varieties

This dataset consists of images of different varieties of apples. Images of four common apple varieties—Golden Delicious, Fuji, Honey Crisp and Red Delicious—were selected for the experiments to explore the influence of different apple varieties on the segmentation results. Some samples are shown in Figure 6.

3.3.4. Dataset of Apple Images with Complex Backgrounds

This dataset is mainly composed of multiple apple images in the natural state. The images are apples under relatively complex backgrounds with branches and leaves, which were used to explore the effect of complex backgrounds on the semantic segmentation of apple images using DeepMDSCBA in this study. Some samples of the dataset are shown in Figure 7.

3.4. Evaluation Indicators

In the experiments, MIoU (Mean Intersection over Union) and PA (Pixel Accuracy) were used as the evaluation indicators for apple image segmentation to analyze the segmentation performance.
(1)
Pixel Accuracy (PA)
PA is the ratio of correctly predicted pixels to total pixels. The calculation formula is as follows:
PA = i = 0 k P i i i = 0 k j = 0 k P i j
In the formula, k denotes the total number of categories, P i j denotes the number of pixels that belong to class i but are predicted to belong to class j ,   P i i denotes the number of correctly predicted pixels, and P i j and P j i denote false positive and false negative results, respectively.
(2)
Mean Intersection over Union (MIoU)
MIoU is the most commonly used metric in semantic segmentation experiments. It is used to calculate the ratio of the intersection and union of two sets of real and predicted values on each class and then to calculate the average value of the intersection and union ratio of all classes, that is, the average intersection and union ratio. The calculation formula is as follows:
MIoU = 1 k + 1 i = 0 k P i i j = 0 k P i j + j = 0 k P j i P i i
In the formula, k denotes the total number of categories, P i j denotes the number of pixels that belong to class i but are predicted to belong to class j ,   P i i denotes the number of correctly predicted pixels, and P i j and P j i denote false positive and false negative results, respectively.

3.5. Experimental Scheme

3.5.1. Determination of Training Parameters

For the original DeepLabV3+ method, with an initial learning rate of 0.007 and a batch size of 16, the average intersection–union ratios of the method on the PASCAL VOC2012 and Cityscapes [25] datasets were 89.1% and 83.2%, respectively, achieving good segmentation results. On this basis, according to the commonly used empirical values of network training hyperparameters, and after repeated testing, the network hyperparameters of DeepMDSCBA used in the experiments were established. They are shown in Table 3.

3.5.2. Test Scheme

In order to test the performance of the proposed DeepMDSCBA method in an apple image segmentation task, it was compared with the traditional semantic segmentation methods of FCN, SegNet, PSPNet, UNet [26] and DeepLabV3+. MIoU and PA were selected as indicators to test the segmentation performance of each method.
To test the segmentation efficiency of each method, the training time and single image prediction time, memory occupancy, and parameter quantity were selected as indicators.
In order to test the generalization ability of DeepMDSCBA and verify its robustness, segmentation and comparison experiments were performed on the constructed training and test sets, which comprised datasets of apple images with different levels of rot, different rot positions, different apple varieties, and complex backgrounds.
In order to verify the effectiveness of the ideas of DeepMDSCBA, such as adopting a more lightweight network (MobileNet) than the original feature extraction network (Xception), changing the ordinary convolution to DSC in the ASPP module, and adding CBAM to the feature extraction module and the ASPP module, the following ablation experiments were performed on the total test set.
(1)
DeepM: Based on the traditional DeepLabV3+ network, the feature extraction network was changed to a more lightweight MobileNetV2 network.
(2)
DeepMDS: On the basis of DeepM, the ordinary convolution in the ASPP module was changed to DSC.
(3)
DeepMCBA: On the basis of DeepM, CBAM was added to the feature extraction and ASPP modules.
(4)
DeepMDSCBA1: Based on DeepMDS, only CBAM was added to the feature extraction module.
(5)
DeepMDSCBA2: Based on DeepMDS, only CBAM was added to the ASPP module.
(6)
DeepMDSCBA: Based on DeepMDS, CBAM was added to the feature extraction and ASPP modules, which reflected the method proposed in this paper.

3.5.3. Dataset Configuration

The training set and the test set adopted completely different pictures with no intersection.
The training set comprised a dataset of 212 images of fully healthy apples of different varieties without any rot, a dataset of 240 images of apples with different degrees of rot, a dataset of 180 images of apples with different positions of rot, and a dataset of 216 images of apples with complex backgrounds. The training set details are shown in Table 4.
For the three experiments related to the unseen cases described in 4.3.5, the training set for each experiment does not contain the corresponding case in the test set as an unseen case, that is a rot degree of (40%, 60%], a rot position of the calyx view, and a variety of Honey Crisp, respectively, as can be seen in Table 5.
The test set was divided into four subsets: a dataset of 120 images of fully healthy apples of different varieties without any rot, a dataset of 200 images of apples with different degrees of rot, a dataset 90 images of apples with different rot positions, and a dataset of 50 images of apples with complex backgrounds. The sum of all test sets was the total test set. There were no repeated pictures in the test set, and apple image appeared multiple times. The details of the test set are shown in Table 5.

4. Results and Analysis

4.1. Performance of Segmentation

In order to verify the segmentation performance of the DeepMDSCBA model, the model trained by the training set was used to perform segmentation tests on the previously divided total test set. In the experiment, the FCN, SegNet, PSPNet, UNet and DeepLabV3+ methods were used for comparison with the proposed DeepMDSCBA method. Some segmentation results are shown in Figure 8.
It can be seen in Figure 8 that compared with the other five methods, DeepMDSCBA showed the highest degree of recognition of the edges of apples with complex backgrounds in the images, as well as fewer omissions and misclassifications, especially for the rotten parts of apples.
Using MIoU and PA as indicators, the segmentation performance of apple images with DeepMDSCBA and the other five methods were analyzed, and the results are shown in Table 6.
It can be seen in Table 6 that the MIoU of DeepMDSCBA was 87.1%, which was 5.5%, 3.1%, 3.5%, 4.1% and 3.4% higher than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+, respectively. The PA of DeepMDSCBA was 95.3%, which was 5.7%, 3.2%, 3.8%, 4.4% and 3.1% higher than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+, respectively. The results of the experiments showed that the adoption of CBAM in DeepMDSCBA improved the feature extraction ability and the segmentation accuracy of various apples in the image test set, further proving the performance of the method.

4.2. Efficiency of Segmentation

It can be seen in Table 7 that the training time of DeepMDSCBA was 3.52 h, which was 33%, 21%, 18% and 21% faster than that of FCN, PSPNet, UNet and DeepLabV3+, respectively. The single image prediction time of DeepMDSCBA was 32 ms, which was 42%, 15%, 26% and 42% faster than FCN, PSPNet, UNet and DeepLabV3+, respectively. DeepMDSCBA occupied 7.1 GB of memory, which was 20%, 10%, 12%, 18% and 8% less than FCN, SegNet, PSPNet, UNet and DeepLabV3+, respectively. In terms of the number of parameters, the number of parameters of DeepMDSCBA was 22.6 MB, which was 89%, 23%, 49%, 76% and 84% lower than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+, respectively. However, compared with the SegNet method, the training time of DeepMDSCBA was almost 6% longer and the single image prediction time increased by 7%. This is because the SegNet method has a simpler method structure and fewer parameters than the proposed method [27], which results in a reduction in the method’s training time and single image prediction time. However, it can be seen in Figure 8 and Table 6 that the detection accuracy of SegNet was not as good as that of DeepMDSCBA. In general, because DeepMDSCBA (the method proposed in this paper) uses the lightweight network of MobileNet as its feature extraction network and changes the ordinary convolution in the ASPP module to DSC, it showed an improved calculation speed compared with the other tested methods.

4.3. Robustness Verification

In order to test the robustness of DeepMDSCBA, the segmentation performance of our DeepMDSCBA model trained using the training set on the four apple test sets was analyzed in comparison with the other five methods.

4.3.1. Segmentation Performance of Apple Images with Different Rot Degrees

Segmentation experiments were performed on the test set of apple images with complex backgrounds, using MIoU and PA as indicators. The comparison results are shown in Table 8.
It can be seen in Table 8 that the MIoU values of DeepMDSCBA on the test set of apple images with rot degrees of (0, 20%], (20%, 40%], (40%, 60%], (60%, 80%], (80%, 100%] were 86.7%, 84.8%, 83.9%, 84.4%, and 85.1%, respectively, and the PA values were 94.4%, 92.5%, 92.2%, 92.3%, and 93.4%, respectively, which were higher than those of FCN, SegNet, PSPNet, UNet, and DeepLabV3+. Furthermore, the segmentation performance of DeepMDSCBA was better than that of the other tested models.
In addition, it can be seen in Table 8 that for all tested methods, the segmentation performance first decreased and then increased as the proportion of the rotten area increased. The PA and MIoU of each method gradually decreased until the proportion of the spoiled area was in the interval of (60%, 80%]. When the proportion of the spoiled area was in the interval of (80%, 100%], the segmentation effect was similar to that of the spoiled area in the interval of (0, 20%].
The analysis showed that with the gradual increase in the rotten area of the entire apple peel, the normal area of the apple became irregular due to the rotten area and the boundary becoming more difficult to distinguish, resulting in a gradual decrease in segmentation accuracy. When the blackened area gradually spread out to the entire apple, the overall contour of the apple and the background color could be clearly distinguished, so the segmentation accuracy again increased.

4.3.2. Segmentation Performance of Apple Images with Different Rot Positions

Segmentation experiments were carried out on the test set of apple images with different rot positions using MIoU and PA as indicators. The results of the performance comparison are shown in Table 9.
It can be seen in Table 9 that the DeepMDSCBA MIoU of the test set of apple images with the view without the stalk or calyx, the calyx view, and the stalk view was 86.7%, 84.6%, and 84.7%, respectively, and the DeepMDSCBA PA was 94.4%, 92.6%, and 92.8%, respectively. These values were higher than those of FCN, SegNet, PSPNet, UNet, and DeepLabV3+, thus proving the segmentation performance of DeepMDSCBA was better than that of the other methods.
In addition, it can be seen in Table 9 that the DeepMDSCBA PA and MIoU values of the test set of apple images with the view without the stalk or calyx were higher than those of the test set of apple images with the calyx view and the stalk view. This is because the existence of the calyx and the stalk had a certain negative impact on the segmentation effect.

4.3.3. Segmentation Performance of Apple Images of Different Varieties

Segmentation experiments were performed on the test set of apple images of different varieties, using MIoU and PA as indicators. The comparison results are shown in Table 10.
It can be seen in Table 10 that the MIoU of DeepMDSCBA on the test set of apple images of the Golden Delicious, Fuji, Honey Crisp and Red Delicious varieties were 87.4%, 87.6%, 87.2% and 86.8%, respectively, and the corresponding PA values were 95.4%, 95.6%, 94.9% and 94.7%, respectively; these values higher than those of FCN, SegNet, PSPNet, UNet and DeepLabV3+.
In addition, it can be seen in Table 10 that there were no significant differences among the PA and MIoU of different apple varieties using the same segmentation method, indicating that apple varieties had little effect on segmentation performance.

4.3.4. Segmentation Performance of Apple Images with Complex Backgrounds

Segmentation experiments were carried out on the test set of apple images with complex backgrounds, using MIoU and PA as indicators. The comparison results are shown in Table 11.
On the apple image test set with complex backgrounds, the MIoU and PA of DeepMDSCBA were 86.8% and 94.4%, respectively, which were the highest of the tested method. The segmentation effect of DeepMDSCBA was better than FCN, SegNet, PSPNet, UNet and DeepLabV3+, and the segmentation accuracy of DeepMDSCBA was also improved.

4.3.5. Segmentation Performance for Unseen Cases

To further verify the robustness of DeepMDSCBA, the following three experiments were carried out. The test set for each following experiment consisted of a certain case, which is listed in Table 5, and the training set for the experiment consisted of the other cases. The segmentation performance of DeepMDSCBA was analyzed in comparison with the other five methods.
As can be seen in Table 5, apple images with a rot degree of (40%, 60%] were used as the test set and apple images with rot degrees of (0, 20%], (20%, 40%], (60%, 80%], and (80%, 100%] were used in the training set for this experiment. The results of the performance comparison are shown in Table 12.
It can be seen in Table 12 that the MIoU and PA of DeepMDSCBA were 84.1% and 91.7%, respectively, which were also the highest of the tested methods. This means that for the apple images with an unseen rot degree, the segmentation effect of DeepMDSCBA was better than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+.
For the positions of rot, apple images with the calyx view were used as the test set, and apple images with the other two views (fruit stalk view and view without the stalk or calyx) were used in the training set for this experiment. The results of the performance comparison are shown in Table 13.
The experimental results showed that the MIoU and PA of DeepMDSCBA were 84.7% and 92.3%, respectively, which were also the highest of the tested methods. The segmentation effect of DeepMDSCBA was better than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+ for apple images with an unseen rot position.
For apple varieties, images of the Honey Crisp apple variety were used as the test set, and images of the Golden Delicious, Fuji and Red Delicious apple varieties were used as the training set for this experiment. The results of the performance comparison are shown in Table 14.
According to the experimental results, the MIoU and PA of DeepMDSCBA were 85.9% and 93.6%, respectively, which were still the highest of the tested methods. The segmentation effect of DeepMDSCBA was better than that of FCN, SegNet, PSPNet, UNet and DeepLabV3+ for apple images of an unseen variety.
A comparison of the experimental results of each method for each test set showed that the PA and MIoU of DeepMDSCBA were higher than those of the other tested methods, which proved that the segmentation accuracy and effect of the proposed method on all kinds of test sets were improved. At the same time, it was verified that the method had a strong generalization ability and robustness.

4.4. Results of Ablation Experiments

Ablation experiments were carried out based on the ablation experiment scheme described in Section 3.5.2, and MIoU and PA were used as indicators. The results of the experiments are shown in Table 15.
According to the ablation experiment results in Table 15, DeepMDSCBA had improved MIoU and PA values compared with those of DeepLabV3+, DeepM, DeepMDS, DeepMCBA, DeepMDSCBA1, and DeepMDSCBA2, indicating that replacing the backbone network with MobileNetV2 improved the segmentation accuracy of the method to a certain extent. At the same time, the MIoU and PA of DeepMCBA, DeepMDSCBA1, DeepMDSCBA2 and DeepMDSCBA were better than those of DeepM, indicating that adding CBAM to the feature extraction module or ASPP module could improve the segmentation accuracy of the method.
Furthermore, DeepMDSCBA, the method proposed in this paper, showed greater improvements than DeepMDSCBA1 and DeepMDSCBA2, indicating that adding CBAM to the feature extraction and ASPP modules at the same time could improve the segmentation accuracy of the method. DeepMDSCBA had the highest MIoU and PA values, which proved the effectiveness of this method.
Furthermore, the training segmentation efficiencies of each ablation experimental method were compared, and the results are shown in Table 16.
It can be seen in Table 16 that the training time and single image prediction time of the DeepM, DeepMDS, DeepMCBA, DeepMDSCBA1, DeepMDSCBA2 and DeepMDSCBA methods were all reduced compared with those of DeepLabV3+, indicating that the adoption of a more lightweight network (MobileNetV2) could shorten training time and improve prediction speed.
The training time and single image prediction time of the DeepMDS, DeepMDSCBA1, DeepMDSCBA2, and DeepMDSCBA methods were shorter than those of DeepLabV3+ and DeepM, indicating that changing the ordinary convolution in the ASPP module to DSC could further shorten training time and improve prediction speed.
The method training time and single image prediction time of DeepMDSCBA were the lowest of the studied methods. According to the ablation experiments, the DeepMDSCBA method proposed in this paper reduced the computational complexity of the network, shortened the training running time, and improved the segmentation accuracy.

5. Conclusions

A network segmentation method based on deep learning, DeepMDSCBA, was proposed in this paper. The method combines DeepLabV3+ with the optimized lightweight MobileNetV2 network and uses DSC to replace the ordinary convolution in the ASPP module, which effectively reduces the number of method parameters and improves the speed of calculation. CBAM was added to the feature extraction module and the ASPP module to better restore the edge information of objects, improve the feature extraction ability of the method, and result in fewer omissions and misclassifications. The method proposed in this paper, DeepMDSCBA, was shown to more effectively extract apple areas in images than other test methods. The PA of the whole dataset of apple images reached 95.3% and the MIoU reached 87.1%, demonstrating a more efficient and accurate segmentation of apple images compared with other tested methods, even for images of rotten apples and apples with complex backgrounds.
By comparing it with five other semantic segmentation methods on test sets of apple rot degrees, rot positions, apple varieties, and complex backgrounds, the robustness of DeepMDSCBA was fully verified. It was also proven that the performance of DeepMDSCBA was better than the other tested methods under the influence of factors such as the degree of rot, position of rot, apple variety, and complex backgrounds.
Although DeepMDSCBA’s segmentation of images of rotten apples and apples with complex backgrounds was faster and more accurate than other tested methods, its segmentation of hidden areas was not accurate for apples that were partially hidden by leaves or branches in more complex situations. It is therefore necessary to construct a relevant dataset and conduct further experimental research in the future.

Author Contributions

Conceptualization, Y.F. and L.M.; methodology, Y.F., G.W., X.Y., X.W. and P.W.; software, Y.F.; validation, Y.F.; formal analysis, Y.F., X.Y., X.W. and P.W.; investigation, Y.F., G.W., X.Y., X.W. and P.W.; data curation, Y.F.; writing—original draft. Y.F.; resources, G.W. and L.M.; writing—review and editing, G.W. and L.M.; visualization, G.W., X.Y. and X.W.; supervision, G.W. and L.M.; project administration. G.W.; funding acquisition, L.M. and P.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Patrício, D.I.; Rieder, R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef] [Green Version]
  2. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  3. Wang, D.; Fu, Y.; Yang, G.; Yang, X.; Liang, D.; Zhou, C.; Zhang, N.; Wu, H.; Zhang, D. Combined Use of FCN and Harris Corner Detection for Counting Wheat Ears in Field Conditions. IEEE Access 2019, 7, 178930–178941. [Google Scholar] [CrossRef]
  4. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid scene parsing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  5. Cho, K.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  6. Deng, J.; Niu, Z.; Zhang, X.; Zhang, J.; Pan, S.; Mu, H. Kiwifruit vine extraction based on low altitude UAV remote sensing and deep semantic segmentation. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Computer Applications (ICAICA), Dalian, China, 28–30 November 2021. [Google Scholar]
  7. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Zhang, D.; Ding, Y.; Chen, P.; Zhang, X.; Pan, Z.; Liang, D. Automatic extraction of wheat lodging area based on transfer learning method and deeplabv3+ network. Comput. Electron. Agric. 2020, 179, 105845. [Google Scholar] [CrossRef]
  9. Sharifzadeh, S.; Tata, J.; Sharifzadeh, H.; Tan, B. Farm area segmentation in satellite images using deeplabv3+ neural networks. In International Conference on Data Management Technologies and Applications; Springer: Cham, Switzerland; Berlin, Germany, 2019. [Google Scholar]
  10. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  11. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  12. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  13. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Cho, K.; Van Merrienboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  15. Xiang, Q.; Wang, X.; Li, R.; Zhang, G.; Lai, J.; Hu, Q. Fruit image classification based on Mobilenetv2 with transfer learning technique. In Proceedings of the 3rd International Conference on Computer Science and Application Engineering, Sanya, China, 22–24 October 2019. [Google Scholar]
  16. Mingxing, T.; Le, Q.V. Mixconv: Mixed depthwise convolutional kernels. arXiv 2019, arXiv:1907.09595. [Google Scholar]
  17. Binh-Son, H.; Tran, M.; Yeung, S. Pointwise convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  18. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  19. Tongxue, Z.; Canu, S.; Ruan, S. Automatic COVID-19 CT segmentation using U-Net integrated spatial and channel attention mechanism. Int. J. Imaging Syst. Technol. 2021, 31, 16–27. [Google Scholar]
  20. Zhu, X.; Cheng, D.; Zhang, Z.; Lin, S.; Dai, J. An empirical study of spatial attention mechanisms in deep networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  21. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  22. Qi, L.; Liu, Y.; Sun, Z. Age progression and regression with spatial attention modules. In Proceedings of the AAAI conference on artificial intelligence, New York, NY, USA, 7–12 February 2020; Volume 34. No. 07. [Google Scholar]
  23. Tung, F.; Little, J.J. Scene parsing by nonparametric label transfer of content-adaptive windows. Comput. Vis. Image Underst. 2016, 143, 191–200. [Google Scholar] [CrossRef]
  24. Li, Q.; Xie, B.; You, J.; Bian, W.; Tao, D. Correlated logistic method with elastic net regularization for multilabel image classification. IEEE Trans. Image Process. 2016, 25, 3801–3813. [Google Scholar] [CrossRef] [Green Version]
  25. Ben, H. Cityscapes: Cultural Readings in the Material and Symbolic City; Red Globe Press: New York, NY, USA, 2005. [Google Scholar]
  26. Olaf, R.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland; Berlin, Germany, 2015. [Google Scholar]
  27. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Structure of DeepMDSCBA method.
Figure 1. Structure of DeepMDSCBA method.
Foods 11 03999 g001
Figure 2. Standard convolution and depthwise separable convolution operation. (a) Standard convolution process. (b) Depthwise convolution process. (c) Pointwise convolution process.
Figure 2. Standard convolution and depthwise separable convolution operation. (a) Standard convolution process. (b) Depthwise convolution process. (c) Pointwise convolution process.
Foods 11 03999 g002aFoods 11 03999 g002b
Figure 3. Structure of CBAM.
Figure 3. Structure of CBAM.
Foods 11 03999 g003
Figure 4. Apple images with different degrees of rot and their labels.
Figure 4. Apple images with different degrees of rot and their labels.
Foods 11 03999 g004
Figure 5. Apple images with different rot positions and their labels. (a) View without the stalk or calyx. (b) View of calyx. (c) View of stalk.
Figure 5. Apple images with different rot positions and their labels. (a) View without the stalk or calyx. (b) View of calyx. (c) View of stalk.
Foods 11 03999 g005
Figure 6. Images and labels of different apple varieties. (a) Golden Delicious. (b) Fuji. (c) Honey Crisp. (d) Red Delicious.
Figure 6. Images and labels of different apple varieties. (a) Golden Delicious. (b) Fuji. (c) Honey Crisp. (d) Red Delicious.
Foods 11 03999 g006
Figure 7. Apple images with complex backgrounds and their labels.
Figure 7. Apple images with complex backgrounds and their labels.
Foods 11 03999 g007
Figure 8. Results of segmentation using different algorithms.
Figure 8. Results of segmentation using different algorithms.
Foods 11 03999 g008
Table 1. MobileNetV2 network structure.
Table 1. MobileNetV2 network structure.
InputOperatortcns
256 × 256 × 3conv2d-3212
128 × 128 × 32bottleneck11611
128 × 128 × 16bottleneck62422
64 × 64 × 24bottleneck63232
32 × 32 × 32bottleneck66442
32 × 32 × 64bottleneck69631
16 × 16 × 96bottleneck616032
8 × 8 × 160bottleneck632011
Table 2. Experimental software and hardware configuration.
Table 2. Experimental software and hardware configuration.
ItemDetail
CPUAMD Ryzen 7 5800H with Radeon Graphics @3.20 GHz
RAM16 GB
Operating systemWindows 11 64-bit
CUDACUDA10.0
Python Python 3.6
Table 3. Optimal hyperparameters.
Table 3. Optimal hyperparameters.
EpochBatch SizeLrInput Shape
100160.0001512 × 512
Table 4. Training Set.
Table 4. Training Set.
Training SetSize (Before Enhancement)Size (After Enhancement)
Rot degrees(0, 20%]53159
(20%, 40%]57171
(40%, 60%]45135
(60%, 80%]47141
(80%, 100%]38114
Subtotal240720
Rot positionsView without stalk or calyx60180
Calyx view60180
Stalk view60180
Subtotal180540
VarietiesGolden Delicious53159
Fuji55165
Honey Crisp46138
Red Delicious58174
Subtotal212636
Complex backgrounds216648
Total8482544
Table 5. Test set.
Table 5. Test set.
ExperimentTest SetSize
PerformanceTotal460
EfficiencyTotal460
RobustnessDifferent Rot Degrees(0, 20%]40
(20%, 40%]40
(40%, 60%]40
(60%, 80%]40
(80%, 100%]40
Subtotal200
Rot PositionsView without stalk or calyx30
Calyx view30
Stalk view30
Subtotal90
VarietiesGolden Delicious30
Fuji30
Honey Crisp30
Red Delicious30
Subtotal120
Complex backgrounds50
Unseen Cases (40%, 60%]40
Calyx view30
Honey Crisp30
AblationTotal460
Table 6. Performance of different apple image segmentation methods.
Table 6. Performance of different apple image segmentation methods.
MethodMIoU (%)PA (%)
FCN82.590.1
SegNet84.492.3
PSPNet84.191.8
UNet83.691.2
DeepLabV3+84.292.4
DeepMDSCBA87.195.3
Table 7. Efficiency of different apple image segmentation methods.
Table 7. Efficiency of different apple image segmentation methods.
MethodTraining Time/hSingle Image Prediction Time/msMemory Occupancy/GBParameter Quantity/MB
FCN5.32568.9209.4
SegNet3.34308.020.44
PSPNet4.43388.144.1
UNet4.27438.795.3
DeepLabV3+4.51557.7143.8
DeepMDSCBA3.52327.122.6
Table 8. Segmentation performance of apple images with different rot degrees.
Table 8. Segmentation performance of apple images with different rot degrees.
MethodIndicator(0, 20%](20%, 40%](40%, 60%](60%, 80%](80%, 100%]
FCNMIoU (%)82.180.179.480.481.4
PA (%)88.987.586.487.988.4
SegNetMIoU (%)83.782.982.082.383.2
PA (%)90.689.788.989.490.4
PSPNetMIoU (%)83.181.781.181.582.8
PA (%)89.988.688.188.789.7
UNetMIoU (%)82.681.380.581.782.4
PA (%)89.588.287.588.689.3
DeepLabV3+MIoU (%)83.882.681.982.483.1
PA (%)90.889.689.189.489.9
DeepMDSCBAMIoU (%)86.784.883.984.485.1
PA (%)94.492.592.292.393.4
Table 9. Segmentation performance of apple images with different rot positions.
Table 9. Segmentation performance of apple images with different rot positions.
MethodIndicatorView without Stalk or CalyxCalyx ViewStalk View
FCNMIoU (%)79.978.478.6
PA (%)87.886.486.7
SegNetMIoU (%)84.682.782.5
PA (%)92.490.790.4
PSPNetMIoU (%)83.981.681.1
PA (%)91.789.488.4
UNetMIoU (%)82.179.679.2
PA (%)89.987.586.9
DeepLabV3+MIoU (%)84.882.582.2
PA (%)92.790.489.9
DeepMDSCBAMIoU (%)86.784.684.7
PA (%)94.492.692.8
Table 10. Segmentation performance of apple images of different varieties.
Table 10. Segmentation performance of apple images of different varieties.
MethodIndicatorGolden DeliciousFujiHoney CrispRed Delicious
FCNMIoU (%)82.181.981.881.7
PA (%)89.889.689.589.4
SegNetMIoU (%)84.784.584.784.8
PA (%)92.492.292.692.8
PSPNetMIoU (%)83.983.883.683.6
PA (%)91.891.691.491.4
UNetMIoU (%)83.482.983.283.6
PA (%)90.990.590.791.3
DeepLabV3+MIoU (%)85.184.984.785.1
PA (%)92.892.692.492.9
DeepMDSCBAMIoU (%)87.487.687.286.8
PA (%)95.495.694.994.7
Table 11. Segmentation performance of apple images with complex backgrounds.
Table 11. Segmentation performance of apple images with complex backgrounds.
MethodMIoU (%)PA (%)
FCN80.387.8
SegNet83.791.2
PSPNet83.290.7
UNet82.589.1
DeepLabV3+84.190.6
DeepMDSCBA86.894.4
Table 12. Segmentation performance for apple images with an unseen rot degree.
Table 12. Segmentation performance for apple images with an unseen rot degree.
MethodMIoU (%)PA (%)
FCN78.786.2
SegNet81.789.2
PSPNet80.487.9
UNet79.987.4
DeepLabV3+82.589.0
DeepMDSCBA84.191.7
Table 13. Segmentation performance for apple images with an unseen rot position.
Table 13. Segmentation performance for apple images with an unseen rot position.
MethodMIoU (%)PA (%)
FCN78.385.8
SegNet82.690.2
PSPNet81.589.1
UNet79.787.3
DeepLabV3+82.590.1
DeepMDSCBA84.792.3
Table 14. Segmentation performance for apple images with an unseen variety.
Table 14. Segmentation performance for apple images with an unseen variety.
MethodMIoU (%)PA (%)
FCN80.688.2
SegNet83.591.1
PSPNet82.490.0
UNet82.089.4
DeepLabV3+83.591.1
DeepMDSCBA85.993.6
Table 15. Segmentation performance of ablation experiments.
Table 15. Segmentation performance of ablation experiments.
MethodMIoU (%)PA (%)
DeepLabV3+83.792.6
DeepM84.192.9
DeepMDS84.393.1
DeepMCBA85.894.3
DeepMDSCBA184.693.5
DeepMDSCBA284.893.8
DeepMDSCBA87.195.3
Table 16. Segmentation efficiency of training in ablation experiments.
Table 16. Segmentation efficiency of training in ablation experiments.
MethodTraining Time/hSingle Image Prediction Time/ms
DeepLabV3+4.5155
DeepM4.2343
DeepMDS3.4134
DeepMCBA4.2344
DeepMDSCBA13.3735
DeepMDSCBA23.3438
DeepMDSCBA3.5232
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mo, L.; Fan, Y.; Wang, G.; Yi, X.; Wu, X.; Wu, P. DeepMDSCBA: An Improved Semantic Segmentation Model Based on DeepLabV3+ for Apple Images. Foods 2022, 11, 3999. https://doi.org/10.3390/foods11243999

AMA Style

Mo L, Fan Y, Wang G, Yi X, Wu X, Wu P. DeepMDSCBA: An Improved Semantic Segmentation Model Based on DeepLabV3+ for Apple Images. Foods. 2022; 11(24):3999. https://doi.org/10.3390/foods11243999

Chicago/Turabian Style

Mo, Lufeng, Yishan Fan, Guoying Wang, Xiaomei Yi, Xiaoping Wu, and Peng Wu. 2022. "DeepMDSCBA: An Improved Semantic Segmentation Model Based on DeepLabV3+ for Apple Images" Foods 11, no. 24: 3999. https://doi.org/10.3390/foods11243999

APA Style

Mo, L., Fan, Y., Wang, G., Yi, X., Wu, X., & Wu, P. (2022). DeepMDSCBA: An Improved Semantic Segmentation Model Based on DeepLabV3+ for Apple Images. Foods, 11(24), 3999. https://doi.org/10.3390/foods11243999

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop