Next Article in Journal
Solving Inverse Electrocardiographic Mapping Using Machine Learning and Deep Learning Frameworks
Previous Article in Journal
Short Single-Lead ECG Signal Delineation-Based Deep Learning: Implementation in Automatic Atrial Fibrillation Identification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Defect Detection of Subway Tunnels Using Advanced U-Net Network

1
Graduate School of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Japan
2
Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-ku, Sapporo 060-0814, Japan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(6), 2330; https://doi.org/10.3390/s22062330
Submission received: 7 February 2022 / Revised: 8 March 2022 / Accepted: 13 March 2022 / Published: 17 March 2022

Abstract

:
In this paper, we present a novel defect detection model based on an improved U-Net architecture. As a semantic segmentation task, the defect detection task has the problems of background–foreground imbalance, multi-scale targets, and feature similarity between the background and defects in the real-world data. Conventionally, general convolutional neural network (CNN)-based networks mainly focus on natural image tasks, which are insensitive to the problems in our task. The proposed method has a network design for multi-scale segmentation based on the U-Net architecture including an atrous spatial pyramid pooling (ASPP) module and an inception module, and can detect various types of defects compared to conventional simple CNN-based methods. Through the experiments using a real-world subway tunnel image dataset, the proposed method showed a better performance than that of general semantic segmentation including state-of-the-art methods. Additionally, we showed that our method can achieve excellent detection balance among multi-scale defects.

1. Introduction

With the growth of economies worldwide, various infrastructures such as tunnels, bridges, and viaducts have been constructed [1], which are indispensable for our daily life and are used by a large number of people in their daily lives [2]. However, infrastructures built more than five decades ago are experiencing aging problems, and the number of dilapidated infrastructures will significantly increase in the near future [1], and their maintenance cost will also increase exponentially [3,4]. Under these circumstances, a more efficient maintenance and management of infrastructures have become an urgent issue. Recently, much interest has been shown in smart maintenance and management technologies, including artificial intelligence (AI), Internet of Things (IoT), and big data analysis. These techniques have already been applied to real-world problems in various fields [5,6,7,8,9]. These techniques are required in the field of infrastructure to improve the efficiency and accuracy of infrastructure maintenance [10,11].
As an important infrastructure, urban railway systems have been mainly constructed during the high-speed economic growth period. In urban areas, the overground transportation network is already dense and its expansion potential is limited. On the other hand, the subway transportation environment, such as subway tunnels, is expected to expand further in the future. However, through the high frequency of use, tunnels that were built decades ago inevitably decay and suffer from a number of defects. Without repairs, these defects lead to significant economic losses and threaten safety.
In order to maintain a high level of security and economic growth, the daily maintenance and inspection of tunnels is necessary. Traditional inspection methods mainly rely on tunnel wall images taken by inspection vehicles or inspectors [12]. Inspectors look for deterioration such as cracks and leaks when taking images and the deterioration of tunnel walls is evaluated and repaired according to their conditions. This process is performed manually, and it takes much time and labor. Technologies that enable the automatic detection of defects are required to facilitate this process [13,14].
The standard strategy for supporting the inspection of subway tunnels is to construct a detector for the estimation of defects from tunnel wall images. Among all kinds of defects, automated crack detection has been studied for a long time, and various methods based on image processing have been proposed [15,16,17,18,19]. Recently, in the field of computer vision, the performance of image recognition has been significantly improved with the emergence of deep learning, which has been useful for various tasks [20,21,22,23,24]. Therefore, it is expected that image recognition technology will enable the development of a detector that can automatically identify defects in infrastructures.
Deep learning-based methods have achieved higher performances in detecting defect in infrastructures than traditional methods that use handcrafted image features [25,26,27]. However, when applying deep learning methods to real-world problems, various characteristics and situations have to be considered. Since there are various kinds of defects in subway tunnels such as cracks, cold joints, and leakages, existing deep learning methods cannot be directly applied to this task. Specifically, the following problems need to be addressed to improve detection performance:
  • Problem 1:
    Subway tunnel images have a high resolution and limited areas of defects. Hence, the problem of imbalance between the background and foreground in semantic segmentation is prominent.
  • Problem 2:
    Defects in subway tunnels have multi-scale variations. It is necessary to distinguish between these types since the repair action differs depending on the type of defect.
  • Problem 3:
    Subway tunnel image contains a complex background. Although there are no defects in the background area, it often contains structures similar to the defects due to the construction conditions.
Hence, it is desirable to devise more effective network architectures that can recover the details of defects in subway tunnel images and improve the detection accuracy of multi-scale defects.
To solve the above problems, we focus on the U-Net architecture [28], one of the most widely used methods in biomedical image segmentation tasks. The U-Net’s skip connection method, which can concatenate up-sampled feature maps with feature maps skipped from an encoder, makes it possible to effectively capture details and location information about objects. U-Net and its variants have achieved impressive segmentation results in computer vision tasks, especially in detecting multi-scale targets [29,30,31,32]. Because the cracks feature in our task is long and thin, we require the network to have the capacity to maintain the feature in high resolution; U-Net is a suitable choice for this. Specifically, the feature of cracks (small targets) is mainly captured by the high-resolution layer, and the water leakage feature is mostly captured by the low-resolution layer. Because of the succinct architecture, it is easy to add extra modules or change the architecture to improve the detection capacity for different kinds of segmentation targets in our task. The U-Net architecture is, therefore, suitable for our task.
In this paper, we propose an improved version of the U-Net architecture to solve the above problems. As a network design for the multi-scale target segmentation of a particular image dataset, the U-Net architecture is a suitable foundation network for our task. To solve Problem 1, we adjust the image dataset to balance background and foreground images to overcome the problem of background examples dominating gradients. To solve Problems 2 and 3, we optimize the network architecture using the following strategies: First, we replace all convolution blocks of the U-Net architecture with inception blocks [33]. Since the inception module consists of four different branches with different kernel sizes and enlarges the network’s receptive field, we can improve the network adaption to different scales of features. For our task, this improvement increases the capacity to detect multi-scale defects. In addition, for the same purpose, we replace the first convolution layer of the bridge layer with an atrous spatial pyramid pooling (ASPP) module from Deeplab-v2 [34]. Combining these two kinds of structures results in more precise detection and mitigates the over-fitting problem.
Our contributions are summarized as follows:
  • We propose a novel advanced U-Net for defect detection using subway tunnel images.
  • We design an architecture that can grasp the characteristics of a variety of defects. The experimental results show the effectiveness of our new architecture.
This paper is organized as follows: Summaries of related works on defect detection and classification are presented in Section 2. Next, Section 3 shows the data characteristics, and Section 4 shows the proposed method and the adopted network architectures. The experimental results are shown in Section 5. Finally, our conclusion is presented in Section 6.

2. Related Works

In this section, we discuss related works of computer vision tasks for application, U-Net family, and defect detection, respectively. Recent application tasks in computer vision are mentioned in Section 2.1, more specific architectures based on U-Net are explained in Section 2.2, and methods for defect detection are presented in Section 2.3.

2.1. Computer Vision Task for Application

Computer vision tasks have made great progress with the rise of deep learning technologies [35]. In the past, computer vision tasks have been studied mainly with the aim of recognizing objects in images; however, with the rise of deep learning, the level of accuracy close to real-world applications has been achieved [25,36]. The recognition level of general objects exceeded human accuracy in a competition held in 2015, and various methods for more advanced tasks such as object detection and pixel-level segmentation have been proposed [37,38]. In parallel, this technology has been applied not only in the field of computer science but also in various other fields. Transfer learning has shown that feature representations acquired by general image recognition can be useful for tasks in other domains [39,40]. In addition, a number of studies have been proposed for tasks where the amount of data is not sufficient [41].
Following general images, medical images are the next area where the technology is expected to be applied to society [42,43,44]. Medical images are highly specialized due to the clarification of imaging standards, but the quality of the captured images is high. Therefore, supervised learning, which is the speciality of deep learning, has succeeded in building relatively accurate models [45].

2.2. Deep Learning with U-Net and Its Variants

As a well-known biomedical image segmentation network, U-Net architecture in 2015 has a completely symmetric encoder–decoder structure, where features extracted from the same size convolutional layers are concatenated with corresponding up-sampling layers; thus, high- or low-level feature maps can be preserved and inherited by the decoder to obtain more precise segmentation. After that, its variants were proposed in the following years and are still applied to real-world segmentation tasks nowadays.
The common improved variants of U-Net are committed to redesigning convolutional modules and modifying down- and up-sampling. Namely, many varying methods such as TernausNet [46], Res-UNet [47], Dense U-Net [48], and R2U-Net [31] have been proposed. For example, TenausNet replaces the encoder part with VGG11, Res-UNet and Dense U-Net replace all submodules with res-connection and dense-connection, and R2U-Net combines recurrent convolution and res-connection as a submodule. U-Net++ [29] and U-Net 3+ [30] hope to increase multi-scale target detection capacity. The main advantage of these variants is that they can capture features of different levels and integrate them through feature superposition.

2.3. Defect Detection in Infrastructures

Before the high development age of deep learning techniques, the defect detection methods were mainly developed by using image processing method. In [10,11], the authors conducted surveys of newly developed robotic tunnel inspection systems and showed that they overcome these disadvantages and achieve high-quality inspection results. Additionally, Huang et al. reported a method for analyzing the morphological characteristics and distribution characteristics of structural damage based on an intelligent analysis method from visible images of tunnel images [13]. Furthermore, Koch et al. reported computer vision-based distress detection and condition assessment approaches related to concrete and asphalt civil infrastructure [49]. In addition, several methods for automatic detection based on computer vision techniques have been proposed [21,22]. Khoa et al. proposed automatic crack detection and classification methods by using morphological image processing techniques and feature extraction based on distance histogram-based shape descriptors [21]. Furthermore, Zhang et al. proposed a method called online CP-ALS to incrementally update tensor component matrices, followed by self-tuning a one-class support vector machine [24] for online damage identification [22].
In recent years, deep learning techniques have been successfully applied to defect detection tasks based on real-world datasets. For instance, Kim et al. [50] used Mask R-CNN to detect and segment defects in multiple kings of civil infrastructure. Bai et al. [51] used Robust Mask R-CNN for the task of crack detection. Specifically, they proposed a two-step method, called cascaded network, in which ResNet is used to classify defects and then some state-of-art segmentation networks are used. Huang et al. [52] proposed an integrated method, which combines a deep learning algorithm and Mobile Laser Scanning (MLS) technology, achieving an automated three-dimensional inspection of water leakages in shield tunnel linings. Choi et al. [53] proposed a semantic damage detecting network (SDDNet) for crack segmentation, which achieves real-time segmentation effectively negating a wide range of various complex backgrounds and crack-like features. Chen et al. [54] present a switch module to improve the efficiency of the encoder–decoder model, demonstrating it with U-Net and DeepCrack as examples. In this way, deep learning-based defect detection methods have shown promising results for the classification and segmentation tasks with the benefit of high representation ability.

3. Dataset

In this section, we explain the inspection data used in our study. Figure 1 shows examples of the subway tunnel image data. We can see that the tunnel image data have different characteristics of natural image data. The size of the images is approximately 12,088 × 10,000 pixels or 12,588 × 10,000 pixels and the resolution is 1 mm/pixel, and so they can be considered high-resolution images. Typically, analyzing high-resolution images requires enormous computer resources and such image sizes are not used in the input of deep learning models. On the other hand, the resizing process results in the loss of fine-scale defects. We solve this problem by the patch division processing.
The subway tunnel image data consist of defect and background images. Figure 2 shows defect patch examples divided from original images shown in Figure 2 (a) cracks, (b) cold joint, (c) construction repair (d) deposition (e) peeling, and (f) trace of water leakage. As shown in Figure 2, we can see that each type of defect has its characteristics such as different texture edges and color features. As for a two-class segmentation task, this intra-class variance will cause false alarms. For instance, the size and color of cracks (Figure 2a) are different from those of traces of water leakage (Figure 2f).
Next, we show divided patch examples of background images that have no defects in Figure 3 (a) cable, (b) concrete joint, (c) connection component of overhead conductor rail, (d) passage tunnels (e) overhead conductor rail, and (f) lighter. In Figure 3, some of them have characteristics similar to those of defect images, which can also cause a serious false alarm problem.

4. Methodology

Inspired by Inception-v4, ASPP module, and U-Net, we propose a new model for defect detection. The proposed network combines the advantages of all three existing models. We explain data augmentation in Section 4.1 and introduce the architecture of our network in Section 4.2.

4.1. Data Augmentation

In this subsection, we propose our data augmentation strategy and patch selection method. First, we divide high-resolution subway tunnel images into multiple patches as shown in Figure 2 and Figure 3. Let P i ( i = 1 , 2 , 3 , , I ) denote divided patches derived from the original images shown in Figure 1, where I represents the number of patches. Because of the imbalance distribution and multi-scale defects, we used an overlap strategy to ensure exhaustive defect patches, which extend the patch dataset. In addition, to construct the dataset via patch selection, we experimentally obtained a large-scale dataset containing background B n ( n = 1 , 2 , , N ) and defect patches D m ( m = 1 , 2 , , M ) . Note that the ratio between M and N is approximately 7:3 and N + M = I.
For the training phase, since the dataset includes superfluous patches and a approximately half of them are background patches, it can cause a data imbalance problem. Under this condition, we randomly excluded some background patches to balance the number of patch samples. It should be noted that this strategy does not influence the detection accuracy. Finally, the ratio between defect and background patches can reach 1:1.
The advantage of data augmentation is that features between distributions of data can be resolved by pseudo-data generation. The model acquires a high degree of generality by learning to identify the transformed images as input. In recent years, this idea has been incorporated into self-supervised learning. In self-supervised learning, a transformation similar to data augmentation is performed, and learning is performed without labels. It has been reported that this method can dramatically improve the representational capability of the model itself. In this paper, we focus on data augmentation because we are interested in supervised learning.

4.2. Network Architecture

In this subsection, we explain the network architecture used in our method. Figure 4 depicts a model architecture of the proposed method, and Table 1 represents the details of our network. We chose U-Net as our backbone model to achieve a high performance in the special data segmentation task. To increase the rate of detection of multi-scale defects in subway tunnel data, first, we replaced the convolution blocks of the U-Net architecture with inception blocks modified from Inception-v3 as shown in Table 1. Inception blocks extend the feature capture area to increase accuracy and mitigate the over-fitting problem. Second, we added the ASPP module to our model, and we imitated the usage of the ASPP in Deeplab-v3+ to set it after the last layer of the encoder (the bridge layer, middle of the network) shown in Figure 5a. In shallow architectures, the size of the encoder’s last layer is no less than 16 × 16. We adjusted the parameter settings of multiple parallel atrous convolutions in the ASPP module for adaptation to our task. In the following, we explain the details of our model.
Our network consists of stacked layers of modified inception blocks shown in Figure 5b in the U-Net-based encoder–decoder network. The inception blocks consist of four parallel branches. Three of them have convolution layers with different kernel sizes, and the last one has one max-pooling layer. We replaced the 5 × 5 convolution layer with 5 × 1 and 1 × 5 convolution layers to decrease the training parameters. In the original U-Net architecture, the encoder part contains 8 convolution blocks. In addition, the output of every 2 convolution blocks is down-sampled by a max-pooling layer, and to construct a deeper network, we add one inception block before each max-pooling layer, increasing the total number of convolution operations in the encoder from 8 to 12.
At the end of the encoder part, we replaced the bridge’s first convolution layer with the ASPP module, which is shown in Figure 5a; the input was split into 5 equal partitions. In the original ASPP module, the atrous rates of three 3 × 3 convolutions were set to 6, 12, and 18 (with 256 filters and batch normalization) to adapt to the input size, which is over 37 × 37. When the rate value is close to the feature map size, the 3 × 3 filter degenerates to a 1 × 1 filter, and the atrous convolution loses its effectiveness. In our task, the input size was limited to 256 × 256 pixels, and after 4 max-pooling operations, the final input size of the ASPP module became 16 × 16, which is less than the required 37 × 37. Therefore, we changed the atrous rates from 4, 8, and 16 to 2, 4, and 6 to adapt to the input size. After the ASPP module, a 1 × 1 convolution operation (with 1024 channels) was added to merge the bridge layer.
In the decoder part, we used a convolution transpose layer (with a kernel size of 3 × 3 and a stride size of 2) to perform the up-sampling operation. Instead of using the deeper architecture as the encoder, we replaced all basic convolution layers with inception blocks.

5. Experiments and Results

This section shows quantitative and qualitative evaluations to confirm our network’s effectiveness for detecting defects in subway tunnel images. The experimental settings are explained in Section 5.1, and the results and discussion are presented in Section 5.2 and Section 5.3, respectively. Experimental data were provided by Tokyo Metro Co., Ltd, a Japanese subway company.

5.1. Settings

In our experiments, 47 images made up the subway tunnel image dataset. The images were obtained from visible cameras with high resolutions (e.g., 12,088 × 10,000 pixels or 12,588 × 10,000 pixels), and we divided the images into multiple patches of 256 × 256 pixels with a sliding interval of 64 pixels.
In the training phase, we filtered the patches using the strategy introduced in Section 4.1. The pixel-ground truth of defects was determined by inspectors. We selected 280,000 patches from 29 images as our training dataset. In this dataset, the ratio between the background and defect patches was set to 1:1. Then, in the validation phase, seven images were divided by the same strategy, as in the training phase, and finally, 71,818 patches were selected. The last 11 images were used in the test phase. We only used the same dividing strategy without abandoning background patches. Therefore, the number of patches used in the test phase was 326,172, which is significantly larger than that in the training phase. After the test phase, we generated estimation images by recombining the estimation results and the average probability of each pixel.
For the semantic segmentation task, Recall, Precision, F-measure, and Intersection over Union (IoU) were used to evaluate the binary classification performance as our estimation metrics. They can be calculated as follows:
R e c a l l = TP TP + FN ,
P r e c i s i o n = TP TP + FP ,
F-measure = 2 × Recall × Precision Recall + Precision ,
I o U = TP TP + FP + FN ,
where TP, TN, FP, and FN represent the number of true-positive, true-negative, false-positive, and false-negative samples, respectively.
We compared our method with classic segmentation methods including Deeplab-v3+ (CM1) [55], FCN (CM2) [56], and SegNet (CM3) [57]. Since the input of the network was set to 256 × 256, the output size of the encoder in Deeplab-v3+ was 16 × 16. According to our method, we adjusted the parameter settings of multiple parallel atrous convolutions in the ASPP module using the same strategy as introduced in Section 4.2. In addition, since our network is based on the U-Net architecture, we added several previous U-Net versions as comparative methods (CM4-CM7). The design of each method is shown in Table 2. Among them, CM5 [58] added additional down-sampling blocks to both the encoder and decoder of the network, changing the down-sampling stride from 16 to 32.

5.2. Results

In this subsection, we show the evaluation results and discuss some important details of the proposed model.

5.2.1. Quantitative Analysis

Table 3 shows the detection rate of all defects. From Table 3, we can compare the defect detection performance of our method and comparative methods (CM1-CM7). In these metrics, IoU, which is the standard metric of the semantic segmentation field, is the most important value to evaluate the total performance. We can see that PM obviously outperformed all CMs in this metric.
Next, Table 4 shows the recall rate of detection of each defect. From Table 4, we can observe the specific defect detection performance of our method and comparative methods (CM1-CM7). It should be noted that the metric Recall was used for the evaluation of each defect detection performance since the small crack defects were directly included. For the evaluation of the detection performance of cracks, IoU is not the best evaluation metric because of the difficulty of pixel-level matching. Moreover, considering the application situation, over-detection is considered preferable to miss-detection for the detection of defects. From the above reasons, we selected the evaluation metric Recall in this evaluation.
The proposed method outperforms all comparative methods. According to Table 3 and Table 4, we can further discuss the importance of each component.
  • Limitation of Deeplab-v3+ (CM1):
    Deeplab-v3+ used atrous convolution, ASPP module, and a simplified decoder branch, achieving great improvement compared with the baseline. There was a slight difference in the detection accuracy for various kinds of defect. Although Deeplab-v3+ applied multiple kinds of modules to improve detection performance for multi-scale defects, it still lacks detection accuracy for large-scale defects as shown in Table 3.
  • FCN and SegNet (CM2, CM3):
    FCN and SegNet, as classic segmentation networks, show a certain degree of incompatibility in our subway tunnel dataset, not only with a low accuracy but also with a large number of false detection instances as shown in Table 3. Especially, the performance of SegNet is extremely poor. Although the detection accuracy of small targets such as cracks can be maintained, it is almost impossible to detect large defects as shown in Table 4. These result in the low overall detection accuracy and precision of the network. Unlike U-Net, the SegNet decoder uses the max-pooling indices received from the corresponding encoder to perform nonlinear upsampling of the input feature map as a typical symmetric encoder–decoder architecture. It is considered that this function did not work well in the subway tunnel dataset.
  • Effectiveness of ASPP module (CM4):
    In CM4, this module increases F-measure from 0.428 to 0.444 and IoU from 0.272 to 0.286 compared with the baseline module (CM7) in Table 3. Additionally, the obtained results from Table 4 suggest that the addition of the ASPP module significantly improved the detection performance of small-, medium-, and large-scale defects. The obtained results show the effectiveness of the ASPP module.
  • Effectiveness of layer extend operation (CM5):
    In CM5, compared with the baseline (CM7), this module increases F-measure from 0.428 to 0.495 and IoU from 0.272 to 0.329 as shown in Table 3. Additionally, Table 4 suggests that CM5 is superior to CM4, CM6, and the baseline (CM7). These results suggest that deeper networks improve the detection of all scales of defects. However, this operation could not be applied to networks with the ASPP module due to patch size limitations in the experimental setting.
  • Effectiveness of Inception module (CM6):
    In CM6, we only replaced all convolution blocks with the inception module. This operation increased F-measure from 0.428 to 0.443 and IoU from 0.272 to 0.285 compared with the baseline (CM7) in Table 3. Additionally, Table 4 shows that the detection rate of each scale significantly improved compared with the baseline. This indicates that the addition of the inception module can contribute to the representation ability of low- and high-level information.
  • Analysis of the proposed method:
    As shown in Table 3, PM outperformed all other methods. Furthermore, from Table 4, we can see that PM achieves better accuracy in detecting large-scale defects but has some limitations in detecting small-scale defects. The limitation of small-scale defects may influence the detection performance of the inspection task. Thus, qualitative analysis is also required.

5.2.2. Qualitative Analysis

In this part, we discuss the visual quality of the results. The estimation results are shown in Figure 6, Figure 7, Figure 8 and Figure 9. Figure 6 shows detection result samples of all regions of the test image. Figure 7 and Figure 8 show the detection results of peeling and cracks. From Figure 6, Figure 7 and Figure 8, we can see that PM achieves a high detection quality when detecting various defects compared to CMs. On the other hand, we show the over-fitting result sample in Figure 9. In some cases, we observed that vertical cracks tend to over-fit in our model. The quantitative analyses show that the proposed method has some limitations in detecting small-scale defects, and according to Figure 8, these limitations may not influence the actual inspection works. Compared with all CMs, the result of PM achieves fewer instances of false detection, which would lead to less unnecessary work for inspectors.

5.3. Discussion

In the field of image recognition, various models have been proposed consistently owing to the AI boom. In the models for general object recognition, the error rate of recognition now exceeds that of humans, and there is a glimpse of a direction to target more advanced tasks. Applications of AI are beginning to be explored in all areas, one of which is infrastructure maintenance. In this paper, we have proposed a method for detecting defects in subway tunnel images. By constructing a model that takes into account the characteristics of the data, the proposed method achieved a higher accuracy in detecting defects compared to conventional methods.
What we should consider here is how much the system should achieve to reach the accuracy that can be applied in the real world. The quantitative evaluation results obtained from this experiment showed that the IoU was around 0.3–0.4. This value may not be sufficient when compared to the accuracy of general image recognition. However, as shown in the results of the qualitative evaluation, cracks and other defects in the image can be detected even if there is some deviation. For example, if we consider the practical applications of the proposed method, such as supporting the registration of defects in CAD systems or identifying dense regions of defects, we can say that the proposed method has reached a system that can be applied in practice.
There are some limitations in this study. This study was conducted using data from a certain subway line in Japan, and there is still room for future studies on the general applicability to a wide variety of data. In this study, 47 high-resolution subway tunnel images were divided into patches to enable the network training; however, it would be desirable to have a larger number of images to verify the robustness of our method. In addition, since the accuracy is considered to vary depending on the year of construction of tunnels, verification using a wide variety of data is necessary. Specifically, the condition of the wall depends on the construction method of the subway tunnel, and furthermore, the new construction method may be completely different from the conventional construction method. When considering the versatility of the model, it will be necessary to verify the versatility of the model for various types of data.

6. Conclusions

In this study, we present a new version of the U-Net architecture to improve the detection performance of defects in subway tunnel images. By introducing ASPP and inception modules in the U-Net-based network architecture, we improved the capacity of the network for defect detection. The experimental results on a real-world subway tunnel image dataset showed that our method outperformed other segmentation methods quantitatively and qualitatively. Different from conventional crack detection methods, our model can detect various types of defects in a single model, which enhances the practicality for supporting tunnel inspections. In future works, we will investigate a new strategy for enhancing detection accuracy and discuss its application to other real-world tasks.

Author Contributions

Conceptualization, A.W., R.T., T.O. and M.H.; methodology, A.W., R.T., T.O. and M.H.; software, A.W; validation, A.W., R.T., T.O. and M.H.; data creation, A.W.; writing— original draft preparation, A.W.; writing—review and editing, R.T., T.O. and M.H.; visualization, A.W.; funding acquisition, T.O. and M.H. All authors read and agreed to the published version of the manuscript.

Funding

This work was supported by KAKENHI Grant Number JP17H01744.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank Tokyo Metro Co., Ltd, for providing the research data used in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ministry of Land, Infrastructure, Transport and Tourism; White Paper on Land, Infrastructure, Transport and Tourism in Japan. 2019. Available online: https://www.mlit.go.jp/en/statistics/white-paper-mlit-index.html (accessed on 26 June 2019).
  2. Merenda, M.; Porcaro, C.; Iero, D. Edge machine learning for AI-enabled IoT devices: A review. Sensors 2020, 20, 2533. [Google Scholar] [CrossRef] [PubMed]
  3. Underwood, B.S.; Guido, Z.; Gudipudi, P.; Feinberg, Y. Increased costs to US pavement infrastructure from future temperature rise. Nat. Clim. Chang. 2017, 7, 704–707. [Google Scholar] [CrossRef]
  4. Onuma, A.; Tsuge, T. Comparing green infrastructure as ecosystem-based disaster risk reduction with gray infrastructure in terms of costs and benefits under uncertainty: A theoretical approach. Int. J. Disaster Risk Reduct. 2018, 32, 22–28. [Google Scholar] [CrossRef]
  5. Lee, J.; Park, G.L.; Han, Y.; Yoo, S. Big data analysis for an electric vehicle charging infrastructure using open data and software. In Proceedings of the Eighth International Conference on Future Energy Systems, Hong Kong, China, 16–19 May 2017; pp. 252–253. [Google Scholar]
  6. Lv, Z.; Hu, B.; Lv, H. Infrastructure monitoring and operation for smart cities based on IoT system. IEEE Trans. Ind. Informatics 2019, 16, 1957–1962. [Google Scholar] [CrossRef]
  7. Wang, J.; Yang, Y.; Wang, T.; Sherratt, R.S.; Zhang, J. Big data service architecture: A survey. J. Internet Technol. 2020, 21, 393–405. [Google Scholar]
  8. Arfat, Y.; Usman, S.; Mehmood, R.; Katib, I. Big data tools, technologies, and applications: A survey. In Smart Infrastructure and Applications; Springer: Berlin/Heidelberg, Germany, 2020; pp. 453–490. [Google Scholar]
  9. Zhu, L.; Yu, F.R.; Wang, Y.; Ning, B.; Tang, T. Big data analytics in intelligent transportation systems: A survey. IEEE Trans. Intell. Transp. Syst. 2018, 20, 383–398. [Google Scholar] [CrossRef]
  10. Montero, R.; Victores, J.G.; Martinez, S.; Jardón, A.; Balaguer, C. Past, present and future of robotic tunnel inspection. Autom. Constr. 2015, 59, 99–112. [Google Scholar] [CrossRef]
  11. Pouliot, N.; Richard, P.L.; Montambault, S. LineScout Technology Opens the Way to Robotic Inspection and Maintenance of High-Voltage Power Lines. IEEE Power Energy Technol. Syst. J. 2015, 2, 1–11. [Google Scholar] [CrossRef]
  12. Dung, C.V. Autonomous concrete crack detection using deep fully convolutional neural network. Autom. Constr. 2019, 99, 52–58. [Google Scholar] [CrossRef]
  13. Huang, Z.; Fu, H.; Chen, W.; Zhang, J.; Huang, H. Damage detection and quantitative analysis of shield tunnel structure. Autom. Constr. 2018, 94, 303–316. [Google Scholar] [CrossRef]
  14. Hastak, M.; Baim, E.J. Risk factors affecting management and maintenance cost of urban infrastructure. J. Infrastruct. Syst. 2001, 7, 67–76. [Google Scholar] [CrossRef]
  15. Mohan, A.; Poobal, S. Crack detection using image processing: A critical review and analysis. Alex. Eng. J. 2018, 57, 787–798. [Google Scholar] [CrossRef]
  16. Yamaguchi, T.; Hashimoto, S. Fast crack detection method for large-size concrete surface images using percolation-based image processing. Mach. Vis. Appl. 2010, 21, 797–809. [Google Scholar] [CrossRef]
  17. Liu, Z.; Suandi, S.A.; Ohashi, T.; Ejima, T. Tunnel crack detection and classification system based on image processing. In Proceedings of the Machine Vision Applications in Industrial Inspection X, San Jose, CA, USA, 21–22 January 2002; Volume 4664, pp. 145–152. [Google Scholar]
  18. Yiyang, Z. The design of glass crack detection system based on image preprocessing technology. In Proceedings of the 2014 IEEE 7th Joint International Information Technology and Artificial Intelligence Conference, Chongqing, China, 20–21 December 2014; pp. 39–42. [Google Scholar]
  19. Nishikawa, T.; Yoshida, J.; Sugiyama, T.; Fujino, Y. Concrete crack detection by multiple sequential image filtering. Comput.-Aided Civ. Infrastruct. Eng. 2012, 27, 29–47. [Google Scholar] [CrossRef]
  20. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road crack detection using deep convolutional neural network. In Proceedings of the International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3708–3712. [Google Scholar]
  21. Khoa, N.L.D.; Anaissi, A.; Wang, Y. Smart infrastructure maintenance using incremental tensor analysis. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 959–967. [Google Scholar]
  22. Zhang, W.; Zhang, Z.; Qi, D.; Liu, Y. Automatic crack detection and classification method for subway tunnel safety monitoring. Sensors 2014, 14, 19307–19328. [Google Scholar] [CrossRef] [PubMed]
  23. Yang, X.; Li, H.; Yu, Y.; Luo, X.; Huang, T.; Yang, X. Automatic pixel-level crack detection and measurement using fully convolutional network. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 1090–1109. [Google Scholar] [CrossRef]
  24. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  26. Deng, L.; Yu, D. Deep Learning: Methods and Applications. Foundations and Trends® in Signal Processing. Signal Process 2014, 7, 197–387. [Google Scholar]
  27. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M.; Gao, M.; et al. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [Green Version]
  28. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  29. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  30. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.W.; Wu, J. UNet 3+: A Full-Scale Connected UNet for Medical Image Segmentation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  31. Alom, M.Z.; Hasan, M.; Yakopcic, C.; Taha, T.M.; Asari, V.K. Recurrent residual convolutional neural network based on u-net (r2u-net) for medical image segmentation. arXiv 2018, arXiv:1802.06955. [Google Scholar]
  32. Diakogiannis, F.I.; Waldner, F.; Caccetta, P.; Wu, C. Resunet-a: A deep learning framework for semantic segmentation of remotely sensed data. ISPRS J. Photogramm. Remote. Sens. 2020, 162, 94–114. [Google Scholar] [CrossRef] [Green Version]
  33. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  34. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  37. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  38. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 7263–7271. [Google Scholar]
  39. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 1–40. [Google Scholar] [CrossRef] [Green Version]
  40. Samala, R.K.; Chan, H.P.; Hadjiiski, L.; Helvie, M.A.; Richter, C.D.; Cha, K.H. Breast Cancer Diagnosis in Digital Breast Tomosynthesis: Effects of Training Sample Size on Multi-Stage Transfer Learning Using Deep Neural Nets. IEEE Trans. Med. Imaging 2019, 38, 686–696. [Google Scholar] [CrossRef] [PubMed]
  41. Togo, R.; Watanabe, H.; Ogawa, T.; Haseyama, M. Deep convolutional neural network-based anomaly detection for organ classification in gastric X-ray examination. Comput. Biol. Med. 2020, 123, 103903. [Google Scholar] [CrossRef]
  42. Togo, R.; Yamamichi, N.; Mabe, K.; Takahashi, Y.; Takeuchi, C.; Kato, M.; Sakamoto, N.; Ishihara, K.; Ogawa, T.; Haseyama, M. Detection of gastritis by a deep convolutional neural network from double-contrast upper gastrointestinal barium X-ray radiography. J. Gastroenterol. 2019, 54, 321–329. [Google Scholar] [CrossRef] [Green Version]
  43. Togo, R.; Hirata, K.; Manabe, O.; Ohira, H.; Tsujino, I.; Magota, K.; Ogawa, T.; Haseyama, M.; Shiga, T. Cardiac sarcoidosis classification with deep convolutional neural network-based features using polar maps. Comput. Biol. Med. 2019, 104, 81–86. [Google Scholar] [CrossRef] [Green Version]
  44. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Togo, R.; Ogawa, T.; Haseyama, M. Synthetic Gastritis Image Generation via Loss Function-Based Conditional PGGAN. IEEE Access 2019, 7, 87448–87457. [Google Scholar] [CrossRef]
  46. Iglovikov, V.; Shvets, A. TernausNet: U-Net with VGG11 Encoder Pre-Trained on ImageNet for Image Segmentation. arXiv 2018, arXiv:1801.05746. [Google Scholar]
  47. Xiao, X.; Lian, S.; Luo, Z.; Li, S. Weighted Res-UNet for High-Quality Retina Vessel Segmentation. In Proceedings of the 2018 9th International Conference on Information Technology in Medicine and Education (ITME), Hangzhou, China, 19–21 October 2018; pp. 327–331. [Google Scholar] [CrossRef]
  48. Guan, S.; Khan, A.A.; Sikdar, S.; Chitnis, P.V. Fully Dense UNet for 2D Sparse Photoacoustic Tomography Artifact Removal. arXiv 2018, arXiv:1808.10848. [Google Scholar]
  49. Koch, C.; Georgieva, K.; Kasireddy, V.; Akinci, B.; Fieguth, P. A review on computer vision based defect detection and condition assessment of concrete and asphalt civil infrastructure. Adv. Eng. Informatics 2015, 29, 196–210. [Google Scholar] [CrossRef] [Green Version]
  50. Kim, B.; Cho, S. Automated Multiple Concrete Damage Detection Using Instance Segmentation Deep Learning Model. Appl. Sci. 2020, 10, 8008. [Google Scholar] [CrossRef]
  51. Bai, Y.; Sezen, H.; Yilmaz, A. End-to-end Deep Learning Methods for Automated Damage Detection in Extreme Events at Various Scales. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2020. [Google Scholar]
  52. Huang, H.; Cheng, W.; Zhou, M.; Chen, J.; Zhao, S. Towards Automated 3D Inspection of Water Leakages in Shield Tunnel Linings Using Mobile Laser Scanning Data. Sensors 2020, 20, 6669. [Google Scholar] [CrossRef]
  53. Choi, W.; Cha, Y. SDDNet: Real-Time Crack Segmentation. IEEE Trans. Ind. Electron. 2020, 67, 8016–8025. [Google Scholar] [CrossRef]
  54. Chen, H.; Lin, H.; Yao, M. Improving the Efficiency of Encoder-Decoder Architecture for Pixel-Level Crack Detection. IEEE Access 2019, 7, 186657–186670. [Google Scholar] [CrossRef]
  55. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  56. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  57. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  58. Wang, A.; Togo, R.; Ogawa, T.; Haseyama, M. Detection of Distress Region from Subway Tunnel Images via U-net-based Deep Semantic Segmentation. In Proceedings of the IEEE 8th Global Conference on Consumer Electronics (GCCE), Osaka, Japan, 15–18 October 2019; pp. 766–767. [Google Scholar]
Figure 1. Examples of subway tunnel images used in this study. (a,b) are sample images taken from a visible camera for inspection. (Resolution: 1 mm/pixel, Image size: 12,088 × 10,000 pixels).
Figure 1. Examples of subway tunnel images used in this study. (a,b) are sample images taken from a visible camera for inspection. (Resolution: 1 mm/pixel, Image size: 12,088 × 10,000 pixels).
Sensors 22 02330 g001
Figure 2. Example of defect images. (af) represent cracks, cold joint, construction repair, deposition, peeling, and trace of water leakage, respectively. (Resolution: 1 mm/pixel, Image size: 256 × 256 pixels).
Figure 2. Example of defect images. (af) represent cracks, cold joint, construction repair, deposition, peeling, and trace of water leakage, respectively. (Resolution: 1 mm/pixel, Image size: 256 × 256 pixels).
Sensors 22 02330 g002
Figure 3. Example of background images. (af) show cable, concrete joint, connection component of overhead conductor rail, passage tunnels, overhead conductor rail, and lighter, respectively. (Resolution: 1 mm/pixel, Image size: 256 × 256 pixels).
Figure 3. Example of background images. (af) show cable, concrete joint, connection component of overhead conductor rail, passage tunnels, overhead conductor rail, and lighter, respectively. (Resolution: 1 mm/pixel, Image size: 256 × 256 pixels).
Sensors 22 02330 g003
Figure 4. Overview of our defect detection network architecture.
Figure 4. Overview of our defect detection network architecture.
Sensors 22 02330 g004
Figure 5. Modules introduced in our method. (a) represents the architecture of ASPP module and (b) represents the inception module.
Figure 5. Modules introduced in our method. (a) represents the architecture of ASPP module and (b) represents the inception module.
Sensors 22 02330 g005
Figure 6. Results of proposed method and comparative methods. (From left to right: (a): original image; (b): ground truth; (c): results obtained by the proposed method; and (dj): results obtained by the comparative methods.)
Figure 6. Results of proposed method and comparative methods. (From left to right: (a): original image; (b): ground truth; (c): results obtained by the proposed method; and (dj): results obtained by the comparative methods.)
Sensors 22 02330 g006aSensors 22 02330 g006b
Figure 7. Example of the result in peeling detection. (a) Original image. (b) Ground Truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Figure 7. Example of the result in peeling detection. (a) Original image. (b) Ground Truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Sensors 22 02330 g007
Figure 8. Example of the result for crack detection. (a) Original image. (b) Ground truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Figure 8. Example of the result for crack detection. (a) Original image. (b) Ground truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Sensors 22 02330 g008
Figure 9. Example of the results of over-fitting parts. (a) Origin image. (b) Ground truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Figure 9. Example of the results of over-fitting parts. (a) Origin image. (b) Ground truth. (c) PM. (d) CM1. (e) CM2. (f) CM3. (g) CM4. (h) CM5. (i) CM6. (j) CM7.
Sensors 22 02330 g009
Table 1. Architecture of the proposed model.
Table 1. Architecture of the proposed model.
TypeSize/StrideOutput SizeDepth
Inception Module3 × 3/1256 × 256 × 643
Inception Module3 × 3/1256 × 256 × 643
Inception Module3 × 3/1256 × 256 × 643
Max Pooling3 × 3/2128 × 128 × 641
Inception Module3 × 3/1128 × 128 × 1283
Inception Module3 × 3/1128 × 128 × 1283
Inception Module3 × 3/1128 × 128 × 1283
Max Pooling3 × 3/264 × 64 × 1281
Inception Module3 × 3/164 × 64 × 2563
Inception Module3 × 3/164 × 64 × 2563
Inception Module3 × 3/164 × 64 × 2563
Max Pooling3 × 3/232 × 32 × 2561
Inception Module3 × 3/132 × 32 × 5123
Inception Module3 × 3/132 × 32 × 5123
Inception Module3 × 3/132 × 32 × 5123
Max Pooling3 × 3/216 × 16 × 5121
The ASPP module16 × 16 × 10242
Inception Module3 × 3/116 × 16 × 10243
Deconvolution3 × 3/232 × 32 × 5123
Cat32 × 32 × 5121
Inception Module3 × 3/132 × 32 × 5123
Inception Module3 × 3/132 × 32 × 5123
Deconvolution3 × 3/264 × 64 × 2561
Cat64 × 64 × 5121
Inception Module3 × 3/164 × 64 × 2563
Inception Module3 × 3/164 × 64 × 2563
Deconvolution3 × 3/2128 × 128 × 1281
Cat128 × 128 × 2561
Inception Module3 × 3/1128 × 128 × 1283
Inception Module3 × 3/1128 × 128 × 1283
Deconvolution3 × 3/2256 × 256 × 641
Cat256 × 256 × 1281
Inception Module3 × 3/1256 × 256 × 643
Inception Module3 × 3/1256 × 256 × 643
Sigmoid1 × 1/1256 × 256 × 11
Table 2. Differences in the proposed method (PM) and U-Net-based comparative methods (CM4-CM7) used in the experiment.
Table 2. Differences in the proposed method (PM) and U-Net-based comparative methods (CM4-CM7) used in the experiment.
MethodInceptionASPPLayer Extend
PM-
CM4--
CM5--
CM6--
CM7 (Baseline)---
Table 3. Defect detection performance of the proposed method (PM) and the comparative methods (CMs).
Table 3. Defect detection performance of the proposed method (PM) and the comparative methods (CMs).
MethodRecallPrecisionF-MeasureIoU
PM0.6600.4360.5250.356
CM1 [55]0.5640.3750.4510.291
CM2 [56]0.4940.3150.3850.238
CM3 [57]0.4100.1360.2040.158
CM40.4930.4050.4440.286
CM50.5320.4630.4950.329
CM60.6170.3460.4430.285
CM70.5880.3360.4280.272
Table 4. Recall of all kinds of defects in each method.
Table 4. Recall of all kinds of defects in each method.
DefectRecall
PMCM1CM2CM3CM4CM5CM6CM7
Peeling0.9210.8660.7290.1910.7950.9050.7110.655
Floating0.8020.7110.5680.1990.7080.7820.6510.533
Crack (0.3 mm–0.5 mm)0.1730.2300.1630.2090.1590.1400.1250.110
Crack (0.5 mm–1 mm)0.3580.3850.4300.3340.4070.3820.3610.326
Crack (1 mm–2 mm)0.4020.4630.3840.4220.4550.4340.4090.388
Crack(2mm+)0.4140.4090.3940.4310.4670.4440.4260.389
Cold joint0.0130.0170.0160.0140.0160.0160.0070.005
Honeycomb0.0840.2510.2300.0100.0300.2100.0900.080
Patching (intermediate pile)0.8190.7340.6160.1590.7210.8160.6560.591
Alligator crack0.3620.3080.2160.0630.3170.3680.3060.244
Early construction repair0.4230.3750.2710.0610.3940.5040.3060.297
Deposition0.0540.0490.0150.0010.0800.0120.0050.010
Construction repair0.5910.3070.1670.0780.4130.5560.3640.375
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, A.; Togo, R.; Ogawa, T.; Haseyama, M. Defect Detection of Subway Tunnels Using Advanced U-Net Network. Sensors 2022, 22, 2330. https://doi.org/10.3390/s22062330

AMA Style

Wang A, Togo R, Ogawa T, Haseyama M. Defect Detection of Subway Tunnels Using Advanced U-Net Network. Sensors. 2022; 22(6):2330. https://doi.org/10.3390/s22062330

Chicago/Turabian Style

Wang, An, Ren Togo, Takahiro Ogawa, and Miki Haseyama. 2022. "Defect Detection of Subway Tunnels Using Advanced U-Net Network" Sensors 22, no. 6: 2330. https://doi.org/10.3390/s22062330

APA Style

Wang, A., Togo, R., Ogawa, T., & Haseyama, M. (2022). Defect Detection of Subway Tunnels Using Advanced U-Net Network. Sensors, 22(6), 2330. https://doi.org/10.3390/s22062330

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop