1. Introduction
According to the National Bridge Inventory, 42% of the nearly 620,000 bridges in the United States were built more than 50 years ago (NBI) [
1]. Furthermore, 12% of all bridges were built more than 80 years ago, exceeding their intended design life of 75 years [
2]. To prevent potential safety hazards caused by bridge deterioration, the Federal Highway Administration (FHWA) requires a routine inspection at least every 2 years to evaluate each bridge’s condition [
3]. The FHWA and the US Department of Transportation (USDOT) allocate a considerable amount of annual funding towards improving bridge safety by standardizing inspections and evaluations. In May 2022, FHWA and USDOT announced the allocation of an additional
$1.14 billion in funding for bridge asset management [
4]. Simultaneously, they announced updates to the specifications to be used for future bridge inspections [
4]. This announcement stated that, effective after the annual submittal date (the anniversary date of the previous year’s inspection) specified in the FHWA’s Implementation Memo, bridge inspection data should be reported according to
The Specifications for the National Bridge Inventory (
SNBI 2022) [
5]. This specification will supersede the former
Recording and Coding Guide for the Structure Inventory and Appraisal of the Nation’s Bridge (
Coding Guide 1995) [
6].
A major change in the
SNBI 2022 includes condition assessments for a more detailed set of bridge elements, adopting the bridge element-level inspection system from the
Manual of Bridge Element Inspection (
MBEI) [
7]. While this new specification has the potential to improve the quality of bridge inspections, it may also increase costs. In the
SNBI 2022, 11 bridge elements require assessment, compared with the six components in the
Coding Guide 1995. The
SNBI 2022 also includes the bridge element-level inspection system defined by
MBEI, which attempts to quantitatively associate element defects with one of four condition states. Their definitions take into consideration the bridge element type, element material/construction type, the relevant damage types, and damage measurements. Consequently, this new system will provide more data to engineers for future deterioration tracking and modeling. However, these updates may significantly increase the workload of bridge inspectors. In addition, it is possible that as bridges continue to age faster than they are rebuilt, inspection requirements may become even more detailed. Thus, applying innovative and efficient techniques to assist this practice will be crucial.
For many years, researchers have been exploring the application of artificial intelligence for infrastructure inspection. Here we organize the previous research into four levels: (1) image organization; (2) damage detection; (3) information extraction; and (4) damage quantification. The first level of this application is the rapid organization of inspection data for future retrieval. Saha et al. utilized a CNN to classify building images based on the function of the buildings, e.g., schools or non-schools [
8]. Yeum et al. used AlexNet [
9] to classify post-event building images based on content, including the perspective of the image and the building elements [
10]. Zhang et al. made use of ResNet 50 to classify bridge images based the perspective and the bridge element shown in the image [
11]. Fend et al. developed a GPU-enabled method to classify pavement distress images in real time [
12]. The second level of application is detecting damage from inspection images. The computer vision methods of image classification and object detection have proven successful for this task. Haciefendioglu et al. detected and classified common damage in wooden structures [
13]. Ale et al. utilized object detection to automatically find pavement cracks from roadway images [
14]. Zhang et al. developed a single-stage detector to mark concrete bridge surface damage [
15]. An assembled method utilizing PCA-compressed residual frequency response functions and neural network has been developed by Li et al. to identify damage in civil engineering structures [
16]. At this level, damage is found and presented to an engineer; however, the engineer generally still needs to interpret or quantify the damage. The third level of application includes generating relevant information for further evaluation, e.g., material information. For example, Yuan et al. was able to automatically classify common building materials through 3D terrestrial laser scan data [
17]. Merono et al. utilized image classification techniques to recognize materials and damage on historical buildings [
18]. Bunrit et al. improved the representation of CNN-based features by autoencoder for a task of construction material image classification [
19]. Rashidi et al. tested different machine learning techniques for detecting construction materials in digital images [
20]. The final level of AI application is damage quantification and measurement. In pavement and concrete crack images, one key step at this level is generating a contour of the damage from images. If a contour can be coupled with a reference object to establish scale, the geometric size of the damage can be calculated. Computer vision algorithms and semantic/instance segmentation techniques are usually used at this level and concrete cracks are a common application. Dinh et al. implemented a computer vision-based method for generating the contours of cracks [
21]. Similarly, Dung et al. utilized the deep fully convolutional neural network and achieved the objective of autonomous concrete crack detection [
22]. Song et al. improved the performance of detecting cracks from images by pre-processing and fusing the crack images with the crack range images [
23,
24]. Finally, Choi et al. developed SDDNet which can segment cracks in real time [
25].
Despite all of this work and the fact that this research has significantly promoted the application of AI for infrastructure inspection, a robust, AI-based framework able to complete the tasks that will soon be required by the
SNBI 2022, e.g., assigning condition states to damage, does not yet exist. Thus, this work aims to assist human inspectors by rapidly performing element-level condition state assignments for reinforced concrete bridge decks, as required by the new specification. More specifically, bridge deck cracking was selected to demonstrate the proposed method. Existing AI techniques, including image classification [
26] and semantic segmentation [
27], were used to generate the information needed for condition state assessment. To build confidence in the accuracy of these techniques, a cost-based decision making method was also developed. This method allows engineers to choose the combination of automation and manual inspection for a given bridge inventory that best aligns with their risk tolerance.
This paper presents a practical process of machine-aided condition state assessment using AI. The first section, known as Technical Approach, includes: (1) an introduction to the element-level inspection tasks required by SNBI 2022; (2) the design of an AI-based method for the condition state task; (3) the experimental dataset used for validating the proposed method; and (4) the cost and decision making functions. The second section, known as Experimental Validation, includes the detailed process and results of applying this method to an experimental dataset. In the third section, known as Illustrative Example, the decision making method is demonstrated with a simulated dataset to describe the decision making process for implementing manual versus automated assessments.
2. Technical Approach
2.1. Element-Level Inspections
In 1968, the National Bridge Inspection Program was introduced due to the collapse of the Silver Bridge spanning the Ohio River. Since then, several bridge inspection manuals have been introduced and revised [
28]. In the 1970s, three manuals—the FHWA
Bridge Inspector Training Manual 70, the American Association of State Highway Officials (AASHO)
Manual for Maintenance Inspection of Bridges, and the
Coding Guide 1995 [
6]—were introduced to support bridge inspections [
28]. In 1995, the FHWA made significant updates to the
Coding Guide 1995 to provide each state DOT with guidelines for compiling inspection reports [
29]. This version of
Coding Guide 1995 requires the assessment of the condition of each bridge’s major components using condition rating numbers from 0 to 9. The ratings are required only to indicate the condition of six major components (e.g., the sub-structure, superstructure, deck, etc.), with 9 being a perfect rating and 0 requiring complete bridge replacement. For the remainder of this paper, this condition rating system will be referred to as component-level inspection.
Component-level inspection has drawbacks. Several studies found that it results in ratings with high uncertainty due to the subjective rating definitions [
28]. Additionally, this system focuses only on the major bridge components and neglects assessing more specific bridge elements [
30,
31]. Thus, a new system was developed for more detailed element-level inspections in 2013, outlined in the first edition of the AASHTO
MBEI and required by the new
SNBI 2022 manual. In this manual, the major bridge components are further divided and categorized into minor elements based on material and functionality. For example, the bridge deck (component) can be categorized as a reinforced concrete deck or a pre-stressed concrete deck (elements). For each element, a table is provided to correlate element defects with one of four condition states (CSs): good (CS1), fair (CS2), poor (CS3), and severe (CS4). These tables provide a more specific and quantitative relationship between the damage and the condition states than past guidelines.
As described in AASHTO MBEI, condition states are typically assigned based on damage type, damage size, and any maintenance actions previously taken. For example, to assess cracks on a reinforced concrete bridge, deck inspectors need to record the crack type (normal or map cracks), crack size (insignificant, moderate-width, or wide), and previous maintenance actions (sealed or non-sealed). CS1 is assigned to insignificant cracks or to moderate-width cracks that have been successfully sealed. CS2 is assigned to moderate-width cracks that remain unsealed or unsealed moderate pattern (map) cracking. CS3 is assigned to wide cracks or heavy pattern (map) cracking that may require more significant repairs. CS4 indicates that the cracking could influence the strength or serviceability of the element or bridge. As defined in the manual, an insignificant crack width is smaller than 0.012 inches, a moderate crack width is larger than 0.012 inches but smaller than 0.05 inches, and a wide crack width is larger than 0.05 inches. Thus, to assess the condition state for the cracking defect using AI-based methods, all of this information must be extracted from images and then compared with the tables defined by MBEI.
2.2. Design of the AI-Based Workflow
As described in the previous section, condition states (hereafter, CSs) can be assigned once damage information is extracted from bridge deck images. Computer vision methods can be utilized to automate the extraction of different types of information. Image classification is used to determine image-level (or same level) information, e.g., what the main object is in an image or the pre-set categories to which an image belongs. Semantic segmentation is generally used to determine pixel-level information, e.g., among all of the pixels in an image, the object/category to which each pixel belongs. Thus, semantic segmentation is useful for determining the position and shape of a specific object captured in an image. Based on the complementary features of these two methods, an information extraction workflow was designed specifically for different bridge damage types. In this work, we used reinforced concrete deck cracking to design and demonstrate an AI-based workflow. As summarized in the previous section, the crack type, crack size, and previous maintenance actions must be extracted. The information in CS4 was not considered in this study because the owner will automatically call for manual inspection and this bridge will likely be closed to the public. Furthermore, very few bridge inspection images are available for this stage. Thus, the AI-based method developed here does not apply to decks in this condition. The overall workflow designed for condition assessment is shown in
Figure 1.
Researchers previously developed an AI-based classification schema which successfully classifies bridge images based on the bridge’s components and elements [
11]. Thus, the proposed workflow starts with reinforced concrete bridge deck images, assuming they are already pre-filtered and, thus, only include deck images. In the first step (Step 1), deck images were classified as “Cracked” or “non-Cracked” using image classification. The “Cracked” category includes images that show cracking in the deck, while the “non-Cracked” category includes images with no visible cracking. An image classifier (Image Classifier I) was designed to perform this step. Per description in
Section 2.1, the “non-Cracked” category was then automatically assigned to CS1, while the “Cracked” images required further analysis. Thus, in the second step (Step 2) the “Cracked” images were further classified based on whether the crack shown in the images had been sealed (defined as “Sealed” category) or not (defined as “non-Sealed” category). An image classifier (Image Classifier II) was designed to execute this step. The “non-Sealed” images continued to the next step (Step 3) to determine the type of cracks present. The workflow does not determine the crack type for the “Sealed” category because sealing often covers the crack pattern. Consequently, crack patterns could be wrongly identified if they are in the “Sealed” category. In practice, crack patterns are usually recorded before the sealing action and, thus, if a crack is sealed it is reasonable to assume the crack type is already in the database. To determine the crack pattern (Step 3), a third classifier was designed (Image Classifier III) and used to classify images into the “Normal Crack” or “Map Crack” categories. Images in the “Map Crack” category have intersecting cracks that extend below the surface of hardened concrete, forming a network of cracks (resembling a road map) [
32]. All other cracks were categorized into the “Normal Crack” category. Image examples of each classifier’s categories, taken from the Bridge Inspection Application System (BIAS) [
33] managed by Indiana Department of Transportation (INDOT), are shown in
Figure 2,
Figure 3 and
Figure 4:
In the last step (Step 4) of this workflow, semantic segmentation was used to find the contour of the cracks to estimate their size. Previous research showed how to evaluate the geometric size of an object in an image using this method. Shan et al. developed a method using canny edge detection and Zernike sub-pixel evaluation [
34]. A real-time crack geometric size generation method was developed by Li and Zhao using an encoder–decoder network [
35]. Other researchers, such as Choi et al. [
36] and Liu et al. [
37], continued to improve the practice of the AI-based crack measurement method by adding angle adaptability and portability. Since the images in the dataset built for this work were taken through bridge inspectors’ daily work and, thus, were not specifically taken for crack assessment research using computer vision, the authors elected not to re-develop advanced crack size measurement methods. Instead, crack contours were generated using semantic segmentation and then used to quantitatively analyze how the accuracy of these contours influences the decision of assigning a CS. With the information generated through these four steps, the CS of a crack image and, in aggregate, the CS of bridge deck cracking can be assigned to a reinforced concrete bridge deck.
2.3. Cost Function to Quantify Loss Due to Misclassification of Condition States
Although researchers have developed many metrics to evaluate the performance of their computer-vision methods, AI methods are not expected to achieve perfect accuracy [
38]. For image classification, the measures of accuracy, precision, recall, and f1-scores may be used to evaluate classification results. For semantic segmentation, the measures of pixel accuracy, pixel precision, pixel recall, and intersection over union (IoU) may be used to evaluate performance. However, these conventional AI metrics do not consider different consequences caused by different types of errors. When the final outputs of an AI method are used for a real-world decision, a specific cost table can be formulated to account for these different consequences and evaluate the performance of the AI methods within the context of the problem and workflow.
We note that the overall goal of this workflow is to determine the CS of cracks on a reinforced concrete deck element. Any errors in assigning CSs have varying consequences. In some situations, the misclassification of the CSs may be inconsequential, while in other cases the misclassifications could result in some hazard. For example, if a CS3 (poor) crack is misclassified as CS2 (fair) or CS1 (good), proper maintenance actions for a poor condition crack will not be taken, which may lead to unexpected failures. Considering these situations, a cost function for the final CS prediction, which is composed of estimated real-world costs, is proposed in
Table 1.
The parameters in
Table 1 reflect conceptual costs for the misevaluation of CSs. The parameters are defined as follows:
—This parameter indicates the misclassification of the CS of the crack on the reinforced concrete bridge deck from CS i to CS j while i, j = 1, 2, 3. When i is equal to j, this parameter should be 0 because this situation reflects the fact that the classification of the CS is correct (no cost).
—This parameter refers to the database modification “fee” when a misclassification of the CS occurs.
The specific values assigned to these parameters should be determined at the local level, considering an inspection region’s economy, natural and traffic hazards, maintenance, and inspection fees. Precisely calculating these values is a difficult problem. Thus, in this work the authors selected values to reflect the anticipated ratio of costs for different errors, rather than precise cost values. One representative setting of these parameters is shown in
Table 2.
As defined in
Table 2, the database modification fee is set at
$200. The remaining costs are defined differently for different cases. For critical situations, e.g., when a CS 3 (poor) is classified as CS 1 (good), the penalty is high while for the less consequential situations, e.g., where a CS 1 (good) is classified as CS 2 (fair), the penalty is low.
Once the costs are defined, the cost function can be used to calculate the total loss of utilizing this AI-based CS evaluation method:
In Equation (1),
i (
i = 1, 2, 3) indicates the true CS of the crack on the reinforced bridge deck, while
j (
j = 1, 2, 3) indicates the evaluated CS through the method proposed in this paper.
represents the number of misevaluations having a true label
i but evaluated as
j, while
represents the cost of this misevaluation and is defined in
Table 2.
2.4. Application of Deep Learning-Based Method as a Decision-Making Problem
In the previous section, a cost function was proposed to calculate the cost due to the limitations of this AI-based method. Since the primary attraction of implementing an AI method is to reduce the costs associated with a human’s work [
39], it is important to develop a method to quantitatively analyze whether the inspector should trust any automated method. For every reinforced concrete bridge deck, an inspector must decide to either accept the automated result and its uncertainty, with the risk of overlooked degradation, or ignore the results in favor of manual inspection. The cost of manual inspection should be considered in this decision. Previous research considered how to make this decision for a very difficult problem involving data entry [
39]. Thus, this method assumes that all classification steps prior to the semantic segmentation step are 100% accurate and all risk in this workflow stems from the semantic segmentation step. The ability of a stakeholder to accept risk is often categorized as unprepared, risk-neutral, risk-averse, or extremely risk-averse [
40]. In this work, the extremely risk-averse category indicates that the inspector allocates the entire inspection budget for manual inspection. The risk-neutral category indicates that the inspectors do not require any extra compensation for taking on additional risk. The risk-averse category represents inspectors who like to pursue a balance between the costs and risks. The unprepared category indicates that the inspector accepts all risks associated with the AI-based method and uses it for all inspections.
To analyze this problem, the impact of the semantic segmentation results was considered, which is the last step in the workflow shown in
Figure 1. Thus, each image in this step is assumed to have some cracking. We assumed that in a given region, there were N reinforced concrete bridge decks requiring crack CS assessment and the indices
identify each bridge deck. The manual inspection fee for cracking on one bridge is denoted as k. The total cost of the CS assignment work can be determined using:
In Equation (2), x is the specific image of the bridge deck containing cracking. Here, we assumed that the condition state of the crack was only determined by one image.
is the CS predicted using image
,
is the cost corresponding to the
i-th bridge deck when the predicted CS is
and the true label is
, and
is a binary decision indicating whether to accept the predicted CS. The
value is determined by a pre-set manual check budget
, reflecting the risk preference of the engineer. For example, a risk-neutral inspector will want to minimize the total expected cost, which can be quantitatively represented by:
In Equation (3),
indicates the expectation of the total cost over the random variables. Another example is an extremely risk-averse inspector, who may wish to keep the variance of the cost as low as possible. The variance is the indicator of the risk because the bridge deck inspection images obtained could be different every time. For example, even for the same bridge, changes in the background environment or weather conditions could occur over time. Though the approximate distribution of the images taken in each inspection period should be close, this difference, to some extent, will cause randomness in the results output from an AI-based method. However, this randomness does not influence the crack CS assessment result through the manual inspection herein as the manual inspection is assumed to be 100% accurate. Thus, increased participation for the AI-based method in crack CS assessment work produces larger randomness in the results, which is reflected by the variance of the total cost.
In Equation (4),
indicates the variance of the total cost. The Pareto frontier, which is used in multi-objective optimization [
41], is established to find the suitable pre-set manual inspection budget. To build the Pareto frontier, a list was built containing all budget levels, [
] and
. Here,
is 0, while
refers to a budget large enough to inspect all
N bridges in this region. Due to the randomness mentioned above, for each budget level we sampled
M times to capture this randomness. The expected cost in Equation (3) can then be approximated by:
The variance in Equation (4) can be calculated by:
With the equations for the expected cost and the variance of the cost for each budget level, the Pareto frontier can be plotted to visualize the relationship among the risks (variance), the total costs (expected costs), and the pre-set manual inspection budget. The bridge inspector can then use this plot with the Pareto frontier to determine the budget for the element-level CS assignment for reinforced concrete deck cracking.
5. Conclusions
With the requirements of the newly released SNBI 2022, bridge inspection involves more tasks and these tasks require more resources. Thus, an AI-based workflow was developed and validated to assist inspectors with one of these new tasks: the condition state assignment of a concrete bridge deck cracking. The workflow utilizes three cascading levels of CNN–based binary classifiers to distinguish cracked deck from non-cracked deck images (step 1), sealed crack from non–Sealed crack images (step 2), and normal cracking from map cracking images (step 3). Semantic segmentation is then used to generate the contour unsealed crack images (step 4). Based on the results, each image is assigned a condition state in accordance with the SNBI 2022. Classification accuracies of 0.94, 0.93, and 0.84 were achieved for a dataset of real-world inspection images for steps 1, 2, and 3, respectively, validating these steps of the workflow. The final step yields a calculated IoU of 0.61. This outcome occurred because the images used were not taken for the purpose of applying machine vision and studying cracks, as shown by their much better performance when the methods were used on the Crack 500 dataset that was designed for this purpose. Rather, the purpose of this step was not to optimize the results but to illustrate the overall workflow and the impact of the results on its application when realistic images collected in the field are used.
Inspectors are unlikely to choose full automation, especially in the case of high-risk bridge inspections. Thus, in addition to the workflow, a decision making method was developed which enables bridge inspectors to adjust the inspection budget and their risk preferences to strike the right balance between automated and manual inspection. This tool was used in an illustrative example where we also varied the cost settings and the quality of the semantic segmentation results as discussed above. The influence of these factors on the costs and risks of applying the AI-based workflow was analyzed and discussed.
Based on this study, the authors have several implementation recommendations for bridge inspectors adopting this method. Firstly, high-quality crack images may be taken by inspectors to optimize the performance of the workflow. While the illustrative example showed that even AI models with low performance can help in inspection work, enabling higher performance by taking higher quality crack images will certainly further reduce the costs and risks. Secondly, pre-analysis of the image before applying this workflow is a necessary step to improve overall performance. An inspector should view all outliers and remove those images not suitable for the workflow, while the AI models can still deal with the rest of the massive data. In this way, an effective collaboration is established that can take advantage of both humans and AI to optimize performance. Thirdly, to apply the method in practice different DOTs should collect information on their inspection costs and evaluate their risk tolerance so that they can make use of the method for their needs. Overall, this work illustrated the practical application of a framework for evaluating concrete bridge deck cracking.
Finally, in this work the CS of the crack is only determined by one image whereas, in reality, bridge inspectors may take several images for the same area. Thus, a method for fusing information from multiple images can be developed in the future, which would likely improve the robustness of the method. Additionally, the costs were analyzed based on the assumption that the classification steps were perfect. In reality, although it is a well-established technique, the accuracy of classification steps will also influence the decision of budget setting. Thus, people can also include such influence in the future. However, the basic idea of the decision making process for practical problems is the focus of this paper. Researchers may also expand these functions for other types of damage and other bridge elements.