1. Introduction
The beneficial possibilities of distributed fiber optic sensors (DFOS) for the structural health monitoring (SHM) of infrastructure, such as bridges [
1,
2,
3], tunnels [
4,
5,
6], dams [
7], pipelines [
8], walls [
9], or other engineering structures, are widely acknowledged [
10,
11,
12]. It is anticipated that the maintenance costs are reduced by minimizing direct human interaction with the structure when DFOS are integrated into SHM systems with automated evaluation capabilities [
13,
14]. One of the benefits of coherent optical frequency domain reflectometry (c-OFDR)-based distributed strain sensing (DSS) is the ability to detect local damages [
15,
16,
17,
18]. However, apart from physical influences (which should be accounted for separately), raw DSS data are contaminated with different types of disturbances caused by the measurement principle, which complicate the evaluation by obscuring the sometimes weak signal caused by a structural changes (e.g., damages) [
19,
20,
21]. With respect to real-world use cases, engineers will have to accept low-quality data even with the best possible setup available due to challenging measurement conditions, such as dirt, moisture, ambient vibrations [
22], sensor degradation due to chemical or mechanical influences [
23,
24], transversal pressure, and further local effects [
25]. A reliable evaluation of DSS data requires the reliable elimination of those disturbances caused by the measurement principle [
26], as well as compensation of the physical influences. This study focuses on removing measurement principle-related disturbances from the raw data. The data cleaning step is called pre-processing within this study as it is an upstream process of data evaluation for the structural information extraction workflow.
While evaluating data, it was observed that the pre-processing capabilities at hand, tailored towards DSS data, were insufficient. The process required manual intervention, an iterative procedure (trial and error), or eyeballing. As pre-processing is an integral part of the data evaluation workflow, the aforementioned benefits can only be achieved if pre-processing can be carried out reliably and automatically. To tackle the challenge of DSS data pre-processing, the following research questions are investigated. Each research question is examined in one of the remaining sections.
Which types of disturbances exist? What are their potential causes and characteristics? How are they dealt with by the DFOS community? These questions are answered through a literature review in
Section 2.
How can the pre-processing of DSS data be generalized? In
Section 3, pre-processing tasks are categorized, and a modular, flexible pre-processing workflow is presented.
Which advanced pre-processing algorithms exist? In
Section 4, selected pre-processing algorithms are presented, their respective advantages and disadvantages are discussed, and modifications are proposed.
How reliable are the selected pre-processing algorithms on DSS data acquired in crack monitoring scenarios? In
Section 5, the benchmark prerequisites are presented. The benchmark was carried out on artificial data, taking into account different scenarios. A simple functional model—based on the DFOS sensitivity—is introduced to simulate crack-induced strain peaks. Performance measures for the pre-processing are presented. In
Section 6, the benchmark results are discussed.
2. Measurement Disturbances in DFOS Data
This section provides an overview over three different disturbance types—strain reading anomalies (SRAs), dropouts, and noise—present in DSS measurements. A schematic of the disturbances is shown in
Figure 1. For each disturbance type, causes and specific characteristics are described, followed by an overview of coping strategies actually employed by other researchers. Although the causes and characteristics of the disturbances are discussed specifically for the c-OFDR-based optical distributed sensor interrogator (ODiSI) manufactured by Luna Inc. (Roanoke, VA, USA), the coping strategies should be transferable to other DSS methods.
2.1. Strain Reading Anomalies (SRAs)
2.1.1. Characteristics and Causes
SRAs are misreadings, characterized by outliers of implausible high or low values; see
Figure 1. SRAs are caused by a failed cross-correlation between the reference and measurement signal in the ODiSI [
27]. There are two types of SRAs. (i) Harmless strain reading anomalies (HL-SRAs) are isolated glitches in the strain readings in space and time. (ii) In contrast, harmful strain reading anomalies (HF-SRAs) persist (in time) after their first occurrence.
2.1.2. Coping Strategies
It is recommended to convert SRAs to dropouts [
19,
27]. The conversion of a SRA to a dropout is a form of outlier exclusion, which is common practice in data science (replacing an “obviously wrong” value with an “unknown” value). This SRA elimination can be divided into three steps: (i) Possible SRA candidates are identified in the first step. Ideally, no false negatives (SRAs categorized as normal data) remain, at the cost of some false positives (normal data categorized as SRAs). (ii) False positives are eliminated in the second step—candidate verification. (iii) In the last step, identified SRAs are replaced with not a number (NaN). Afterwards, the resulting dropout can be dealt with as discussed in
Section 2.2.
The ODiSI manufactured by Luna Inc. implements the conversion of SRAs to dropouts with spectral shift quality (SSQ) as a reliability/outlier measure [
21,
22,
28]. SSQ evaluates the plausibility of the cross-correlation between the spectra of the reference reading and the current measurement [
29,
30,
31,
32]. If the SSQ falls below a certain threshold, the reading should be considered an outlier (SRA) and ignored in further processing steps (set to NaN). Different SSQ thresholds are recommended (depending on the use case), e.g., <0.2 [
29] or <0.15 [
31]. However, SSQ has been shown to be an insufficient measure in [
27,
31]. In earlier series of the ODiSI, SSQ could be exported along with the raw measurements and the threshold was configurable [
28], which is no longer possible with the ODiSI 6100 series [
22]. The SRAs contained in the exported data are exclusively false negative classifications obtained using the SSQ method.
Other SRA elimination approaches for DSS data employ outlier detection algorithms, such as (i) based on a local outlier factor [
32], (ii) the geometric threshold method (GTM) [
19], (iii) the polynomial interpolation comparison method [
19], (iv) based on the deviation from sliding median filtering [
33], (v)
z-score-based algorithms [
21], or (vi) the outlier-specific correction procedure (OSCP) proposed for the processing of surface scans [
34,
35].
2.2. Dropouts
2.2.1. Characteristics and Causes
Dropouts are missing values represented by NaN in the exported DSS data; see
Figure 1. A reading becomes a dropout if it is flagged as anomalous by the ODiSI’s built-in SRA detection method (SSQ) [
21]. However, large total deviations from the reference state or large gradients cause false positives. According to the manufacturer’s manual [
22], the limit for the absolute strain difference compared to the reference state is approximately 12,000
. The permissible strain difference between two readings is specified in [
28] as 517
for a gage pitch of 1.25
and 129
for a gage pitch of 5
. Exceeding these limits results in dropouts. The influence of dropouts on c-OFDR-based crack monitoring is discussed in [
36].
2.2.2. Coping Strategies
There are two strategies to cope with the aforementioned limitations: (i) avoid the technical conditions leading to dropouts and (ii) reconstruct missing values by interpolation during the pre-processing stage.
The a priori approach is to avoid exceeding the technical limits of the ODiSI in the first place. Here, the strain transfer between the host material and the optical fiber becomes crucial [
37,
38,
39]. The DFOS type and installation method must always be selected according to the specific measurement task. In [
36], an empirical method is proposed for the design of DFOS monitoring systems, accounting for the DFOS type, installation method, crack pattern, and measurement settings.
The a posteriori approach is to eliminate dropouts from the raw data during the pre-processing stage by (i) replacing them with estimated values using interpolation [
20] or by (ii) excluding the dropout readings from data (which might be equivalent to interpolation depending on the later evaluation steps) [
16].
Several approaches for interpolation can be applied. Linear interpolation between the remaining valid readings along one axis (either space or time) is the standard approach [
15,
16,
19,
20,
29]. However, linear interpolation results in a drastic underestimation of the crack widths if the tip of the strain peaks is missing due to the technical limits of the ODiSI. Interpolation with higher-degree polynomials are discussed in [
27]. However, polynomials can approximate complex strain profiles only locally due to their trend towards positive or negative infinity when the argument increases indefinitely. Piece-wise polynomial functions—so-called splines—can overcome this limitation of polynomials. One specific spline type—
Akima splines—have the advantages to not require assumptions, are fast, and do not show overshooting behavior [
40]. Hence,
Akima splines are used to interpolate dropouts in [
41].
DSS data sets are time-series of one-dimensional spatially distributed data. Due to the incompatible units (
and
, there is no meaningful common metric of the space and time. Hence, two-dimensional interpolation methods (e.g., as proposed in [
34] for two-dimensional spatial data using distance weighting) are not directly transferable to DSS data.
2.3. Noise
2.3.1. Characteristics and Causes
Noise is a high-frequency, low-amplitude component usually present in measurement data [
34,
42]; see
Figure 1. Noise should not be confused with local effects, such as the “concrete signature” caused by inhomogeneities and variations in
Young’s modulus within the concrete matrix [
43] or local strain variation due to bond interactions [
25,
33,
44]. Noise fluctuates in both dimensions (spatial and temporal), in contrast to those local effects, which are relatively stable in time. The intensity and characteristics of noise depend on several factors, such as the quality of fiber connectors, splices, or fibers; measurement settings and the operational mode of the ODiSI, [
22]; vibration [
45]; or loss (e.g., caused by small bending radii) [
46].
2.3.2. Coping Strategies
Noise is commonly reduced by applying low-pass filters [
47,
48,
49]. Examples for low-pass filters used for smoothing DSS data are the sliding average [
25,
33,
50,
51], the sliding median [
52], locally weighted scatterplot smoothing (LOWESS) [
53],
Butterworth filters [
20,
54],
Bessel filters [
20],
Savitzky-Golay filters [
20,
49], wavelet filtering [
55], and extremal filtering [
56]. However, filtering might be of questionable benefit or even detrimental (by distorting the signal), especially when applied excessively [
33,
47]. Another approach is to apply aggregate functions to combine several readings when reducing data size, also called “ensemble averaging” [
20,
39,
47]; see
Section 3. However, ensemble averaging reduces the effective measurement rate or spatial resolution [
55,
56].
3. A Generic Workflow for Pre-Processing of DFOS Data
An appropriate pre-processing strategy depends on the data at hand and the goal of the data evaluation. Hence, the individual pre-processing steps may change for each application. Therefore, a generic pre-processing workflow should fulfill the following requirements.
Firstly, it should be possible to freely select and configure algorithms for each step independently from each other. Secondly, the workflow should allow us to carry out the chosen algorithms in an arbitrary order. Thirdly, the workflow should respect that the data array’s shape (number of rows and columns) might be changed by any of the algorithms. Finally, each algorithm should be able to deal with both one-dimensional data (distributed in time or space) and two-dimensional data (distributed in time and space). The operation of inherently one-dimensional algorithms on two-dimensional data should be applied in a transparent way along one axis—row-wise (along space domain) or column-wise (along time domain).
To meet the previously stated requirements, the free open source software framework
fosanalysis implements from its version v0.4 onwards (available at
https://github.com/TUD-IMB/fosanalysis, published and accessed 18 November 2024) the concepts of task objects and workflow objects to achieve a flexible pre-processing pipeline. The This concept is visualized in
Figure 2.
Task objects implement one specific algorithm, its methods, settings, and parameters. Task objects are interchangeable because they expose the same interface. The group hierarchy of the pre-processing tasks is shown in
Figure 2a. The terminology for pre-processing tasks is as follows:
- Masking
algorithms identify SRAs using anomaly detection. For each element, a decision is made of whether this element is considered anomalous (SRA) or a normal data point. The values of elements considered anomalous are replaced with NaN. The data array’s shape (number of rows and columns) is not changed. In this paper, three SRA detection algorithms are presented: (i) the GTM discussed in
Section 4.1, (ii) the OSCP discussed in
Section 4.2, and (iii) the sliding modified
z-score (SMZS) discussed in
Section 4.3.
- Repair
algorithms replace dropouts with values estimated from the surrounding data, usually by interpolation. The data array’s shape (number of rows and columns) might be changed. Linear interpolation and spline interpolation for dropout reconstruction are addressed in
Section 6.3.
- Filtering
algorithms modify the values of the data to smooth the signal by reducing high-frequency components (low-pass filters). The data array’s shape (number of rows and columns) is not changed. In this paper, three filter algorithms are presented: (i) the sliding average, (ii) the sliding median, and (iii) the cluster filter; see
Section 4.4.
- Resizing
algorithms change the data array’s shape. They can be divided into three groups: (i) Downsampling is used to reduce the data volume by either taking only every
mth element or consolidating several elements into a single element using aggregate functions (e.g., from two-dimensional to one-dimensional or even to a singleton) [
47]. (ii) Resampling is used to obtain readings at different sampling points (in time and/or space) compared to the original sampling points by means of interpolation. (iii) Cropping is used to restrict the data to a specific area (i.e., cut away uninteresting parts).
Note that algorithms can have effects associated with multiple groups, e.g., downsampling will also reduce the noise or a filtering algorithm might cancel out SRAs too.
Workflow objects are containers that implement the pre-processing pipeline and coordinate multiple task objects; see
Figure 2b. A workflow object holds the order of the tasks and calls them sequentially. This container will pass its input to the first task object and pipe the output of one task object to the input of the next task object. Because the tasks are interchangeable, any sequence of pre-processing tasks can be established. An example for such a workflow is shown in
Figure 2c.
While the actual sequence of the pre-processing steps might depend on the data set and the use case, particular sequences make sense:
SRA removal should be followed up by dropout elimination.
SRA removal should be carried out before smoothing.
Downsampling with the average should be preceded by SRA removal.
Downsampling with the median might not need to be preceded by SRA removal since it is more stable against outliers and might stabilize the signal instead.
Depending of the density of the target points, resampling might be preceded by a combination of the other pre-processing steps.
4. Pre-Processing Algorithms
This section presents three masking algorithms (for SRA removal): (i) GTM in
Section 4.1, (ii) OSCP in
Section 4.2, (iii) SMZS in
Section 4.3. It also presents one filtering algorithm (for smoothing): the cluster filter in
Section 4.4. Implementations for all four algorithms were made available with
fosanalysis. Each of the following subsections is structured into three parts: (i) a presentation of the algorithm’s original concept, (ii) a discussion of the algorithm’s advantages and disadvantages, and (iii) a presentation of modifications made to the algorithm to overcome the disadvantages.
Note that zero-based numbering (as is common in the Python 3 programming language) is used for all algorithm descriptions. Assignment statements are denoted by a left pointing arrow (e.g., means “the variable a is set to 0”, and means “the value of a is incremented by one”), while equality comparison statements are denoted by the equals sign (e.g., denotes a check, whether the variables a and b hold the same value).
All presented algorithms operate on one-dimensional strain data. The operation of two-dimensional strain data (time-series of spatially distributed measurements) is supported by carrying out the one-dimensional operation for each row or column independently. The algorithms can be applied both spatially (i.e., along the DFOS’ length for one specific time stamp) or temporally (i.e., on a time-series of a specific position in the DFOS). However, the relation along the DFOS’ length is governed by mechanical laws, whereas there is no intrinsic relationship between consecutive measurements at different times. Hence, the space-axis is recommended as the primary operational axis.
4.1. Geometric Threshold Method (GTM)
4.1.1. Original Concept
The geometric threshold method (GTM) is an algorithm intended to detect and neutralize SRAs based on the comparison of local strain increments [
19]. A flowchart representation of the GTM is shown in
Figure 3a; additions to the original concept are highlighted in gray. The original (raw) strain array is denoted as
, and the processed array is denoted
. The GTM traverses the strain array with
n elements and compares the candidate element with index
i to the current reference element with index
k. The initial reference element is the first element of the array. Elements that are dropouts are skipped. The candidate is accepted as the new reference element if the absolute strain increment
from the current candidate element to the reference element does not exceed the threshold
. Else, the candidate element is considered an SRA and set to NaN in the processed strain array. The modifications of a forward neighbor comparison (FNC) and reverse sweep (RS) are described in
Section 4.1.3.
4.1.2. Advantages and Disadvantages
The GTM has the advantages of (i) being a simple algorithm, with a linear time complexity class and (ii) featuring a threshold with a physical meaning. Hence, the GTM is fast and easy to understand, and its results can be manually verified.
A disadvantage of the original GTM version is that it struggles to recover from large, but legitimate increments in steep strain gradients (which are encountered, for example, in the slopes of crack-induced strain peaks). Since the reference element is updated only when the candidate’s strain value is again within , in strain data with large signal amplitudes, this condition might be fulfilled again only after a considerable distance or not at all. As a result, extensive parts of the processed data might be incorrectly marked as SRAs and replaced with NaN.
4.1.3. Modifications
To overcome the aforementioned flaw, two modifications are introduced. The first modification—reverse sweep (RS)—is a backtracking mode, which is triggered if the previous reference element and the new reference element are not direct neighbors after accepting a candidate as a new reference element. In this case, the “normal” forward sweep is paused, and the strain array is traversed in the reverse direction, starting from the new reference element towards the old reference element. When the old reference point is reached, the forward sweep is resumed.
The original algorithm uses backward-directed comparisons (to previously encountered elements) only. The second improvement—forward neighbor comparison (FNC)—is an additional verification step. A flowchart representation of the FNC is shown in
Figure 3b. The FNC is triggered when a candidate is initially flagged as an SRA and compares the strain increment to the elements ahead of the candidate element with the index
i. In order to handle multiple adjacent SRAs, a configurable number of neighbors
are taken into account. Here,
p is the number of neighbors exceeding the strain increment to the candidate:
. The FNC is successful and the candidate is reaccepted if the ratio
does not exceed the configurable tolerance ratio
t. It should be noted that the FNC makes RS mostly dispensable because elements to be reaccepted by RS would already be accepted by FNC.
4.2. Outlier-Specific Correction Procedure (OSCP)
4.2.1. Concept
The outlier-specific correction procedure (OSCP) is an algorithm for outlier identification on regular rasterized topological data [
34,
35] and can be used within the scope of DSS to detect and neutralize SRAs. The algorithm is structured into two phases: (i) candidate detection and (ii) candidate verification. A flowchart representation of the OSCP is shown in
Figure 4.
In the first stage, outlier candidates are detected based on the the elements’ relative height , which is defined as the absolute difference between the original value and the value of the central sliding median with its window’s inradius r. After the array of relative heights, is calculated in a vectorized manner, and the cumulative density function is established. For reasons of numerical stability, is resampled, e.g., using equidistantly spaced quantiles of 1%. The deduction of the threshold is based on two assumptions: (i) Outliers have a large relative height compared to normal data points. Hence, the threshold has to be located in the upper half (). (ii) SRAs are few in comparison to normal data points with large distances in . Hence, the threshold has to be located in the asymptotic part of , i.e., where the curve is “flat enough”. The second condition is represented by , with the flatness given as the reciprocal of the cumulative density function’s first derivative and the flatness level L configuration parameter.
An element is flagged as an outlier candidate in the array of boolean values if its relative height exceeds the threshold: . The steps described above (calculation of relative height, determination of the threshold, and candidate flagging) are carried out for variable inradii to detect different-sized outlier clusters. The inradius of the largest window is a configuration parameter that determines the largest detectable outlier cluster. An element enters the verification stage as a candidate if it was flagged in at least one pass among all inradii.
The second stage—outlier verification—is based on the absolute strain increment between neighboring elements
. The estimation of the strain increment threshold
with the cumulated density function
is analogous to the first stage. Then, the strain array is divided into groups of contiguous neighboring elements. The set of group boundaries
B contains the indices of the groups’ last elements, found where the strain increment
is larger than the strain increment threshold
. While the process of grouping is straightforward for 1D, it is more intricate for 2D and not described in [
34,
35]. Grouping in 2D is carried out as follows. Firstly, the steps from the strain increment calculation through group boundary detection are carried out separately for each direction. In 2D, indices are an ordered pair. Initial primitive groups are built by iterating over the array (still separated for each dimension) row-wise and column-wise. Before adding the current index to the current group, the conditions for the beginning of a new group are evaluated: (i) the current index is contained in the set of group boundaries
B or (ii) the end of the array in this direction is reached and the iteration resumes with the first element of the next row or column. Then, the primitive groups are merged when they contain at least one common index, until only groups with pairwise distinct indices are left. Note that this results in non-rectangular shaped 2D groups.
Finally, elements belonging to groups consisting of only candidates are confirmed as SRAs and converted into dropouts, replacing their values with NaN. Candidates in mixed groups (consisting of normal elements and of candidates) are considered normal elements.
4.2.2. Advantages and Disadvantages
The advantageous properties of the OSCP are that it (i) can deal with clustered SRAs, (ii) is direction agnostic, and (iii) supports 2D operation, since it does not depend on a common metric for space and time.
Disadvantageous properties of the OSCP are (i) its complexity with relatively high numerical cost (governed by the repeated sliding median, repeated sorting for the cumulated density function, and the group merging necessary in the 2D case), (ii) the unintuitive flatness level parameter L, and (iii) possible instabilities in the threshold estimation because the empiric cumulated density function is not a smooth function, even with resampling.
4.2.3. Modifications
The modifications to the original description presented in [
34] are as follows. (i) The cumulative density function
C is calculated without the intermediate histogram, and
C is resampled afterwards with a configurable number of quantiles. (ii) In the original use case for surface scanning, both directions have the same unit, and the increments in both directions can be put into the same cumulated density function
. This is not reasonable for time-series of spatially distributed measurements due to the different meaning of time and space. To support direct 2D operation, the group boundaries are determined for each dimension separately. Hence, the steps from the calculation of
through the initial group are carried out for each dimension separately. (iii) The threshold can be set directly, which enables us to skip the determination via the cumulative density function.
4.3. Outlier Detection Based on the z-Score
4.3.1. Concept
The
z-score is a statistical measure for the distance of sample values from the sample center. There is a family of simple outlier detection algorithms based on the concept that a value is considered an outlier if its
z-score exceeds a given threshold
. Being an outlier measure, the
z-score can be used to identify and neutralize SRAs. A representation of the general algorithm is shown in
Figure 5. There,
denotes the original (raw) strain array, and
is the processed strain array.
Four algorithms are presented, which differ in the calculation of the
z-score. The basic
z-score (BZS) uses the arithmetic average
as the sample center and the standard deviation
as the unit [
57].
The modified
z-score (MZS) uses the median
as the sample center and the median absolute deviation (MAD)
as the unit instead of the standard deviation
. To obtain results comparable to those of the BZS, the factor 0.6745 is used because the expected value of the MAD is
for large sets of normally distributed data [
57].
The
Whitaker–Hayes z-score (WHZ) is a modification of the MZS, which uses the array of increments between values
instead of the original values [
58]. This approach has the benefit that the influences of trends and low-frequency changes in the data are eliminated.
The sliding modified z-score (SMZS) (a modification of the MZS) uses the relative heights
as a basis [
21]. The relative heights are given by the difference between the central element
and the median of elements in the surrounding window with the inradius
r. When the mean absolute deviation is used as a fallback option, the correction factor 0.7979 is used instead of 0.6745 because the expected value for the mean absolute deviation is
for large sets of normally distributed data [
59].
4.3.2. Advantages and Disadvantages
Advantageous properties of the z-score-based methods are (i) their algorithmic simplicity and low computational effort with linear time complexity , (ii) that the z-score highlights outliers by exaggerating how much an outlier stands out from the normal data, and (iii) that outliers are recognized in relation with the natural variability of the data. Apart from that, the strengths and weaknesses of the algorithm are governed by the strengths and weaknesses of the calculation rule for the z-score array.
The BZS has the advantage that it is a common statistical measure. However, a weakness is its vulnerability to the distorting effect that individual outliers have on the z-score of all members (caused by the use of the average). Because the BZS takes all members of a sample into account, its use is restricted to stationary data. This makes the BZS unsuitable for non-stationary data, such as DSS data of cracked concrete specimens, where the signal changes over a wide range of values.
The MZS is more robust against outliers due to the use of the median. However, since the MZS uses the complete sample, it is not suitable (like the BZS) to use with DSS data exhibiting large signal variation.
The WHZ solves this problem by operating on the increment. This comes at the expense of false positives, since the z-score of normal data adjacent to SRAs is also anomalous. Another weakness is that adjacent SRAs with similar values are not detectable. This might lead to problems in the presence of anomalous areas.
The SMZS intended to solve this problem by taking only the local vicinity into account. The disadvantage is the introduction of an additional parameter r for the sliding window, which dependents on the data properties.
4.4. Cluster Filter
4.4.1. Concept
The cluster filter is an iterative distance-weighted filtering algorithm [
42]. It is intended to smooth data by means of outlier resistant noise reduction while preserving local features. A flowchart representation of the cluster filter algorithm is shown in
Figure 6. For each element
k for all
n elements in the strain array, with its position
and original strain value
, the smoothed value is estimated in an iterative process. Here,
n denotes the total number of elements in the one-dimensional array. The influence weight
of another element identified by the index
on the element
k falls off exponentially with the squared
Euclidean distance
multiplied with negative
. The falloff factor
is the main configuration parameter and influences the scale of the cluster filter. The initial estimate is calculated as the weighted average over all elements. In each iteration step
t, an intermediate parameter
is calculated representing the local variance before proceeding to calculate the next estimate
. The iterative estimation is guaranteed to converge and is stopped if the change in estimation between two consecutive iteration steps falls below the threshold
, the second configuration parameter. Then, the value of the final iteration value is stored in the processed strain array
, and the algorithm moves on to the next element.
4.4.2. Advantages, Disadvantages, and Modifications
An advantageous property of the cluster filter is its robustness against SRAs, as their influence is demoted with the parameter . However, the cluster filter’s high computational cost with a quadratic time complexity class is a disadvantage.
The algorithm was modified to replace the weight and strain values of dropout elements with 0. This enables the filter to yield decent interpolation for singular or very short dropout fields.
5. Benchmarks and Quality Measures
The following benchmarks are used to assess the performance of the presented algorithms to remove c-OFDR measurement principle-related disturbances from DSS data. This section presents the fundamentals of those benchmarks—firstly, the data in
Section 5.1, and secondly, the accuracy measures in
Section 5.2.
5.1. Benchmark Data
Because all real data contain disturbances (related to the measurement principle and physical effects), the ground truth of real measurements cannot be known. In such a case, pre-processing algorithms could only be compared against a reference algorithm—which would naturally perform best in this benchmark (when compared to itself). The benchmarks need to be carried out in a controlled and repeatable environment for a fair performance comparison. Hence, an artificial data set is synthesized, which enables us to compare the result of the pre-processing algorithms to the “true” signal. The “true” signal—hereafter called clean signal (described in
Section 5.1.1)—simulates different monitoring use cases. The simulated disturbances with realistic properties—described in
Section 5.1.2—enable the transfer of the results to real monitoring data sets.
The disturbed strain array
—simulating the raw DSS data—is composed as the element-wise superposition of the following components: (i) clean signal
, (ii) noise
, (iii) SRAs
, and (iv) dropouts
:
The composition of the clean signal
is described in
Section 5.1.1, and the synthesis of the disturbances is presented in
Section 5.1.2. A plot of the benchmark data is shown in
Figure 7, and the data set was made available as open data [
60]. Note the different scaling of the vertical axes.
5.1.1. Clean Signal
To simulate different use cases commonly encountered with DFOS-based DSS monitoring, five different scenarios are generated: (a) constant zero signal, (b) ramps with abrupt changes in the level of the signal, (c) crack pattern with weak bond between the concrete matrix and the DFOS, (d) crack pattern with normal bond between the concrete matrix and the DFOS, and (e) crack pattern with stiff bond between the concrete matrix and the DFOS.
The first scenario (see
Figure 7a) simulates a measurement with a stationary signal, encountered in both the space and time domains. In the space domain, such a flat strain profile is encountered in an uncracked concrete specimen (after taring). In the time domain, constant load results in such a stationary time-series for one gage. Hence, the clean signal is the zero vector.
The second scenario (see
Figure 7b) simulates ramps and steps encountered with, e.g., varying cross-sectional geometries in steel construction. The spikes and jumps are deliberately modeled to exaggerate potentially difficult cases for the algorithms. Hence, it is expected that this scenario is difficult for the algorithms, resulting in a high number of misidentifications.
The last three scenarios (see
Figure 7c–e) simulate strain curves obtained in the use case of the DFOS-based crack monitoring of concrete structures. The crack positions and widths are fixed. From
Figure 7c–e, the bond stiffness between the DFOS’ core and the substrate increases, resulting in increasingly pronounced strain peaks. The bond stiffness can be described by means of the DFOS sensitivity [
61]:
with the crack width
and the maximum strain reading
. The higher the sensitivity, the stiffer the bond between optical fiber and substrate [
61]. Values for this benchmark are chosen based on the results of preliminary studies [
15,
36]. A weak bond is simulated with
, which is associated with a layered robust DFOS. A medium bond is simulated with
, corresponding to a monolithic robust DFOS. A stiff bond is simulated with
, corresponding to a filigree DFOS with an acrylate coating.
The clean signal
for the strain profiles (
Figure 7c–e) is the superposition of several peak shape functions. Due to its resemblance to strain peaks in real DSS data, a
Gaußian function is chosen as the base peak shape function
with the expected value
and standard deviation
. Considering Equation (
2),
becomes
By inserting Equation (
4) into Equation (
3), replacing the expected value
with the position of the crack
, and scaling with the crack width
, the strain function for the
kth crack
becomes
Finally, the clean signal is synthesized by superposing all strain peaks for all
k cracks:
Equation (
6) is applied in a vectorized manner for the positional array
, which is generated in the range of 0 m to 3 m with a spatial resolution (gage pitch) of 0.65
. The parameters
,
, and
for different
values are given in
Table 1.
5.1.2. Disturbance Parameters
The parameters for the disturbances—noise, SRAs, and dropouts—are chosen to simulate measurement data of mediocre quality. The noise component is simulated with a standard normal distribution with the expected value and a standard deviation drawn as a consecutive sequence with a random number generator.
SRAs are simulated using a two-step process. Firstly, the positions of the SRAs are randomly distributed through a
Bernoulli trial using a 5% probability for an element to be an SRA. Then, the SRA values are sampled randomly from a data set extracted from real experiments [
60,
62] and added to the signal.
An element is a dropout if any of the following criteria is fulfilled. (i) The element is chosen in a
Bernoulli trial with a dropout rate of 10%. (ii) The absolute value exceeds
, simulating the ODiSI’s absolute measurement range [
22]. (iii) The absolute strain increment exceeds
, simulating the ODiSI’s technical limit for the measurable the strain gradient. This strain increment threshold is related to the ODiSI’s maximum measurable strain gradient of 1480
/
for a gage pitch of 0.65
[
36]. Values of dropout elements are replaced with NaN in the strain signal. Note that the dropout takes precedence if a reading is both a dropout and a SRA.
Noise, randomly distributed dropouts (due to the Bernoulli process criterion) and SRAs are fixed across all scenarios for comparability, i.e., the values are generated once and added to all scenarios. However, the dropouts simulating the technical limits of the ODiSI differ between the scenarios.
5.2. Performance Measures
5.2.1. SRA Detection Accuracy Measures
The SRA detection accuracy is estimated based on the rate of false positives (normal data classified as SRAs) and false negatives (SRAs classified as normal data) in relation to the correct number of SRAs. Because the confusion matrix does not take the severity of the misidentification into account, another measure is developed: the weighted confusion measure (WCM). The WCM is a penalty function that reflects both frequency and the impact of misidentifications. Hence, the WCM is higher as more misidentifications occur and with larger identification mistakes. The WCM has two components: one for false positives () and one for false negatives ().
Regarding false negatives, the larger the missed SRA, the larger is severity of the mistake. The penalty for a false negative at the index
i is
the SRA-value added to the clean signal, which is known for the synthetic benchmark data. Then,
is the sum over all penalties:
For false positives, the opposite is the case: the false positiveis the worse, the more similar an element is to its surrounding. However, since
false positives, an additional penalty value
is introduced. It is compared to the actual absolute average strain increments
and
to the left and right neighbors of the noisy strain signal’s element. The penalty for a false positive is the absolute difference between
and the average of those two strain increments. Then,
is the sum over all penalties:
To avoid artifacts,
should be set about the largest expected false positive SRAs. In the following benchmarks,
is used.
5.2.2. Filter Quality Measures
The filtering accuracy is assessed based on the deviation between the clean signal
and the processed signal
with the positional index
i and the total number of gages
n. Then, the mean error (ME) is
Another measure—which emphasizes larger derivations—is the root mean square error (RMSE):
5.2.3. Crack Width Error Measure
The accuracy of dropout reconstruction is assessed with the relative crack width estimation error
with
being the crack width integrated from the pre-processed (reconstructed) strain peak and
being the reference crack width.
6. Results and Discussion
This section presents the carried out benchmarks and is divided into four subsections. The three first subsections are dedicated to SRA detection algorithms (
Section 6.1), noise reduction (
Section 6.2), and the reconstruction of strain peaks with dropouts (
Section 6.3). In each of those subsections, the specific benchmark setup is described followed by a presentation and discussion of the results. General limitations of the benchmark are pointed out in
Section 6.4. All benchmarks were carried out using
fosanalysis.
6.1. SRA Detection Accuracy
In this study, the SRA detection algorithms were applied on the benchmark data and their performances were evaluated.
6.1.1. Setup
Figure 8,
Figure 9,
Figure 10,
Figure 11 and
Figure 12 compare the algorithms GTM, OSCP, and SMZS in terms of their accuracy for the five benchmark scenarios examined, as shown in
Figure 8,
Figure 9,
Figure 10,
Figure 11 and
Figure 12. The
x-axes show the threshold value (according to the respective definition). The threshold is varied in the intervals (i)
for the GTM, (ii)
for the OSCP, and (iii)
for the SMZS. For the GTM, FNC with a range of
and a tolerance
and RS were activated. The OSCP and SMZS additionally depend on the radius of the underlying sliding windows, which is varied
. The legend is used for both the OSCP and SMZS. The
y-axes show the WCM components for false negative
and false positive
misidentifications depending on the configuration, with lower values for
and
indicating a better accuracy. The logarithmic
y-axes allow for a direct comparison. To circumvent the problem that 0 cannot be displayed in a logarithmic scale, 1 is added to all values as the base line.
6.1.2. Scenario Comparison
The optimal algorithm configuration is dependent on the data set [
21]. Hence, the different scenarios are used to identify the strengths and weaknesses of the algorithms. The trade-off between the rate of false negatives to false positives is seen in the general tendency that higher thresholds result in smaller
but larger
.
The results in scenarios (a) and (c) (see
Figure 8 and
Figure 10) are very similar. Both scenarios (a) and (c) have a low variance in the signal, so the SRAs stand out clearly, making them the easiest scenarios for the SRA detection. All algorithms perform well, and there exists a configuration range for perfect results with neither false positives nor false negatives for the GTM and SMZS but not for the OSCP in scenario c, where the OSCP reports more false positives.
The results for the both crack scenarios (d) and (e) (see
Figure 11 and
Figure 12) are closely related. With a higher variance in both scenarios (d) and (e), no configuration can be found for neither false negatives nor false positives. Still, scenario (e) is more difficult due to the dropouts and steeper gradients, resulting in a slightly tighter configuration leeway. In the plot for the GTM, the value used to simulate the technical strain gradient limit is clearly visible represented in the drop in
.
Scenario (b) (see
Figure 9) stands out as the most difficult one among the five scenarios, and the algorithms have a considerably higher
and
. There is no configuration yielding perfect or near-perfect results. The cause for these difficulties are the steep ramps and sharp tips of the triangular ramps. Here, the random dropouts in the ramps impact the OSCP and SMZS. The large variance in the ramps reduces the
z-score of the SRAs in the SMZS. Avoiding enormous amounts of false positives in the ramps with the GTM requires
to be larger that any genuine strain increment. Hence, scenario (b) would require an infeasible high threshold, preventing the GTM from identifying smaller SRAs.
6.1.3. Algorithm Comparison
With a larger radius for the OSCP and SMZS, fewer false negatives but more false positives are generated. However, the sensitivity to the radius seems to be higher for the number of false positives than the number of false negatives. The explanation for this is that the benchmark data contain HL-SRAs only and no anomalous areas. Tips of peaks are more likely flagged as SRAs the larger the radius is.
The presence of dropouts in the vicinity of SRAs generally increases the probability of misidentifications and makes configuration more difficult. Hence, the largest influence due to dropout is encountered in scenarios (d) and (e). The classification of the OSCP and SMZS is unreliable for the r gages neighboring wide continuous dropout fields because the reduced number of values makes the sliding median less robust. Hence, locally extreme values (e.g., in the flanks) become false positives, and SRAs embedded in the dropout field cannot be detected and become false negatives.
GTM
Compared to the other SRA detection algorithms, the GTM has the advantages (i) that it has only one primary configuration parameter (), which has a physical meaning (intuitive to understand) and a comfortable configuration corridor; (ii) that it retains regular data at the edge of compact dropout fields; (iii) that it should be able to detect even HF-SRAs with appropriate settings; and (iv) its low computational requirements. However, it struggles with steep inclinations (seen in scenario b). The lower limit for the threshold is the technical limit of the ODiSI, up to which strain increments should be considered as regular. The upper limit is given by the lowest SRAs contained in the data set. Hence, the threshold should be set to about , which proved reliable for all scenarios.
OSCP
For the most scenarios, the OSCP takes the middle ground between the GTM and SMZS, but has slightly better results than the other two algorithms in scenario (b). Compared to the other two algorithms, the OSCP’s advantages are (i) the data-driven adaptive threshold with rich configuration options, which is not as intuitive to understand; (ii) that it can identify clustered SRAs; and (iii) it being native in 2D operation mode, which could find more HF-SRAs but at the cost of a serious performance impact.
However, the OSCP has weaknesses in that it tends to flag normal data in the vicinity (up to ) of compact dropout fields in peak flanks, cf. scenarios (d) and (e). This behavior is due to the fact that the median is distorted and the values close to the dropout field are the most extreme ones in the sliding window. These additional false positives contribute to the growths of the dropout field. As rule of thumb, the larger the , the longer the band of false positives, but the larger the threshold and the fewer false positives. False negatives are most likely when enclosed by large compact dropout fields.
Another drawback is the high computational cost of the OSCP because the sliding median is carried out multiple times over the entire data set. This will become more relevant the higher that is.
To reduce , the radius should be larger than , but since increases with the radius, r should not be excessively large. Hence, a value of is appropriate for the benchmark data. The flatness should be set at about . The best-performing setup for the benchmark data at hand would be and .
Instead of the data-driven estimation of t, the OSCP allows the user to set t directly as well. In this case, the OSCP yields an accuracy comparable to that of the GTM, and seems to be a good threshold, but the disadvantageous canceling of genuine values at dropout fields remains.
SMZ
As visible in
Figure 8,
Figure 9,
Figure 10,
Figure 11 and
Figure 12, with the exception of
Figure 10, the results of the SMZS and OSCP are similar. The SMZS’s configuration range is tighter than that of the OSCP in most scenarios. While the SMZS perform worse than the OSCP in scenario (b), the SMZS performs better than the OSCP in crack pattern scenarios.
The
z-score algorithm family comprises several variants [
21]. Each variant implements options for different data set characteristics. Advantages of the
z-score algorithms are (i) that the threshold
is based on the statistical distribution of the data; (ii) that a dynamic threshold (
(that varies over the data set) for signals with varying characteristics is possible [
21]; (iii) that it can take the complete measurement into account, which works best for nearly constant signals; and (iv) that it supports native 2D operation.
On the other hand, the SMZS comes with some disadvantages as well: (i) its tendency to flag false positives at the edges of dropout fields in the peak slopes (up to ), (ii) its configuration corridor for good results is tighter than that of GTM with more sensitive parameters (especially the radius), and (iii) its limited capability in anomalous areas.
The numerical cost of the SMZS is between those of the GTM and OSCP. The best-performing parameter combination for the benchmark data would be and .
6.2. Noise Reduction Benchmark
6.2.1. Setup
This benchmark is used to compare (i) the sliding average, also known as sliding mean, (ignoring NaNs); (ii) the sliding median (ignoring NaNs); and (iii) the cluster filter for their effectiveness in reducing noise and their suitability in neutralizing SRAs. Two cases are investigated. The first case considers noise as the only disturbance type: the “dirty” signal is composed of the clean signal
and the noise component
. The second case additionally incorporates SRAs and dropouts and uses the benchmark data as shown in
Figure 7 to test the robustness of these three methods. This benchmark will highlight the importance of dedicated SRA detection algorithms and show that the sliding average is particularly inappropriate for removing SRAs.
The algorithm’s filter width parameter (which configures how far one element influences the filtered signal) was varied to investigate the sensitivity. The filter width parameter for the sliding filters is the window radius, which was varied between . The higher the r, the higher the width. The filter width parameter for the cluster filter is the scaling factor , which was varied . The lower the , the higher the width. Additionally, the cluster filter’s tolerance convergence criterion is fixed at 0.1, and the filling of the dropouts is deactivated. Note that a radius of or is equivalent to no filtering.
6.2.2. Results
The accuracy of the filtering methods is assessed by means of the
according to Equation (
9), and the
, according to Equation (
10). The combined interpretation of both
and
enables a more detailed assessment. For reference, the noisy signal has the error measures
and
. Note the logarithmic scale for the error measures (
y-axis) and the parameter
(
x-axis) for the cluster filter in
Figure 13 and
Figure 14.
Case 1: Noise Only
The first benchmark case’s results are shown in
Figure 13. Overall, all three filters show similar results. It is notable that both
and
show the same qualitative behavior, indicating that there is no redistribution of errors. The filter width-dependent behavior results can be distinguished into three families: (i) the quasi-stationary (i.e., no or very little variance in the clean signal) scenario a), where the errors approach 0 with increasing filter width; (ii) the dynamic (i.e., mild to strong variance in the clean signal) scenarios (c), (d), and (e), where errors first decrease up to a certain filter width, but after that increase again; and (iii) the sharp-edged (i.e., strong variance in the clean signal with sharp edges) scenario (b), where the error is strictly increasing with an increasing filter width. The sliding average performs slightly better than the sliding median in scenario (a), as the sliding average is statistically optimal filter with white (i.e., normally distributed random) noise with respect to reaction sharpness [
47].
For the medium dynamic scenarios, the error measures first decrease, but increase again with increasing range. The points of the minimum errors depend on the sharpness of the changes in the signal. Eventually, the errors exceed the raw signals’ errors at the point where the negative effect of signal manipulation dominates the smoothing effect. This shows the tight limits for the beneficial use of filtering. The beneficial range of the sliding median is larger than the sliding averages’ one.
The results in scenario (b) show that the sliding median preserves edges better than the sliding average. The sliding average and the cluster filter round out curves, resulting in a distorted signal; see
Figure 15.
Case 2: Noise, SRAs, and Dropouts
The second benchmark case’s results are shown in
Figure 14. Due to the simulated dropouts, the different scenarios have varying values for the
of the unfiltered result from 245.8
to 252.4
and for the
from 1814.8
to 1836.3
.
The results of the algorithms differ notably for the second case (with dropouts and SRAs). Firstly, for the sliding average, it is visible that the
is on a relatively constant, high level. The non-stationary scenarios do not have a beneficial range for the filter width. The
increase comes from the “interpolation” effect on boundaries of large compact dropout fields, which grows in strength with the filter width. Conversely, the sliding average has a smooth interpolation effect on isolated dropouts. However, the
is reduced with increasing filter width. This shows that the sliding average does not eliminate the errors introduced by the SRA. Instead, the sliding average redistributes the error value of the SRAs to the adjacent
r elements. Hence, an SRA is softened and results in bumps in the signal; see
Figure 15. This has a serious implication, as those bumps might be indistinguishable from legitimate local effects (e.g., cracks) and alter the actual signal.
A considerable reduction in both
and
is achieved when employing the sliding median. The drastic drop in errors indicates that the sliding median remediates the errors introduced by the SRAs. The sliding median’s filtered signal is not affected by isolated SRA. Again, the points of minimum errors depend on the variance in the clean signal and the error measure, with the corresponding radius varying
. The increasing part is again due to the distortion governed by the distortion on the edges of dropout fields; see
Figure 15.
The cluster filter takes the middle ground between the sliding average and the sliding median. Similar to to the sliding average, the cluster filter softens SRAs to bumps in the resulting signal. However, due to the large dissimilarity of SRAs with the surrounding data, the cluster filter assigns a smaller weight to SRAs. Hence, the disruptive influence of SRAs is drastically reduced. Similarly to the other algorithms, the cluster filter exhibits a boundary effect on large dropouts fields. In comparison to the other filters, the cluster filter requires enormous numerical effort.
Filtering Example
Figure 15 shows examples of raw and filtered data for a visual comparison.
Figure 15a shows a section from scenario (b), highlighting the different capabilities of the filters to preserve edges.
Figure 15b shows a section from scenario (d), where the distorting effect on the edge of the dropout field is visible. It is visible in both figures how the sliding average and the cluster filter smooth out SRAs and the sliding median cancels out SRAs.
Recommendations
As with anomaly detection, the best parameters are use case-dependent. The sliding average is a very solid choice for noisy, but SRA free data, but are inappropriate for eliminating SRA. While less prone to SRAs, the same judgment holds for the cluster filter. In the case of data with SRAs, it is recommended to employ a dedicated SRA detection algorithm (or alternatively, the sliding median) instead. The SRA detection algorithm could be followed by an optional filter pass. However, the filtering pass is not always necessary. Filtering might improve the signal quality but bears the risk of signal distortion if overdone. As an aside, this investigation showed that there was no benefit to chaining a sliding median and a sliding average.
6.3. Dropout Reconstruction
6.3.1. Setup
This section is dedicated to comparing two simple, yet common repair methods: (i) linear interpolation and (ii) spline interpolation using an Akima spline. For demonstration purposes, a Gaußian curve, normalized to , , and , was chosen. The influence of scaling and sensitivity is negligible for the sake of this benchmark.
In the first case of the benchmark, the steepest parts of the slope—located around the inflections points—are erased. Dropouts in the peak flanks caused by strain gradients that are too steep are the most common cause of dropouts [
21,
36]. With the factor
t for the widths of the dropout field in terms of the curve’s standard deviation
computed according to Equation (
4), all elements
within
around the curve’s inflection points
are set to NaN.
In the second case of the benchmark, the highest part of the peak is erased by setting all elements
to NaN, for which
holds. In other words, the part of the tip that would remain in the first case when
is erased as well; see
Figure 16. Then, the dropouts are replaced with interpolated values, and the crack widths are calculated by integrating the strain profile [
31]. Finally, the accuracy of the reconstruction is evaluated using Equation (
11).
6.3.2. Results
Figure 16 shows some strain profiles for selected
for the first case. Both cases are identical for
. The left-hand side shows the reference signal and the signal with the dropouts. The right-hand-side plots show the same strain profiles interpolated using the two approaches. While the difference is not visible or very small for
, the difference between the reference and the reconstructed signal in the third example, where the peak tip is missing, is striking.
In
Figure 17, the relative crack width error according to Equation (
11) is plotted against
t. Solid lines correspond to the first case (erased slopes), and dashed lines correspond to the second benchmark case (erased tip). As already mentioned, both benchmark cases are the same for
. However, for
, they differ fundamentally.
In the second benchmark case, the crack width is underestimated by both approaches, and the errors increase continuously with increasing t. As expected, the Akima spline is closer to the reference than the linear interpolation. Note that for the second benchmark case, the peak tip between the inflection points is already erased for .
In the first benchmark case, the errors stay close to 0 until the critical point
. Then, the complete peak tip is lost, and both approaches lead to prone errors. For better visibility, the error of the first benchmark case’s errors for
—where the peak tip is still present—is magnified in the right-hand-side plot. It is visible that the errors do not exceed ±5%. These observations are consistent with those from experimental tests [
36]. While the linear interpolations tend to underestimate the crack width, the
Akima spline tends to overestimate it, which is on the safe side. Whether the crack width is underestimated or overestimated depends on the general shape function of the crack. For an exponential peak function [
63], both approaches might overestimate the actual crack width. In general, the
Akima spline offers a smaller deviation from the reference curve than the linear interpolation.
The conclusions from this benchmark are as follows. As long as the tip of a strain peak is preserved, the reconstruction results are acceptable with both interpolation methods. Although the
Akima spline has an overall better fit than the linear interpolation, both approaches are not able to reconstruct the complete upper peak tip with reasonable error. For reconstructing strain peaks with a severe amount of dropout, more sophisticated estimation approaches are required. Those approaches—based on mechanical models [
63,
64], fitting analytical functions [
36], or machine learning models [
65,
66]—could leverage the knowledge of DFOS sensitivity for specific combinations of sensor, adhesive, and installation methods.
At the moment, preserving the strain peak tips is imperative for reliable crack monitoring. This conclusion imposes implications for both the “hardware” and the “software” part. Regarding the “hardware”, it is most important to choose a sensor-adhesive combination appropriate for the measuring task, e.g., based on preliminary crack pattern simulation [
36]. For the “software”, it is most important to not remove the peak tips in the pre-processing, especially in the SRA masking stage. In particular, when only a few gages with valid values are left, distinguishing them from SRAs is challenging.
6.4. Limitations
Some limitations of this study shall be pointed out. The benchmark contained several scenarios to simulate different use cases. However, real data might contain further interfering influences caused by physical effects, such as inhomogeneities in the concrete material [
43], effects of local bonds [
16,
33,
44], or temperature changes [
67]. Because accounting for those physical effects (e.g., by compensation measures) is another step in the data evaluation; these were not taken into account in the benchmark. The SRAs and dropouts were distributed independently of each other in the benchmark data. However, it was observed that SRAs and dropouts do have a dependence and that it is more likely that an SRA neighbors a dropout or another SRA than a genuine value [
62]. Additionally, the simulation did not take HF-SRAs into account. Only a rather simple approach was taken to simulate the dropouts due to technical limitations, as a more detailed investigation can be found in [
36]. The algorithms may therefore perform differently on real data. The benchmark was carried out on only one instance of the benchmark data, only on one-dimensional data. However, the algorithms could utilize additional information when operating in a native 2D mode.
7. Conclusions
Even with the best monitoring setup, DSS measurement data might contain serious signal disturbances. Depending on their severity, these disturbances make further data processing and information extraction difficult or even impossible. Therefore, the pre-processing of the raw data with advanced approaches is required. The three different disturbance types found in DSS data—SRAs, dropouts, and noise—were investigated in this study. Their causes and commonly applied remediation approaches were discussed. Selected pre-processing approaches were discussed in detail, implemented and benchmarked with a synthetic data set. A generic pre-processing workflow for DSS data was proposed and implemented into the free open source software framework fosanalysis. The following conclusions could be drawn.
7.1. SRA Detection
In general, it is recommended to use a dedicated SRA detection algorithm. All three investigated algorithms (GTM, OSCP, and SMZS) can detect SRAs with acceptable reliability. Only a few minor and hard-to-detect SRAs were undetected. Given the benchmark results, the GTM would be equally or slightly preferred over SMZS, with the OSCP taking third place. The algorithms differ mostly in their tendency to false positives in the vicinity of dropout fields and their numerical costs. For all algorithms, a trade-off between false negatives and false positives was observed, and suitable parameter ranges were identified. However, false positives are not problematic, unless they remove primary features (e.g., the peak tips) of the signal. Hence, some manual quality control is still recommended.
7.2. Filtering
Filtering has a limited beneficial range with an optimum in the trade-off between noise removal and preserving local effects. However, some use cases might not require filtering at all. For example, the noise of 10
is unproblematic for crack width monitoring [
16]. All three investigated filters (sliding average, sliding median, and the cluster filter) work well on SRA free signals. In this case, the benchmark results of the sliding median are comparable to those of the sliding average. In the second case, where strain curves with SRAs and dropouts were analyzed, the sliding median was shown to be stable against solitary anomalies with more extreme deviations from the true signal. As these outliers are the main challenge for the subsequent processing steps, the sliding median is often the better choice over the cluster filter or the sliding average.
7.3. Reconstruction
Dropouts most commonly occur in the flanks of strain peaks due to exceeding the technical limitations of the ODiSI. The reconstruction of dropout sections using simple interpolation methods (such as linear and
Akima spline interpolation) yields good results as long as the strain peaks tips are preserved. In this case, the relative error for crack width calculation stays within 5%. However, if the strain peak tip is lost, the associated crack width is drastically underestimated. Hence, the best strategy is to avoid excessive absolute strain values and strain gradients by selecting an appropriate DFOS-adhesive combination according to the monitoring task [
36]. Other approaches could yield more accurate peak reconstruction results, e.g., based on machine learning [
65,
68], mechanical models [
63,
69], or numerical simulations [
70,
71].
7.4. Final Remarks
Since the requirements for the pre-processing workflow are specific to the data set at hand, there is no “one size fits all” solution for pre-processing. Hence, a flexible workflow is proposed with these recommended steps (although the order might require changes):
Use a dedicated SRA detection algorithm;
Use downsampling along the time axis with the median as an aggregation function to leverage its noise reduction feature which is robust against HL-SRAs when analyzing static loading situations;
Use the sliding median as a smoothing filter;
Use an Akima spline for interpolation.
Pre-processing cannot replace high-quality raw data in the first place. Maintaining the best possible quality of raw data is a critical task in the design phase of DFOS-based monitoring systems. For example, preserving the strain peak tips is critical for crack monitoring. In the value chain of data processing, the trustworthiness of the final result is limited to the lowest trustworthiness of all the links in the entire chain. Hence, no matter what pre-processing is used, if the raw data are of poor quality, the results cannot be trusted.
Author Contributions
Conceptualization, B.R.; methodology, B.R.; software, B.R. and L.U.; validation, B.R.; formal analysis, B.R.; investigation, B.R.; writing—original draft preparation, B.R. and L.U.; writing—review and editing, B.R. and M.H.; visualization, B.R.; project administration, S.M. and M.H.; funding acquisition, S.M. and M.H. All authors have read and agreed to the published version of the manuscript.
Funding
This paper presents some results of the research project IDA-KI (Automated assessment of monitoring data for infrastructure constructions using AI and IoT) funded by the Federal Ministry for Digital and Transport, Germany, within the innovation program mFUND (funding reference: 19FS2013A and 19FS2013C). The Article Processing Charge (APC) was funded by the joint publication funds of the TU Dresden, including Carl Gustav Carus Faculty of Medicine, and the SLUB Dresden, as well as the Open Access Publication Funding of the German Research Foundation (DFG).
Data Availability Statement
The investigated algorithms were implemented into the Python software framework
fosanalysis, which is made available as free open source software at
https://github.com/TUD-IMB/fosanalysis (accessed on 18 November 2024). The benchmark data along with the Python scripts for generating the data and carrying out the benchmarks were made available as open data [
60]. The experimental data on the characteristics of SRAs were made available as open data [
62].
Acknowledgments
We would like to thank the funding authority for their financial support as well as our research partners for their productive cooperation within the project.
Conflicts of Interest
Author Lisa Ulbrich was employed by the company Hentschke Bau GmbH. The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Abbreviations
BZS | basic z-score |
c-OFDR | coherent optical frequency domain reflectometry |
DFOS | distributed fiber optic sensor |
DSS | distributed strain sensing |
FNC | forward neighbor comparison |
GTM | geometric threshold method |
HF-SRA | harmful strain reading anomaly |
HL-SRA | harmless strain reading anomaly |
MAD | median absolute deviation |
MZS | modified z-score |
OSCP | outlier-specific correction procedure |
NaN | not a number |
ODiSI | optical distributed sensor interrogator |
RS | reverse sweep |
SHM | structural health monitoring |
RMSE | root mean square error |
ME | mean error |
SRA | strain reading anomaly |
SMZS | sliding modified z-score |
SSQ | spectral shift quality |
WCM | weighted confusion measure |
WHZ | Whitaker–Hayes z-score |
| weighted confusion measure for false positives |
| weighted confusion measure for false negatives |
References
- Cheng, L.; Cigada, A.; Zappa, E.; Gilbert, M.; Lang, Z.Q. Dynamic monitoring of a masonry arch rail bridge using a distributed fiber optic sensing system. J. Civ. Struct. Health Monit. 2024, 14, 1075–1090. [Google Scholar] [CrossRef]
- Bednarski, Ł.; Sieńko, R.; Howiacki, T.; Zuziak, K. The Smart Nervous System for Cracked Concrete Structures: Theory, Design, Research, and Field Proof of Monolithic DFOS-Based Sensors. Sensors 2022, 22, 8713. [Google Scholar] [CrossRef]
- Howiacki, T.; Sieńko, R.; Bednarski, Ł.; Zuziak, K. Structural monitoring of concrete, steel, and composite bridges in Poland with distributed fibre optic sensors. Struct. Infrastruct. Eng. 2023, 20, 1213–1229. [Google Scholar] [CrossRef]
- Vorwagner, A.; Kwapisz, M.; Lienhart, W.; Winkler, M.; Monsberger, C.; Prammer, D. Verteilte Rissbreitenmessung im Betonbau mittels faseroptischer Sensorik—Neue Anwendung von verteilten faseroptischen Messsystemen. Beton-und Stahlbetonbau 2021, 116, 727–740. [Google Scholar] [CrossRef]
- Monsberger, C.; Bauer, P.; Buchmayer, F.; Lienhart, W. Large-scale distributed fiber optic sensing network for short and long-term integrity monitoring of tunnel linings. J. Civ. Struct. Health Monit. 2022, 12, 1317–1327. [Google Scholar] [CrossRef]
- Zhang, X.; Zhu, H.; Jiang, X.; Broere, W. Distributed fiber optic sensors for tunnel monitoring: A state-of-the-art review. J. Rock Mech. Geotech. Eng. 2024, 16, 3841–3863. [Google Scholar] [CrossRef]
- Brezzi, L.; Schenato, L.; Cola, S.; Fabbian, N.; Chemello, P.; Simonini, P. Smart Monitoring by Fiber-Optic Sensors of Strain and Temperature of a Concrete Double Arch Dam. In Geotechnical Engineering in the Digital and Technological Innovation Era; Springer Nature: Cham, Switzerland, 2023; pp. 161–168. [Google Scholar] [CrossRef]
- Wang, D.Y.; Zhu, H.H.; Wu, B.; Ye, X.; Wang, J.; Tan, D.Y.; Shi, B. Performance evaluation of underground pipelines subjected to landslide thrust with fiber optic strain sensing nerves. Acta Geotech. 2024, 19, 6993–7009. [Google Scholar] [CrossRef]
- Lienhart, W.; Monsberger, C.M.; Buchmayer, F. How to Make a Self-sensing House with Distributed Fiber Optic Sensing. In European Workshop on Structural Health Monitoring; Springer International Publishing: Cham, Switzerland, 2022; pp. 718–726. [Google Scholar] [CrossRef]
- Wijaya, H.; Rajeev, P.; Gad, E. Distributed optical fibre sensor for infrastructure monitoring: Field applications. Opt. Fiber Technol. 2021, 64, 102577. [Google Scholar] [CrossRef]
- Kishida, K.; Imai, M.; Kawabata, J.; Guzik, A. Distributed Optical Fiber Sensors for Monitoring of Civil Engineering Structures. Sensors 2022, 22, 4368. [Google Scholar] [CrossRef]
- Wang, J.; Garg, A.; Satyam, N.; Zhussupbekov, A.; Sushkova, S. DFOS Technology in Geoengineering Monitoring in the Past 35 Years: A Bibliometric Analysis. Sensors 2024, 24, 5051. [Google Scholar] [CrossRef]
- Barrias, A.; Casas, J.; Villalba, S. A Review of Distributed Optical Fiber Sensors for Civil Engineering Applications. Sensors 2016, 16, 748. [Google Scholar] [CrossRef] [PubMed]
- Bado, M.F.; Casas, J.R. A Review of Recent Distributed Optical Fiber Sensors Applications for Civil Engineering Structural Health Monitoring. Sensors 2021, 21, 1818. [Google Scholar] [CrossRef] [PubMed]
- Herbers, M.; Richter, B.; Gebauer, D.; Claßen, M.; Marx, S. Crack Monitoring on Concrete Structures—Comparison of Various Distributed Fiber Optic Sensors with Digital Image Correlation Method. Struct. Concr. 2023, 24, 6123–6140. [Google Scholar] [CrossRef]
- Richter, B.; Herbers, M.; Marx, S. Crack monitoring on concrete structures with distributed fiber optic sensors—Toward automated data evaluation and assessment. Struct. Concr. 2023, 25, 1465–1480. [Google Scholar] [CrossRef]
- Paul, A.; Sanio, D.; Mark, P. Monitoring tendon breaks in concrete structures at different depths using distributed fiber optical sensors. e-J. Nondestruct. Test. 2024, 29. [Google Scholar] [CrossRef]
- Howiacki, T.; Sieńko, R.; Bednarski, Ł.; Zuziak, K. Crack Shape Coefficient: Comparison between Different DFOS Tools Embedded for Crack Monitoring in Concrete. Sensors 2023, 23, 566. [Google Scholar] [CrossRef]
- Bado, M.F.; Casas, J.R.; Gómez, J. Post-processing algorithms for distributed optical fiber sensing in structural health monitoring applications. Struct. Health Monit. 2021, 20, 661–680. [Google Scholar] [CrossRef]
- Janiak, T.; Becks, H.; Camps, B.; Classen, M.; Hegger, J. Evaluation of distributed fibre optic sensors in structural concrete. Mater. Struct. 2023, 56, 159. [Google Scholar] [CrossRef]
- Ulbrich, L. Detection and elimination of strain reading anomalies in distributed strain sensing readings. In Proceedings of the Tagungsband des 35. Forum Bauinformatik 2024, Hamburg, Germany, 18–20 September 2024; Stührenberg, J., Al-Zuriqat, T., Chillón Geck, C., Eds.; Technische Universität Hamburg: Hamburg, Germany, 2024. Tagungsband 35. Forum Bauinformatik 2024. pp. 131–138. [Google Scholar] [CrossRef]
- Luna Innovations Inc. User’s Guide ODiSI 6: Optical Distributed Sensor Interrogator Model ODiSI 6: User’s Guide ODiSI 6 Software. 2020. Available online: https://lunainc.com/sites/default/files/assets/files/resource-library/ODiSI%206100%20User%20Guide.pdf (accessed on 17 October 2024).
- Alj, I.; Quiertant, M.; Khadour, A.; Grando, Q.; Benzarti, K. Environmental Durability of an Optical Fiber Cable Intended for Distributed Strain Measurements in Concrete Structures. Sensors 2021, 22, 141. [Google Scholar] [CrossRef]
- Bremer, K.; Alwis, L.S.M.; Zheng, Y.; Weigand, F.; Kuhne, M.; Helbig, R.; Roth, B. Durability of Functionalized Carbon Structures with Optical Fiber Sensors in a Highly Alkaline Concrete Environment. Appl. Sci. 2019, 9, 2476. [Google Scholar] [CrossRef]
- Lemcherreq, Y.; Galkovski, T.; Mata-Falcón, J.; Kaufmann, W. Application of Distributed Fibre Optical Sensing in Reinforced Concrete Elements Subjected to Monotonic and Cyclic Loading. Sensors 2022, 22, 2023. [Google Scholar] [CrossRef] [PubMed]
- Gómez, J.; Casas, J.R.; Villalba, S. Structural Health Monitoring with Distributed Optical Fiber Sensors of tunnel lining affected by nearby construction activity. Autom. Constr. 2020, 117, 103261. [Google Scholar] [CrossRef]
- Bado, M.F.; Casas, J.R.; Barrias, A. Performance of Rayleigh-Based Distributed Optical Fiber Sensors Bonded to Reinforcing Bars in Bending. Sensors 2018, 18, 3125. [Google Scholar] [CrossRef] [PubMed]
- Luna Innovations Inc. User’s Guide ODiSI-B: Optical Distributed Sensor Interrogator Model ODiSI-B: User’s Guide ODiSI-B Software 5.2.0. 2017. Available online: https://www.advancedphotonix.com/wp-content/uploads/2014/05/ODiSI-B-Users-Guide.pdf (accessed on 17 October 2024).
- Barrias, A.; Casas, J.R.; Villalba, S. Embedded Distributed Optical Fiber Sensors in Reinforced Concrete Structures-A Case Study. Sensors 2018, 18, 980. [Google Scholar] [CrossRef] [PubMed]
- Chamoin, L.; Farahbakhsh, S.; Poncelet, M. An educational review on distributed optic fiber sensing based on Rayleigh backscattering for damage tracking and structural health monitoring. Meas. Sci. Technol. 2022, 33, 124008. [Google Scholar] [CrossRef]
- Fischer, O.; Thoma, S.; Crepaz, S. Distributed fiber optic sensing for crack detection in concrete structures. Civ. Eng. Des. 2019, 1, 97–105. [Google Scholar] [CrossRef]
- Li, Y.; Sharif-Khodaei, Z. Accuracy of Distributed Strain Sensing with Single-Mode Fibre in Composite Laminates under Thermal and Vibration Loads. Struct. Control. Health Monit. 2023, 2023, 9269987. [Google Scholar] [CrossRef]
- Galkovski, T.; Lemcherreq, Y.; Mata-Falcón, J.; Kaufmann, W. Fundamental Studies on the Use of Distributed Fibre Optical Sensing on Concrete and Reinforcing Bars. Sensors 2021, 21, 7643. [Google Scholar] [CrossRef]
- Bin Ismail, M.F.; Yanagi, K.; Fujii, A. An outlier correction procedure and its application to areal surface data measured by optical instruments. Meas. Sci. Technol. 2010, 21, 105105. [Google Scholar] [CrossRef]
- Ismail, M.F.B.; Jaafar, T.R.; Che Mat, S.; Pahmi, M.A.A.H. Evaluation of Outlier Specific Correction Procedure for Areal Surface Texture. Appl. Mech. Mater. 2014, 661, 137–142. [Google Scholar] [CrossRef]
- Herbers, M.; Richter, B.; Marx, S. Rayleigh-based crack monitoring with distributed fiber optic sensors—Experimental study on the interaction of spatial resolution and sensor type. Struct. Control Health Monit. 2024; forthcoming. [Google Scholar]
- Ansari, F.; Libo, Y. Mechanics of Bond and Interface Shear Transfer in Optical Fiber Sensors. J. Eng. Mech. 1998, 124, 385–394. [Google Scholar] [CrossRef]
- Her, S.C.; Huang, C.Y. Effect of Coating on the Strain Transfer of Optical Fiber Sensors. Sensors 2011, 11, 6926–6941. [Google Scholar] [CrossRef] [PubMed]
- Weisbrich, M.; Holschemacher, K.; Bier, T. Comparison of different fiber coatings for distributed strain measurement in cementitious matrices. J. Sens. Sens. Syst. 2020, 9, 189–197. [Google Scholar] [CrossRef]
- Akima, H. A New Method of Interpolation and Smooth Curve Fitting Based on Local Procedures. J. ACM 1970, 17, 589–602. [Google Scholar] [CrossRef]
- Weisbrich, M.; Holschemacher, K.; Bier, T. Validierung verteilter faseroptischer Sensorik zur Dehnungsmessung im Betonbau. Beton-und Stahlbetonbau 2021, 116, 648–659. [Google Scholar] [CrossRef]
- Lou, S.; Tang, D.; Zeng, W.; Zhang, T.; Gao, F.; Muhamedsalih, H.; Jiang, X.; Scott, P.J. Application of Clustering Filter for Noise and Outlier Suppression in Optical Measurement of Structured Surfaces. IEEE Trans. Instrum. Meas. 2020, 69, 6509–6517. [Google Scholar] [CrossRef]
- Weisbrich, M.; Messerer, D.; Holschemacher, K. The Challenges and Advantages of Distributed Fiber Optic Strain Monitoring in and on the Cementitious Matrix of Concrete Beams. Sensors 2023, 23, 9477. [Google Scholar] [CrossRef]
- Koschemann, M.; Curbach, M.; Marx, S. Investigation of local bond behavior using distributedoptical fiber sensing. In Proceedings of the Bond in Concrete—Bond, Anchorage, Detailing: 5th International Conference, Stuttgart, Germany, 25–27 July 2022; Hofmann, J., Plizzari, G., Eds.; Universität Stuttgart: Stuttgart, Germany, 2022; pp. 133–145. [Google Scholar] [CrossRef]
- Wheeler, L.N.; Pannese, E.; Hoult, N.A.; Take, W.A.; Le, H. Measurement of distributed dynamic rail strains using a Rayleigh backscatter based fiber optic sensor: Lab and field evaluation. Transp. Geotech. 2018, 14, 70–80. [Google Scholar] [CrossRef]
- Xiang, G.; Sun, A.; Liu, Y.; Gao, L. An improved denoising method for Φ-OTDR signal based on the combination of temporal local GMCC and ICEEMDAN-WT. Opt. Fiber Technol. 2024, 87, 103949. [Google Scholar] [CrossRef]
- O’Haver, T.C. A Pragmatic Introduction to Signal Processing 2023, 2023rd ed.; Kindle Direct Publishing: College Park, MD, USA, 2023. [Google Scholar]
- Smith, S.W. Digital Signal Processing: A Practical Guide for Engineers and Scientists; Newnes, Elsevier Science: Burlington, MA, USA, 2003; ISBN 0-750674-44-X. [Google Scholar] [CrossRef]
- Ershov, I.A.; Stukach, O.V.; Trubin, I.V.; Gladyshev, S.A. Features of Digital Filters in Raman DTS. In Proceedings of the 2023 Dynamics of Systems, Mechanisms and Machines (Dynamics), Omsk, Russian, 14–15 November 2023; pp. 1–4. [Google Scholar] [CrossRef]
- Piątek, B.; Howiacki, T.; Kulpa, M.; Siwowski, T.; Sieńko, R.; Bednarski, Ł. Strain, crack, stress and shape diagnostics of new and existing post-tensioned structures through distributed fibre optic sensors. Measurement 2023, 221, 113480. [Google Scholar] [CrossRef]
- Monsberger, C.M.; Lienhart, W. Distributed Fiber Optic Shape Sensing of Concrete Structures. Sensors 2021, 21, 6098. [Google Scholar] [CrossRef] [PubMed]
- Ouellet, S.M.; Dettmer, J.; Lato, M.J.; Cole, S.; Hutchinson, D.J.; Karrenbach, M.; Dashwood, B.; Chambers, J.E.; Crickmore, R. Previously hidden landslide processes revealed using distributed acoustic sensing with nanostrain-rate sensitivity. Nat. Commun. 2024, 15, 6239. [Google Scholar] [CrossRef] [PubMed]
- Liu, H.; Zhang, S.; Coulibaly, A.A.S.; Cheng, J.; DeJong, M.J. Monitoring Reinforced Concrete Cracking Behavior under Uniaxial Tension Using Distributed Fiber-Optic Sensing Technology. J. Struct. Eng. 2021, 147, 04021212. [Google Scholar] [CrossRef]
- Ma, K.F.; von Specht, S.; Kuo, L.W.; Huang, H.H.; Lin, C.R.; Lin, C.J.; Ku, C.S.; Wu, E.S.; Wang, C.Y.; Chang, W.Y.; et al. Broad-band strain amplification in an asymmetric fault zone observed from borehole optical fiber and core. Commun. Earth Environ. 2024, 5, 402. [Google Scholar] [CrossRef]
- Ershov, I.A.; Stukach, O.V. Choice of Wavelet for Filtering of Signal from Fiber-Optic Temperature Sensor. In Proceedings of the 2022 Moscow Workshop on Electronic and Networking Technologies (MWENT), Moscow, Russia, 9–11 June 2022. [Google Scholar] [CrossRef]
- Ershov, I.A.; Stukach, O.V.; Myasnikova, N.V. Features of the Implementation of the Extremal Filtration Method in the Distributed Optic-Fiber Temperature Sensor. In Proceedings of the 2021 International Seminar on Electron Devices Design and Production (SED), Prague, Czech Republic, 27–28 April 2021; pp. 1–5. [Google Scholar] [CrossRef]
- Iglewicz, B.; Hoaglin, D.C. How to Detect and Handle Outliers; Number 16 in The ASQC basic references in quality control: Statistical techniques; ASQC Quality Press: Milwaukee, WI, USA, 1993. [Google Scholar]
- Whitaker, D.A.; Hayes, K. A simple algorithm for despiking Raman spectra. Chemom. Intell. Lab. Syst. 2018, 179, 82–84. [Google Scholar] [CrossRef]
- Pham-Gia, T.; Hung, T. The mean and median absolute deviations. Math. Comput. Model. 2001, 34, 921–936. [Google Scholar] [CrossRef]
- Richter, B. Artificial data set for benchmarking pre-processing algorithms for distributed fiber optic strain data. 2024. [Google Scholar] [CrossRef]
- Zhang, S.; Liu, H.; Coulibaly, A.A.S.; DeJong, M. Fiber optic sensing of concrete cracking and rebar deformation using several types of cable. Struct. Control. Health Monit. 2021, 28, e2664. [Google Scholar] [CrossRef]
- Ulbrich, L. Statistical Evaluation of Strain Reading Anomalies in Distributed Fiber Optic Strain Sensing Measurements. 2024. [Google Scholar] [CrossRef]
- Bassil, A.; Chapeleau, X.; Leduc, D.; Abraham, O. Concrete Crack Monitoring Using a Novel Strain Transfer Model for Distributed Fiber Optics Sensors. Sensors 2020, 20, 2220. [Google Scholar] [CrossRef]
- Hong, C.; Rao, W.; Qiu, T.; Chen, X.; Dai, J.; Wu, C.; Li, M.; Chen, W. Monitoring and assessment of mechanical response of large-scale prefabricated structures in underground subway stations during construction process. Measurement 2024, 235, 115015. [Google Scholar] [CrossRef]
- Liu, Y.; Bao, Y. Automatic interpretation of strain distributions measured from distributed fiber optic sensors for crack monitoring. Measurement 2023, 211, 112629. [Google Scholar] [CrossRef]
- Liu, Y.; Bao, Y. Intelligent monitoring of spatially-distributed cracks using distributed fiber optic sensors assisted by deep learning. Measurement 2023, 220, 113418. [Google Scholar] [CrossRef]
- Murray, M.J.; Murray, J.B.; Ogden, H.M.; Redding, B. Dynamic temperature-strain discrimination using a hybrid distributed fiber sensor based on Brillouin and Rayleigh scattering. Opt. Express 2022, 31, 287. [Google Scholar] [CrossRef] [PubMed]
- Venketeswaran, A.; Lalam, N.; Wuenschell, J.; Ohodnicki, P.R.; Badar, M.; Chen, K.P.; Lu, P.; Duan, Y.; Chorpening, B.; Buric, M. Recent Advances in Machine Learning for Fiber Optic Sensor Applications. Adv. Intell. Syst. 2021, 4, 2100067. [Google Scholar] [CrossRef]
- Zhang, S.; Liu, H.; Cheng, J.; DeJong, M.J. A mechanical model to interpret distributed fiber optic strain measurement at displacement discontinuities. Struct. Health Monit. 2020, 20, 2584–2603. [Google Scholar] [CrossRef]
- Alj, I.; Quiertant, M.; Khadour, A.; Grando, Q.; Terrade, B.; Renaud, J.C.; Benzarti, K. Experimental and Numerical Investigation on the Strain Response of Distributed Optical Fiber Sensors Bonded to Concrete: Influence of the Adhesive Stiffness on Crack Monitoring Performance. Sensors 2020, 20, 5144. [Google Scholar] [CrossRef]
- Billon, A.; Hénault, J.M.; Quiertant, M.; Taillade, F.; Khadour, A.; Martin, R.P.; Benzarti, K. Qualification of a distributed optical fiber sensor bonded to the surface of a concrete structure: A methodology to obtain quantitative strain measurements. Smart Mater. Struct. 2015, 24, 115001. [Google Scholar] [CrossRef]
Figure 1.
Schematic of the different measurement disturbances: SRAs, dropouts, and noise. The variation in the signal marked as local effects is not caused by the measurement principle.
Figure 1.
Schematic of the different measurement disturbances: SRAs, dropouts, and noise. The variation in the signal marked as local effects is not caused by the measurement principle.
Figure 2.
The task and workflow duality implemented by the pre-processing module of fosanalysis. (a) Class inheritance hierarchy of task objects. (b) Structure of a pre-processing workflow object. (c) Pre-processing workflow with an exemplary sequence of task objects.
Figure 2.
The task and workflow duality implemented by the pre-processing module of fosanalysis. (a) Class inheritance hierarchy of task objects. (b) Structure of a pre-processing workflow object. (c) Pre-processing workflow with an exemplary sequence of task objects.
Figure 3.
GTM for one-dimensional case, as presented in [
19]; additions are highlighted in grey.
Figure 3.
GTM for one-dimensional case, as presented in [
19]; additions are highlighted in grey.
Figure 4.
OSCP, adapted from [
34].
Figure 4.
OSCP, adapted from [
34].
Figure 5.
Algorithm of the z-score family.
Figure 5.
Algorithm of the z-score family.
Figure 6.
Cluster filter, according to [
42].
Figure 6.
Cluster filter, according to [
42].
Figure 7.
Benchmark data scenarios: (
a) zero signal; (
b) ramps; (
c) strain profile with weak DFOS bond; (
d) strain profile with medium DFOS bond; (
e) strain profile with stiff DFOS bond. The data set is available in [
60].
Figure 7.
Benchmark data scenarios: (
a) zero signal; (
b) ramps; (
c) strain profile with weak DFOS bond; (
d) strain profile with medium DFOS bond; (
e) strain profile with stiff DFOS bond. The data set is available in [
60].
Figure 8.
SRA detection accuracy for the algorithms for scenario (a) zero.
Figure 8.
SRA detection accuracy for the algorithms for scenario (a) zero.
Figure 9.
SRA detection accuracy for the algorithms for scenario (b) ramps.
Figure 9.
SRA detection accuracy for the algorithms for scenario (b) ramps.
Figure 10.
SRA detection accuracy for the algorithms for scenario (c) weak bond.
Figure 10.
SRA detection accuracy for the algorithms for scenario (c) weak bond.
Figure 11.
SRA detection accuracy for the algorithms for scenario (d) normal bond.
Figure 11.
SRA detection accuracy for the algorithms for scenario (d) normal bond.
Figure 12.
SRA detection accuracy for the algorithms for scenario (e) stiff bond.
Figure 12.
SRA detection accuracy for the algorithms for scenario (e) stiff bond.
Figure 13.
Results of filter benchmarks for case 1: noisy signal without SRAs or dropouts.
Figure 13.
Results of filter benchmarks for case 1: noisy signal without SRAs or dropouts.
Figure 14.
Results of filter benchmarks for case 2: noisy signal with SRAs and dropouts.
Figure 14.
Results of filter benchmarks for case 2: noisy signal with SRAs and dropouts.
Figure 15.
Section of strain data pre-processed with the filters: sliding average (), sliding median (), cluster filter (). (a) Scenario (b), (b) scenario (d).
Figure 15.
Section of strain data pre-processed with the filters: sliding average (), sliding median (), cluster filter (). (a) Scenario (b), (b) scenario (d).
Figure 16.
Left column: normalized strain peak with applied dropouts; right column: reconstructed strain peak with (i) , (ii) , and (iii) . The right-hand-side plot details the highlighted part in the left-hand-side plot.
Figure 16.
Left column: normalized strain peak with applied dropouts; right column: reconstructed strain peak with (i) , (ii) , and (iii) . The right-hand-side plot details the highlighted part in the left-hand-side plot.
Figure 17.
Error in crack widths for linear and Akima spline interpolation, depending on the amount of dropout t. The right-hand-side plot is the highlighted detail of the left plot.
Figure 17.
Error in crack widths for linear and Akima spline interpolation, depending on the amount of dropout t. The right-hand-side plot is the highlighted detail of the left plot.
Table 1.
Crack pattern parameters for the simulated strain profiles
Figure 7c–e.
Table 1.
Crack pattern parameters for the simulated strain profiles
Figure 7c–e.
Number | Position | Crack Width | Maximum Strain in |
---|
| in | in | | | |
1 | 0.2 | 181 | 1267 | 4525 | 6335 |
2 | 0.37 | 295 | 2065 | 7375 | 10,325 |
3 | 0.53 | 354 | 2478 | 8850 | 12,390 |
4 | 0.716 | 450 | 3150 | 11,250 | 15,750 |
5 | 0.91 | 321 | 2247 | 8025 | 11,235 |
6 | 1.03 | 220 | 1540 | 5500 | 7700 |
7 | 1.13 | 308 | 2156 | 7700 | 10,780 |
8 | 1.24 | 21 | 147 | 525 | 735 |
9 | 1.35 | 273 | 1911 | 6825 | 9555 |
10 | 1.42 | 211 | 1477 | 5275 | 7385 |
11 | 1.54 | 109 | 763 | 2725 | 3815 |
12 | 1.64 | 213 | 1491 | 5325 | 7455 |
13 | 1.75 | 77 | 539 | 1925 | 2695 |
14 | 1.895 | 293 | 2051 | 7325 | 10,255 |
15 | 2.08 | 1253 | 8771 | 31,325 | 43,855 |
16 | 2.3 | 369 | 2583 | 9225 | 12,915 |
17 | 2.455 | 836 | 5852 | 20,900 | 29,260 |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).