1. Introduction
Modern societies are strongly dependent on the use of complex structures. Because an early detection of structural faults can greatly reduce the maintenance cost over time and prevent catastrophic events [
1], being able to keep civil constructions safe and reliable is fundamental for community welfare [
2]. For these reasons, in the last decades, civil engineering has focused on Structural Health Monitoring (SHM) [
3], which aims at detecting, localizing, and quantifying damage occurrence. Especially, data-driven approaches [
4,
5] are becoming increasingly widespread thanks to their capacity of easily manage the large amount of data, acquired through pervasive sensor networks. In particular, by processing raw vibrational signals (e.g., acceleration recordings that are shaped as multivariate time series), they can extract useful
features, in order to determine the damage state of the structure. To this aim, Deep Learning (DL) algorithms can automatically extract damage-sensitive features [
6] and relate them with the corresponding structural states, by exploiting temporal correlations within and across time recordings.
When considering a finite number of predefined damage scenarios, the detection and localization of damage in structures can be treated within a classification framework. By employing a supervised classifier, the goal is to predict the categorical class (i.e., a label referring to a predefined damage scenario) of new incoming data, on the basis of past observations. Supervised techniques require labeled data on the possible damage conditions of the structure, which, however, are hard (if not impossible) to acquire for civil applications. This drawback calls for a Simulation-Based Classification (SBC) [
1,
7,
8], in order to numerically simulate the effect of damage on the structural response. In such hybrid model-data SHM approach, a dataset of synthetic time-signals, accounting for relevant operational conditions and varying environmental effects, is generated through simulations of a physics-based model, for the whole set of considered damage scenarios and, thus, assimilated with the DL algorithm.
In order to replace the expensive numerical models, relying on the Finite Element (FE) method, a Model Order Reduction (MOR) strategy, such as the Reduced Basis (RB) method [
9,
10], can be adopted in order to accelerate the dataset construction.
The proposed methodology exploits an offline-online decomposition. During the preliminary
offline phase, our DL-based classifier is trained on a numerically pre-built dataset of labeled inputs; during the
online phase, the trained classifier processes unseen experimental signals that were acquired on the fly, returning, as output, the structural state that might have most likely produced them. Our classifier is based on a Fully Convolutional Network (FCN) architecture, which was already successfully applied in [
8,
9].
Additional to damage phenomena, environmental conditions could affect measured signals. Thermal fluctuations (both daily and seasonal) can influence a wide range of material properties and induce structural displacements. It is not easy to distinguish these effects from those of damage. For this reason, the thermal effects are simulated in the numerical model of the structure and temperature measurements are used together with vibrational ones as inputs to the classifier.
2. SHM Methodology: Dataset Definition and Damage Classifier
Considering an observation windows , short enough to assume frozen operational, environmental, and damage conditions, the damage state of the structure is monitored by collecting vibrational and temperature data through a sensor network featuring vibrational sensors, with a fixed sampling period , and thermometers. The network arrangement has been designed starting from an initial placement involving a high number of sensors, progressively reduced evaluating the classifier performances on multiple datasets that were generated according to different sensor configurations. Vibrational recordings consist of displacement and/or acceleration measurements () of length (for the sake of simplicity, in this section we only consider displacement measurements), while each thermometer outputs a single value . A single data instance is composed of a set of displacement recordings and temperature measurements , where: labels the damage state that is undergone by the structure in the i-th instances, modeled as a localized stiffness reduction in pre-designated regions; and label the set of parameters that control the mechanical and the temperature field, respectively. The dataset is made of I instances , ; to relieve the computational burden of its generation, the Full-Order Model (FOM), which relies on the FE method, is replaced by a Reduced-Order Model (ROM).
According to the adopted classification framework, only a discrete number of damage scenarios
has to be defined on the basis of the mechanical behavior, load conditions, and aging processes interesting similar structures [
4]. The baseline undamaged state is labeled as
.
A continuous probability density function (pdf) describes the occurrence of each entry of and , while a discrete pdf governs the occurrence of . The parameter set , which is sampled from pdfs that take the locality of interest and the seasonality of temperature fluctuations into account, controls the temperature profiles that are imposed at the domain edges. The set parametrizes the external loads (e.g., amplitude and frequency of a dynamical load) and the damage level , intended as the intensity of the stiffness reduction involving the subdomain that is related to . A suitable sampling rule (e.g., a Latin Hypercube Sampling) has been adopted in order to explore the parametric space defined combining , and g. Each sampling uniquely identifies the corresponding dataset instance .
Once built, is employed to train and validate a classifier . During the training phase, instances are employed by to learn the underlying mapping between the i-th instance and , while instances (with ) are used in order to validate the learning process. Once trained, should be able to correctly map an unseen instance into the correct damage state . In the absence of experimental data, the generalization capabilities have been assessed on a test set built through FOM simulations, in this way ensuring better fidelity to the experimental framework. For sake of clarity, from now on, we disregard specifying the instance index i.
Our FCN architecture, resembling the one that is proposed in [
8], processes
by exploiting a stack of three Convolutional Units (CU), followed by a Global Average Pooling (GAP), whose output is merged with
through a Concatenation Block (CB). These latter support the classifier in recognizing thermal effects of material contraction/expansion and stiffening/softening within the observed dynamics, in order to not confuse the environmental variability with damage [
11]. Thus, a dense layer operates a linear mapping, ruled by a weight matrix
, allowing a final Softmax layer to perform the classification task. Each of the three CUs is formed by a Convolutional Layer (CL), together with Batch Normalization, in order to stem gradient instability issues during training, and ReLU activation function. In a CL, connection weights
can be imagined as filters of kernel size
, with
, to be applied to the output of the previous layer. Each convolutional layer applies
filters to its inputs, yielding an output that is made of
feature maps. This composition of nonlinear transformations makes each damage target class linearly separable and allows for addressing the temporal pattern recognition, exploiting inter-sensor correlations, by simultaneously analyzing
. The GAP condenses the resulting feature maps, which elicits a single, yet highly informative, description of its input channel
.
While training , the learning algorithm tunes the weights and , by iteratively minimizing a loss function over the instances. The adopted loss function is the cross entropy, as is usually done in classification frameworks; Adam, a first-order stochastic gradient descend algorithm, has been employed in order to perform the iterative minimization process. At each iteration, a certain number of instances B, called mini-batch, are simultaneously analyzed in order to update the connection weights; each time all of the instances have been processed is said to be an epoch.
The FCN hyperparameters (
), initially set according to [
8], have been tuned through the repeated evaluation of the classification accuracy. Herein, we have adopted:
,
, and
, as the number of filters;
,
and
, as kernel sizes;
as mini-batch size;
, as number of instances, with the ratio
; 1000 training epochs.
3. SHM Methodology: Dataset Population
Adopting a SBC approach, the generation of the I instances has been carried out by evaluating the physics-based model of the structure, for multiple values of the input parameters and for each considered scenario .
The thermo-mechanical behavior of the structure has been modeled through the standard linear thermo-elasticity theory employing a one-way coupling approach; the thermal field is independently determined from the kinematic one, but it still influences the material deformations. Moreover, having supposed monitoring windows of fixed duration, being significantly lower than the time that is required to experience notable temperature excursions, the thermal field has been evaluated while disregarding its temporal dependence. In order to reflect an oscillation of the material temperature into the dynamical response, a local dependency of the Young modulus
E on temperature has been introduced. Lastly, damping effects have been disregarded due to the small relevance in the identification of continuously excited dynamic systems [
12,
13]; besides, modeling the structural damage as a selective reduction in stiffness, frozen in
, the mechanical behavior has been treated as linear [
3]. The FOM reads, as follows:
and results from a Galerkin–FE discretization of a stationary diffusion problem and of a elasto-dynamic problem with a coupling term, respectively. In particular:
denotes the time coordinate;
is the temperature vector;
is the thermal conductivity matrix;
is the thermal right hand side vector;
is the displacement vector;
is the mass matrix;
is the elastic stiffness matrix;
is the coupling term;
is the mechanical right hand side vector;
and
denote, respectively, the number of degrees of freedom (dofs) of the temperature and the displacement FE spaces.
Pb. (
1) is first solved in order to determine the temperature field
. Vector
is transformed into equivalent nodal forces through
. A displacement field, compatible with
, is determined by solving a static mechanical problem under the coupling action only, and it is assumed as the reference around which the dynamics that are governed by Pb. (
2) oscillates. Discretization in time has been made according to the sensors sampling rate. For the integration in time of Pb. (
2), we have exploited a generalized-
method.
The FOM relies upon a number of dofs and , depending on the adopted (potentially fine) discretization. The RB method, which is used to construct the dataset , exploits a Galerkin-Proper Orthogonal Decomposition (POD) ROM, whose POD basis is built, starting from a set of FOM solutions (snapshots), which are computed within the prescribed parameters range. The ROM solution is then sought by solving the reduced-order problem resulting from the Galerkin projection of the FOM onto the reduced basis.
Displacements and temperatures that are related to the i-th sampling are time integrated and collected in and , respectively. The monitored dofs and , mimicking the sensors recordings, are extracted through two boolean matrices and , as and .
4. Numerical Test Case
The proposed approach has been assessed on the monitoring of the two-dimensional (2D) frame that is depicted in
Figure 1. When considering a structural thickness of
m, the plane stresses formulation has been adopted. The bottom edges are assumed perfectly clamped to the ground. The geometry has been discretized in 2938 constant strain triangle finite elements. The adopted mechanical and thermal properties are those of an ordinary reinforced concrete: Young modulus
GPa; Poisson ratio
; density
= 2500 kg/m
; thermal expansion coefficient
C
; stiffness thermal coefficient
C
. The last two proprieties allow for relating the local material temperature to the thermally induced anelastic deformations and the material stiffening/softening, respectively.
The structure is excited by low intensity seismic loads. To this aim, we have employed the empirical equations for predicting the attenuation of ground motion proposed in [
14] and implemented in [
15]. The main advantage of this tool is the possibility to generate random spectrum-compatible accelerograms as a function of: local magnitude
Q; epicentral distance
R; and, site geology. In order to ensure the structure to behave in elastic regime, those parameters have been limited within the following ranges:
;
km; rocky conditions. The parameters
Q and
R have been described by two uniform pdfs
and
, respectively.
Nine damage scenarios have been simulated by means of a localized stiffness reduction. Each structural state identifies a structural damage that occurs in the subdomain , , respectively; the damage-free baseline is labeled as . The occurrence of damage scenarios has been modeled with a (discrete) uniform pdf describing g. The damage level , which represents the intensity of the stiffness reduction to be applied, has been modeled by a uniform pdf .
In order to simulate potential thermal distributions, resembling those that are experienced by the structure, Pb. (
1) is solved under the action of thermal profiles that are imposed on all edges. Those profiles have been modeled by interpolating temperature values at edges midpoints and corners, which are nothing but the components of the parameter vector
. Temperatures have been assumed to be constant and equal for the three edges in contact with the ground, while parabolic profiles have been used elsewhere. In this way, a total of 28 parameters
(with
) are involved. From the temperature data of the city of Milan, a Gaussian pdf
,
, has been defined for each month, with
and
being the monthly average and standard deviation, respectively. Each time Pb. (
1) is solved, the month occurrence is sampled from a (discrete) pdf
, thus the
are inferred from the corresponding pdf
.
The sensor network consists of
sensors, recording structural accelerations
, and of
thermometers, recording temperatures
, with
, being arranged as depicted in
Figure 1. We have considered dual output sensors, recording both accelerations and temperatures at the same location. The dynamical response is monitored with a sampling frequency of 20 Hz, so as to sample the first two structural frequencies, respectively, 2.79 Hz and 7.14 Hz, without incurring in aliasing.
Thanks to the adoption of the ROM for the dataset construction, the number of dofs decreases from to 28 for the stationary diffusion problem and from to 63 for the elasto-dynamic problem. Consequently, the CPU time that is required by each simulation, over the time interval , passes from 421 s to s, entailing a speed-up of about 86 times (computations have been run on a PC featuring an Intel (R) Core™, i5 CPU @ 2.6 GHz and 8 GB RAM).
The evolution of the loss function and global accuracy of classifier
, obtained during the training, are respectively reported in
Figure 2 and
Figure 3. The iteration number accounts for the number of times the FCN weights have been updated. The greatest gains in terms of classification accuracy (i.e., the ratio of correctly classified instances over the total) are obtained in the first portion of the graph.
The generalization capabilities of the classifier
have been assessed on a test set made of 108
pseudo-experimental instances that are generated through the FOM. The classifier
carries out the classification task with a global accuracy of
and a testing time of
s, which means roughly about
s for each test instance. The obtained results are summarized by the confusion matrix in
Figure 4. Two different sources of error stand out. In particular, a small number of test instances labeled as
are misclassified as
(the same also occurs between
and
); this might be due to the similar influence of those scenarios on the mechanical behavior. Moreover, half of the test instances that are labeled as
(undamaged) are also misclassified; this might be due to the variability of
. Indeed, a low value of
not only implies an augmented difficulty in distinguishing between damaged and undamaged conditions, but it also causes the ROM to be less accurate.
5. Conclusions
In this work, we have proposed a computational framework, integrating model-order reduction and deep learning, for structural health monitoring under varying operational and environmental conditions. This hybrid model-data strategy enables online damage localization, making use of vibrational and temperature measurements. In order to overcome the lack of experimental data for civil applications, a database of synthetic recordings has been built offline, for a set of predefined damage scenarios, through simulations of a physics-based model, explicitly accounting for varying operational and environmental conditions. A parametric reduced order model, built through the reduced basis method, has been exploited in order to accelerate the dataset generation. Finally, a classifier exploiting a convolutional neural network has been adopted to perform automatic feature extraction and relate raw sensor data to the corresponding structural health conditions.
In the presented example, the classification outcomes show a global accuracy of about , and they offer the possibility to identify the nature behind the misclassification errors.
In future works, we aim to couple the classifier with a further neural network branch, playing the role of first line damage identifier, in order to reduce the possibility of an incorrect classification of undamaged scenarios. Moreover, a sensor placement according to a Bayesian optimization approach is going to be envisaged in order to maximize the effectiveness for the damage assessment.