Next Article in Journal
Modeling of Perforated Piezoelectric Plates
Next Article in Special Issue
Is NSGA-II Ready for Large-Scale Multi-Objective Optimization?
Previous Article in Journal
Three-Dimensional Non-Linearly Thermally Radiated Flow of Jeffrey Nanoliquid towards a Stretchy Surface with Convective Boundary and Cattaneo–Christov Flux
Previous Article in Special Issue
Multi-Objective Optimization of an Elastic Rod with Viscous Termination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scarce Sample-Based Reliability Estimation and Optimization Using Importance Sampling

Advanced Design, Optimization and Probabilistic Techniques (ADOPT) Laboratory, Department of Engineering Design, Indian Institute of Technology Madras, Chennai 600036, India
*
Author to whom correspondence should be addressed.
Math. Comput. Appl. 2022, 27(6), 99; https://doi.org/10.3390/mca27060099
Submission received: 15 September 2022 / Revised: 4 November 2022 / Accepted: 18 November 2022 / Published: 22 November 2022

Abstract

:
Importance sampling is a variance reduction technique that is used to improve the efficiency of Monte Carlo estimation. Importance sampling uses the trick of sampling from a distribution, which is located around the zone of interest of the primary distribution thereby reducing the number of realizations required for an estimate. In the context of reliability-based structural design, the limit state is usually separable and is of the form Capacity (C)–Response (R). The zone of interest for importance sampling is observed to be the region where these distributions overlap each other. However, often the distribution information of C and R themselves are not known, and one has only scarce realizations of them. In this work, we propose approximating the probability density function and the cumulative distribution function using kernel functions and employ these approximations to find the parameters of the importance sampling density (ISD) to eventually estimate the reliability. In the proposed approach, in addition to ISD parameters, the approximations also played a critical role in affecting the accuracy of the probability estimates. We assume an ISD which follows a normal distribution whose mean is defined by the most probable point (MPP) of failure, and the standard deviation is empirically chosen such that most of the importance sample realizations lie within the means of R and C. Since the probability estimate depends on the approximation, which in turn depends on the underlying samples, we use bootstrap to quantify the variation associated with the low failure probability estimate. The method is investigated with different tailed distributions of R and C. Based on the observations, a modified Hill estimator is utilized to address scenarios with heavy-tailed distributions where the distribution approximations perform poorly. The proposed approach is tested on benchmark reliability examples and along with surrogate modeling techniques is implemented on four reliability-based design optimization examples of which one is a multi-objective optimization problem.

1. Introduction

Reliability-based design optimization (RBDO) is a design approach that is used to generate reliable designs by accounting for uncertainties in the system variables such as material properties, loading conditions and geometry. Probabilistic approaches use probability distributions to model the uncertainties. In such approaches, the reliability of the system is measured as the probability of failure to satisfy a performance criterion. Mathematically, this involves calculating the hyper-volume of a multi-dimensional probability density function (PDF) under the failure region. This calculation becomes infeasible when the number of dimensions are large. Even in low dimensions, this calculation could become intractable due to complicated geometry of the failure region [1]. Attractive alternatives are analytical methods or sampling-based approaches.
Analytical methods, such as first-order and second-order reliability methods (FORM and SORM), transform the probability distributions to standard normal space and use linear or quadratic approximations of the performance function to estimate the failure probability at the most probable point (MPP) of failure [2,3,4]. In essence, analytical methods can estimate failure probability within a reasonable number of evaluations for linear or slightly non-linear performance functions. However, if the failure boundary is highly non-linear, analytical approaches are likely to lead to erroneous estimation. Additionally, other factors, such as island failure regions and multiple failure modes, limit their performance [5,6].
Sampling-based approaches such as Monte Carlo methods are highly effective for complex failure boundary problems and multi-modal failure problems [7]. In high-reliability applications where the failure probability is very low, MCS requires a very large number of model evaluations to obtain an accurate estimate. Most RBDO applications involve evaluating high-fidelity models which are computationally expensive, thus rendering MCS prohibitive for reliability estimation.
The computational burden of RBDO can be reduced by using surrogate-based methods wherein surrogate models which are cheaper to evaluate are constructed using limited high-fidelity model evaluations. Uncertainty from the random variables is then propagated through these surrogate models to obtain reliability estimates. Various surrogate modeling approaches, such as polynomial response surface (PRS), radial basis function (RBF), support vector machine (SVM) and Kriging among others have been adopted for reliability estimation [8]. Li and Xiu [9] proposed using cheaper to evaluate surrogates away from the limit state and high-fidelity model evaluations close to the limit state to improve accuracy and reduce cost. Dai et al. [10] proposed an SVM-based radial basis function to approximate the limit state function which is then used to estimate failure probability. Dubourg et al. [11] used error measure derived from Kriging to refine the surrogate during subset-simulation-based reliability estimation. Reliability estimation using surrogates may carry forward the bias from the surrogate approximation [12]. Surrogate modeling can also be used to build approximation models for reliability metrics instead of limit state function. Foschi et al. [13] used the combination of response surface, FORM and importance sampling to perform reliability estimation. Qu and Haftka [14] compared the accuracy of surrogates of different reliability metrics, such as failure probability, reliability index and probabilistic sufficiency factor (PSF) during RBDO. They conclude that inverse measure, such as PSF operating in performance space, performs better compared to a classical reliability metric. A survey of various surrogate-based RBDO frameworks is provided in [15]. The choice between surrogate for reliability estimation versus surrogate for limit state is a trade-off based on the number of design variables and the number of random variables as well as the cost of building the surrogate and cost of reliability estimate [16].
Employing variance reduction techniques to improve the efficiency and accuracy of Monte Carlo estimations is another way to alleviate the computational burden of RBDO. Several variance reduction techniques, such as importance sampling (IS) [17,18,19], subset simulation (SS) [1,20] and separable Monte Carlo (SMC) [21,22], are used to improve the accuracy of failure probability estimates while reducing the required number of high-fidelity model evaluations.
Importance sampling improves the accuracy of the estimate by drawing the sample realizations of the input random variables that have greater impact on the estimate, more often. In order to do that, an alternate sampling density known as importance sampling density (ISD) is chosen which enables sampling the important values more frequently. This introduces a bias in the estimator which is corrected by weighing the sample realizations. A good choice of ISD is proven to improve accuracy and in contrast, incorrect choice of ISD could lead to spurious estimates. Hence, the choice of ISD is critical, and several approaches have been explored to find the optimal ISD. Melchers [17] applied importance sampling for assessing reliability of parallel and series structural systems, where a multi-normal PDF centered at MPP was chosen as the ISD. Using the original distribution shifted to MPP, multivariate normal distribution located at MPP with various choices for the co-variance matrix ranging from the same as the original distribution to using only diagonal elements of the original co-variance matrix have been studied [23].
In cross-entropy-based methods [24], ISD is found by minimizing the Kullback–Leibler (KL) divergence between theoretical optimal ISD and a chosen family of distributions. Kurtz and Song [25] used a Gaussian mixture to obtain a non-parametric multi-modal PDF for ISD. It was observed that the coefficient of variation of the failure probability estimate converged as the number of Gaussian densities in the mixture increased. Cao and Choe [26] used an expectation-maximization (EM) algorithm to obtain a Gaussian mixture as the near optimal ISD. Cross-entropy information criterion (CIC) was used to select the number of Gaussian densities in the mixture. Geyer et al. [27] proposed a modified version of EM algorithm for updating the Gaussian-mixture-based ISD. For selecting the number of distributions in the mixture, density-based spatial clustering of applications with noise (DBSCAN) algorithm was used. In cross-entropy-based IS, one still requires the joint PDF of the original random variables to compute the near optimal ISD.
Kernel-based IS is another way to obtain the alternate sampling density wherein a kernel sampling density is constructed instead of choosing from a family of distributions [28,29]. Au and Beck [30] proposed Markov Chain Monte Carlo (MCMC) to distribute samples asymptotically according to the optimal ISD, and subsequently a kernel sampling density is constructed from these samples. Dai et al. [31] proposed a wavelet density estimation technique to construct the ISD from the MCMC samples. Botev et al. [32] proposed combining MCMC and IS to address the issues of biased sampling estimators that result from MCMC. Various adaptive importance sampling procedures have gained popularity recently. Dalbey and Swiler [33] proposed a Gaussian-process-based adaptive importance sampling where a Gaussian process surrogate is used to identify the likely regions of failure to adaptively improve the sampling density estimate. Zhao et al. [34] constructed a Kriging surrogate from the initial MCMC samples which is improved using an active learning process, and subsequently the surrogate is used for limit state evaluations for adaptive improvement of ISD. Wang and Song [35] used cross-entropy-based adaptive importance sampling for high-dimensional reliability analysis. Here, ISD is obtained by minimizing the KL divergence between a von Mises–Fisher mixture model and near optimal ISD.
In all the literature discussed above, while employing IS, determination of the appropriate ISD is considered to be the challenge. The probabilistic distributions and the corresponding parameters of the original random variables, however, are assumed to be known [22,36]. However, this is not necessarily always true. Here, we propose a framework to employ the importance sampling method for separable failure boundary of the form Capacity (C)–Response (R) using only scarce realizations of the R and C where no information about their distributions is known. We make use of tail-modeling techniques to approximate the cumulative distribution function (CDF) of the capacity and response. These approximations are used to locate a Gaussian ISD whose standard deviation is empirically chosen. Sample realizations drawn from the ISD are used to compute the failure probability. During this computation, kernel density estimates of PDF of response are utilized along with CDF approximation of capacity. Furthermore, bootstrap samples are generated from the original scarce samples of R and C. The proposed framework is applied on the bootstrap samples to obtain the confidence bounds for the reliability estimate. Reliability estimates obtained from scarce samples using the proposed approach are used to construct a surrogate of reliability index which is used for constraint evaluation during RBDO.
The rest of the paper is organized as follows. Section 2 presents the theoretical background of importance sampling in the context of reliability estimation. In Section 3, the methodology is discussed, and in later sections we present the results from the proposed method when applied to test cases and RBDO applications including a multi-objective optimization (MOO) application. Appendix A and Appendix B are used to provide short descriptions of kernel density estimation (KDE) and the third-order polynomial normal transformation (TPNT) technique.

2. Reliability Estimation Using Importance Sampling for Separable Limit States

In structural engineering, limit states are useful in prescribing performance requirements of a design. Thus, a limit state decomposes the design space of a system into safe and failure regions. Violation of a limit state is considered to be failure. and reliability is a measure of probability of such violations. For most structural problems, limit state can be expressed as the difference of Capacity ( C ) and Response ( R ) , as presented in Equation (1), where R and C are functions of independent sets of random variables. This is referred to as a separable limit state [21,22,37], and we consider such limit states in this work.
G ( C , R ) = C R
The system is said to fail when C R and safe when C > R . When either capacity or response or both are functions of variables that are random, a probabilistic measure such as the probability of failure is as presented in Equation (2)
p f = G 0 f C R ( c , r ) d c d r
where f C R is the joint probability distribution function (PDF), and  G 0 is the failure region. In the case of separable limit state, the joint PDF of capacity and response can be decomposed into the product of the marginal PDFs of capacity and response as presented in Equation (3)
p f = c r f C ( c ) f R ( r ) d c d r
p f = F C ( x ) f R ( x ) d x
where f C ( c ) and f R ( r ) are the marginal density functions of C and R, respectively. Equation (3) can also be written in a single integral form as presented in Equation (4). Here, F C ( c ) is the cumulative distribution function (CDF) of capacity. For low failure probabilities of order 10 4 , to obtain a Monte Carlo estimate with 10 % coefficient of variation, the sample size required is 100 × 1 p f = 10 6 . Generating 10 6 instances of expensive computer models is not feasible. To reduce such computational burden, importance sampling is used as in Equation (5).
p f I S = F C ( x ) f R ( x ) h x ( x ) h x ( x ) d x
where h x ( x ) is the alternate sampling density, and  p f I S is the failure probability computed using the importance sampling approach. It is the expectation of the integrand F C ( x ) f R ( x ) h x ( x ) computed with respect to the ISD, h x . Consequently, if  x 1 , x 2 , , x N is an independent identically distributed (i.i.d.) random sample from h x , then the expectation has an unbiased estimator:
p f I S ^ = i = 1 N F C ( x i ) f R ( x i ) h x ( x i ) , x i h x
Alternatively, if the marginal distribution of the response ( f R ( r ) ) is integrated in Equation (3), then the importance sampling estimator of p f would be expressed as:
p f I S ^ = i = 1 N ( 1 F R ( x i ) ) f C ( x i ) h x ( x i ) , x i h x
As mentioned in Section 1, in most of the literature, importance sampling is used when the distributions of R and C are known, and it becomes computationally expensive to draw samples in the region where tails of R and C overlap. Figure 1 presents a schematic of such a scenario where the limit state is of the separable form. Here, the region of interest lies in the tails of the distributions of R and C, and it contains the most probable point (MPP) of failure. Hence, it is only logical to locate the ISD at the MPP [3,17,22].
Figure 1 depicts a Gaussian ISD for which MPP serves as the mean of the distribution and is usually found by solving a constrained optimization problem. Once the mean of the ISD is determined, the spread of the distribution needs to be chosen. In order to have a finite variance for the importance sampling estimator, the ISD should not have a lighter tail than the original distribution [38]. In the literature [22,23], using the same co-variance matrix and sometimes using strictly the diagonal elements of the co-variance matrix of the original distribution is suggested. This has been shown to work for a range of possible limit states. However, sometimes a designer does not necessarily know the original distributions of input variables but has only a few realizations of R and C either through expensive computer simulations or physical experiments. Generally, response is considered to be the source of randomness since it is a measured quantity. However, capacity can also be a random variable. For instance, an example of capacity is yield strength which is a measured quantity. We know that it is random and hence modeled using probability distributions [39,40]. Other examples of capacity that are considered to be random include: member capacity under seismic loading [41], maximum flow capacity in measuring hydraulic reliability of water distribution system [42], and material fatigue properties [43,44].
In this work, we first characterize the distributions of R and C from limited samples and use the concept of importance sampling to estimate low failure probabilities. A Gaussian distribution is chosen as the ISD which is fully described by its two parameters, the mean and variance. Section 3 describes the procedure employed to identify the parameters of ISD. In the proposed framework, to demonstrate the methodology, distribution information of original input random variables is used only to generate the initial limited samples. However, the distribution information is not utilized anywhere afterwards.

3. Identifying Parameters of Gaussian ISD

Figure 2 presents an example scenario [23] of a separable limit state where the response and capacity follow normal distributions. It can be observed that the functional F C f R integrated in Equation (4) is maximum at the point ( x 11.55 ) , near the point of intersection between the PDF of response and the CDF of capacity, i.e.,  f R ( x ) = F C ( x ) . In importance sampling, MPP is used to locate the ISD to maximize sampling of the failure probability content. Hence, it is only logical to define this point of intersection as the MPP and use it as the mean of Gaussian ISD.
This point can be anywhere within the bounds of capacity and response. In high-reliability scenarios, it is located at the tails of one of the distributions or both, so we need to extrapolate into the tails of R and C. The available PDF approximation methods capture the central part of the distribution better than the tails, whereas the main focus of the tail modeling techniques applied for CDF estimation is at the tails of the distribution. Hence, we can reduce the errors in finding the MPP by using the intersection of 1 F R and F C instead of the intersection point of the curves f R and F C . Here, 1 F R is the complementary CDF of R. It is to be noted that the suitability of such a modification has been tested for only uni-modal type response. For multi-modal response, using a single MPP will only capture partial failure region. However, the approach can be extended by using a mixture of Gaussian distributions by using a summed failure probability integral as described in [45]. Algorithm 1 presents the bracketing procedure followed in the current work to find the MPP, while the same process is visually presented in Figure 3.
In Steps 3 and 6 of Algorithm 1, it is stated that the CDF approximations of capacity ( F C ^ ) and response ( F R ^ ) are obtained through a TPNT technique (details provided in Appendix B); however, it is an independent block in the proposed approach and hence can be replaced by any other suitable technique based on user discretion. In effect, the MPP finding problem reduces to finding the zeroes of the function F C ^ ( x ) ( 1 F R ^ ( x ) ) . Hence, root finding algorithms can also be applied to find the MPP. However, it is advised to use derivative-free approaches to avoid numerical issues.
Algorithm 1 Finding μ h .
1:
Obtain samples of response r = { r 1 , r 2 , , r M } and capacity c = { c 1 , c 2 , , c N } as in Figure 3a.
2:
for i = 1 , 2 , , M do
3:
     F C ^ ( r i ) , F R ^ ( r i ) TPNT approximations of CDF of response and capacity at response sample r .
4:
    Obtain point A in Figure 3b ← maximum of r x = { r i | F C ^ ( r i ) ( 1 F R ^ ( r i ) ) < 0 } .
5:
for i = 1 , 2 , , N do
6:
     F C ^ ( c i ) , F R ^ ( c i ) TPNT approximations of CDF of response and capacity at capacity sample c .
7:
    Obtain point B in Figure 3c ← minimum of c x = { c i | F C ^ ( c i ) ( 1 F R ^ ( c i ) ) > 0 } .
8:
Generate a set of evenly spaced points x = { x i : i = 1 , 2 , , 100 } between points A and B.
9:
Obtain mean of ISD, μ h maximum of x = { x i | F C ^ ( x i ) ( 1 F R ^ ( x i ) ) < 0 } . (Alternatively, minimum of x = { x i | F C ^ ( x i ) ( 1 F R ^ ( x i ) ) > 0 } can also be used.)
In very low failure probability estimation, the region corresponding to the estimation is quite small, and it has been observed that aspects such as non-uniqueness of MPP and non-linearity of the limit states have little effect on the estimation [5]. Similarly, we find that errors in the MPP estimation have lesser effect on the reliability estimation compared to a poor choice of standard deviation.
The standard deviation of the ISD determines the importance ascribed to region around the MPP during sampling. The region of sampling itself could be bounded by the supports of capacity and response distributions. Different measures of spread could be applied based on the available knowledge. In this work, a 10% coefficient of variation (CoV) was used to calculate the standard deviation of the ISD to investigate the effect. Such a measure was found to be suitable for scarce samples of normally distributed capacity and response. Different measures of spread were investigated while testing the formulation on samples simulated to be belonging to a different distribution. The spread parameter obtained through Equation (8) was found to be appropriate as such a measure allows one to restrict 68% percent of importance sample realizations within the means of capacity and response.
σ h = x ¯ C x ¯ R 2
where x ¯ C is the sample mean of capacity, and  x ¯ R is the sample mean of the response.

4. Estimation of Reliability and Its Confidence Bounds

p f ^ = i = 1 N F C ^ ( x i ) f R ^ ( x i ) h x ( x i ) , x i h x : N ( μ h , σ h 2 )
Using the now defined ISD, sample points are drawn at which the PDF of response and CDF of capacity are obtained to be used in the computation of failure probability as per Equation (6). However, instead of the actual values, approximations F C ^ ( x i ) , f R ^ ( x i ) are used during the computation as presented in Equation (9). As mentioned earlier, TPNT is used to approximate CDF ( F C ^ ), whereas for PDF approximation ( f R ^ ) an adaptive kernel density estimation method is employed. Both of these methods are distribution-free methods; however, other suitable methods of approximation [46,47] can also be used. Despite the accuracy of the method chosen, errors from the approximations are bound to result in loss of accuracy in the failure probability estimate. Hence, it becomes necessary to quantify the confidence on the estimate. Here, confidence bounds on the estimate are computed using a non-parametric bootstrap method.
The underlying idea of a non-parametric bootstrap method is to recreate samples from the original sample by sampling with the replacement. The sample size of bootstrap samples must be the same as the original sample. From each of the bootstrap samples, a statistical quantity of interest (such as failure probability) can be estimated. By repeating the process many times (say B times), one can obtain a distribution around the estimate from the original sample. In the current work, the proposed approach is applied to the bootstrap samples of C and R to obtain confidence bounds on the reliability estimate. Since the original samples of C and R are themselves scarce, this process of bootstrapping is repeated for T ( = 100 ) original samples of C and R. Thus, the quantiles of β ^ obtained from T ( = 100 ) iterations are compared with mean and standard deviation of quantiles of β ^ b o o t from the bootstrap samples. Figure 4 presents a schematic representation of the bootstrapping procedure followed in the current work.
In order to test the efficacy of the formulation, different tails for capacity and response were considered. To simulate samples of response and capacity belonging to different tail types, generalized extreme value (GEV) distribution is used which takes the shape parameter ( ξ ) as an input along with location ( θ ) and scale ( σ ) parameters. Shape parameter or tail index is a measure of heaviness of tails of a distribution. In GEV distributions, the shape parameter affects the lower-tail and upper-tail differently. In this study, the parameters are chosen such that response distributions are positively skewed, while capacity distributions are negatively skewed to better simulate the difficulties of sampling in tails. Individually, the shape parameters ( ξ r , ξ c ) correspond to three types of tail heaviness: heavy, medium and light. Here, tail heaviness is considered for the upper-tail of R and the lower-tail of C. The location parameter of the capacity distribution ( θ c ) is changed while keeping the response location ( θ r ) stationary so that each combination corresponds to a failure probability of 10 4 . Nine study cases result because of the different combination of tails possible for both R and C. The parameters for the nine study cases are presented in Table 1. The distribution parameters are only utilized to generate scarce samples of R and C. It is to be noted that the GEV distributions used here are continuous over the real line ( R ) and can have negative values. Though this does not reflect the real-world scenario, it does not deter in evaluating the performance of the proposed formulation.
To capture sampling variability, multiple iterations are carried out, and the procedure employed for identifying the variability in the estimates and bootstrap bounds is presented in Algorithm 2.
Algorithm 2 Confidence bounds using bootstrap.
1:
for j = 1 , 2 , , T do
2:
     r j × M G E V ( ξ r , θ r , σ r ) , c j × N G E V ( ξ c , θ c , σ c ) Generate samples of response and capacity.
3:
     p f ^ j Estimate failure probability by applying proposed approach to samples r j × M and c j × N .
4:
    for  i = 1 , 2 , , B  do
5:
         rb i × M , cb i × N Generate bootstrap samples from original samples r j × M and c j × N .
6:
         p f b ^ i Estimate failure probability by applying proposed approach to bootstrap samples rb i × M and cb i × N .
7:
25th, 50th, 75th percentiles { p f ^ } T × 1 .
8:
25 th 1 × T , 50 th 1 × T , 75 th 1 × T percentiles { p f b ^ } B × T .
β I S = Φ 1 ( 1 p f ^ )
The failure probability estimates calculated through Algorithm 2 for each study case are converted into reliability indices using Equation (10). To facilitate comparison, the estimates are divided by the actual reliability index ( β a = 3.71 ) . For all nine study cases, the sample size of response ( M ) and capacity ( N ) are considered to be 50. The results of applying the formulation to each study case are presented in Table 2. Values under the ‘Original sample’ column represent the percentiles of ratio of the estimates from the original samples (from Step 7 of Algorithm 2) repeated for T (=100) iterations. Mean and standard deviation of the percentiles (from Step 8 of Algorithm 2) from the bootstrap repetitions (B = 100) are presented under the ‘Bootstrap’ column. An accurate estimate of the reliability index would be indicated with a value of one, and high precision is indicated by low variability between the percentiles and low standard deviation in the bootstrap percentiles. It is observed that the estimation is poorer in the case of heavy-tailed response and medium-tailed capacity and heavy-tailed response and light-tailed capacity. In both these cases, the failure region is situated further into the tail of the heavy-tailed response where the scarce sample-based PDF estimation is prone to high errors. In many instances, the probability density drops to zero prematurely which results in overestimation of reliability.
In certain reliability applications, either f R or f C is known, or it is easier to obtain samples from one of the distributions of R and C. The proposed method is tested for such scenarios where only one of the distributions is unknown, and during this exercise, the same combinations of tails for R and C are assumed. Table 3 and Table 4 present the results for R unknown and C unknown cases, respectively, using the proposed approach. From Table 3, it is observed that both the accuracy and precision of the estimation improved for most of the tail combinations compared to both R and C unknown cases. In the case of heavy-tailed response and medium-tailed and light-tailed capacity, and in addition, medium-tailed response and light-tailed capacity, the bias in the estimate has increased. This is again due to the erroneous PDF estimation of heavier-tailed response distribution using a scarce sample. Lesser bias in both R and C unknown cases is due to the interaction between the PDF and CDF approximation.
In Table 4, larger errors are observed for heavy-tailed capacity and medium-tailed response and heavy-tailed capacity and light-tailed response. This suggests that estimating heavy-tailed distribution further into the tails results in larger errors; this is akin to the observations made for the R unknown case. Furthermore, the largest errors observed in the C unknown case are smaller compared to the largest errors from the R unknown case which suggests that TPNT is a better approximation for CDF of heavy tailed distributions compared to adaptive KDE used for PDF approximation. To demonstrate that this is indeed the case, the alternate form of the importance sampling estimator presented in Equation (7) is used for the R unknown scenario, wherein CDF of the response distribution is approximated instead of the PDF. Table 5 presents the results, from which it is observed that the bias has reduced compared to the PDF approximation-based estimation. However, this also results in underestimation of reliability in the case of light-tailed response and light-tailed capacity. This is a trade-off between the choice of adaptive-kernel-based PDF approximation and TPNT-based CDF approximation. From these observations, it can be surmised that it would be beneficial to know the heaviness of R and C so that appropriate estimator can be chosen based on the efficacy of the approximation methods available.
An additional observation is that for one (either R or C) unknown case, the bootstrap standard deviation is larger when the bias in the estimates from the original samples are larger and smaller for more accurate cases. Thus, in practice where there is no actual value to compare with, bootstrap standard deviation can be used to discern the accuracy of the estimate. However, this does not track well in the case of both R and C unknown.

Tail-Index Estimation

Estimation of tail index is sometimes part of the distribution identification process and in turn CDF estimation where methods such as Hill estimator [48], Pickands estimator, Generalized Pareto fits and others are applied. Hill estimator considers k upper order statistics from a sample to evaluate the tail heaviness. Equation (11) presents the estimator for a positive sample of size n with the order statistics, X 1 , n X 2 , n X n , n .
H k , n = 1 k i = 0 k 1 log X n i , n X n k , n
The estimate of tail heaviness is sensitive to the number of order statistics ( k ) used and shifting of sample. Different modifications that address such sensitivity issues exist in the literature [49,50]. As we only require comparing between tails of R and C, especially when they are very different, obtaining exact estimates is of low priority. Hence, a modified Hill estimator is used where the sample is mean-shifted, and k is considered to be 10 % of the sample size. For upper-tail estimation, the largest k values are used and for lower-tail estimation, the absolute values of the smallest k values are used. The Modified Hill estimator is applied to assess the tail-heaviness of upper-tail of response and lower-tail of capacity, and the sample with the heavier tail is chosen for CDF approximation, and thereby the appropriate form of the estimator is chosen for failure probability estimation.
In both unknown scenarios, the heavy-tailed R and the light-tailed C which showed the largest deviation from the actual value was chosen to test the tail index estimate-based improvement on the proposed formulation. Results presented in Table 6 show the use of tail index estimation improved the estimates. It is to be noted that the modified Hill estimator was successful in contrasting between heaviness of tails of R and C 81 out of 100 iterations.

5. Reliability Estimation Examples

Based on the flowchart of the proposed approach in Figure 5, benchmark reliability estimation examples are tested in this section. In all examples, to account for sample variability 100 iterations are used with samples of size 50 for R and C in each iteration. For each iteration, bootstrap samples for R and C are generated 100 times.

5.1. Example 1: Concave Limit State 1

This is a concave limit state example taken from [51].
G = 2.62 u 2 0.15 u 1 2
Both u 1 and u 2 are standard normal variables, and the concave limit state is of the separable form. Here, R is taken as 0.15 u 1 2 , and C is taken as 2.62 u 2 .
From tail-index estimates in 88 out of 100 iterations, response R is determined to be heavier-tailed than C which is consistent with the analytical form of R and C. The response distribution resulting from squaring a standard normal variable is an χ 2 -distribution with one degree of freedom. In this case, the  χ 2 -distribution has a heavier tail than the standard normal distribution.
Table 7 provides the results obtained for a target reliability β t = 2.39 . It is observed that the true value is contained within the percentile bounds (0.84–1.12) from the original sample estimates, and the bootstrap bounds obtained for each percentile also contain the true value. The variability of the estimates from the original sample is high, and the standard deviation computed from the bootstrap estimates is reflective of the variability presented in the original sample estimates.

5.2. Example 2: Concave Limit State 2

Example 2 is a concave limit state taken from [51].
G = u 1 2 5 u 2 2 + 45
Both u 1 and u 2 are standard normal variables, and the concave limit state is of the separable form where R is taken as 5 u 2 2 and C is taken as u 1 2 + 45 .
Table 8 presents the numerical results obtained for this example by applying the proposed IS approach. From tail-index estimates, in 100 out of 100 iterations, response R is determined to be heavier-tailed than C. Here, both the response and capacity distributions resulting from squaring standard normal variables u 1 and u 2 are χ 2 -distributions. However, the upper tail of response is heavier because scaling a χ 2 -distribution results in an increase of tail heaviness.
The percentiles of M M β I S / β t from the original samples show that the target value is contained within the first and third quartile bounds. For all the percentiles, it is observed that the bootstrap bounds also contain the target reliability.

5.3. Example 3: Roof Truss Example

This example is discussed as a case study for reliability estimation in [45]. The schematic in Figure 6 presents a roof truss subjected to a uniformly distributed load q which is transformed into the nodal load P = q l 4 . Equation (14) presents the limit state constructed for perpendicular deflection Δ C at node C as:
G ( x ) = 0.03 Δ C
where
Δ C = q l 2 2 3.81 A c E c + 1.13 A s E s
Here, A c and A s are areas of cross sections of the concrete reinforced bars and steel bars, respectively. Similarly, the variables E c and E s represent the moduli of elasticity, while l denotes the length. The input variables are considered to be mutually independent random variables, distributed normally. The parameters of their distributions are as presented in Table 9.
The limit state is transformed into G ( C , R ) = C R form where C = 0.03 q and R = Δ C q becomes a function of five random variables that correspond to the geometry and material properties of the truss members.
Table 10 provides the estimates of reliability and the corresponding bootstrap bounds for a target reliability of β t = 3.4 . It is observed that at each of the percentiles, the bootstrap bound captures the target reliability. From tail-index estimates, in 77 out of 100 iterations, response R is determined to be heavier-tailed than C.

5.4. Example 4: Propped Cantilever Beam Example

Figure 7 presents the schematic of the example which is taken from [52]. Equation (16) presents the original limit state for the maximum deflection of the beam ν m a x against the maximum allowable deflection ν c r i t :
G = ν m a x ν c r i t
where deflection of the beam ν ( x ) is measured as per Equation (17), and for the considered loading condition, the maximum deflection ( ν m a x ) is obtained at x = 0.5528 L .
ν ( x ) = q 0 x 2 120 L E I ( 4 L 3 8 L 2 x 5 L x 2 x 3 )
where
I = b f d 3 ( b f t w ) ( d 2 t f ) 3 12
The limit state is modified to convert into G ( C , R ) = C R form, where C ( ν c r i t , E ) = ν c r i t E , and R is a function of remaining seven random variables. Mean and standard deviation (SD) of the normally distributed random variables is presented in Table 11. The critical displacement ν c r i t is changed between 4–5 mm to generate three target reliability situations.
Table 12 presents the results obtained from the proposed IS formulation for different target reliability indices ( β t ) . As the target reliability increases, the variability from both the original sample estimates and bootstrap sample estimates remain similar.

6. Application to RBDO Examples

The proposed importance sampling approach is demonstrated on two benchmark and two real-world RBDO examples. Algorithm 3 delineates the steps followed to perform RBDO using the proposed approach of reliability estimation. For the benchmark examples, efficiency between the proposed approach and a crude Monte Carlo approach is compared using the total number of limit state evaluations which is computed as N d o e × N i s for the proposed approach and N d o e × N m c s for the MCS-based approach. For the real-world examples, MCS is used only to validate the optima obtained through the proposed approach.
Algorithm 3 RBDO using proposed importance sampling approach.
1:
d = { d 1 , d 2 , , d N d o e } Generate a design of experiment (DoE) in design variable space.
2:
for  i = 1 , 2 , , N d o e do
3:
    Propagate uncertainty at DoE points.
4:
     r , c Obtain response and capacity samples through simulations or physical experiments. In the examples, analytical functions are utilized.
5:
     p f ^ i Apply proposed importance sampling formulation to samples r , c .
6:
     β I S i Transform failure probability to reliability index using (10).
7:
β IS ^ Construct surrogate model for reliability index using { β I S i } and d .
8:
d * Optimization using β IS ^ for evaluation of reliability constraint.

6.1. Cantilever Beam Example

This is a standard RBDO example taken from [53]. In this example, the weight of a cantilever beam (shown in Figure 8) is minimized while considering two failure modes: bending stress does not exceed yield strength (19) and the tip displacement does not exceed the allowable displacement limit of 2.5 i n (20).
G s = σ y 6 L w t X w + Y t
G d = D 0 4 L 3 E w t Y t 2 2 + X w 2 2
The length of the beam ( L ) and the density are held constant, while the width ( w ) and thickness ( t ) of the beam are allowed to change which transforms the objective function from weight to area of cross section, A = w t . The horizontal ( X ) and vertical ( Y ) loads along with modulus of elasticity ( E ) are random variables whose uncertainty characteristics are presented in Table 13.
RBDO of the cantilever beam requires that the failure probability of both the limits states does not exceed 1.35 × 10 3 which translates to a target reliability index β t 3 . Thus, the optimization is formulated as:
min w , t A = w t s . t . g 1 = Φ 1 ( 1 P r ( G s 0 ) ) 3 g 2 = Φ 1 ( 1 P r ( G d 0 ) ) 3
As stated in Algorithm 3, a DoE of size ( N d o e = 40 ) is created in d = ( w , t ) space using latin hypercube sampling (LHS). Additionally, four corner points are added to the DoE. At each design point ( d i ) , samples of response and capacity for both limit states are obtained by simulating uncertainty as per Table 13, considering a sample size of N i s = 50 . For limit state G s , σ y is considered as capacity, while the rest of the equation is considered as a response. Similarly, for limit state G d , D 0 × E is taken as capacity by regrouping the random variables in the limit state function to convert it into a separable form [54]. Using these samples, failure probability estimates for both limit states are obtained through the proposed approach. A surrogate model for reliability index β is constructed, and this surrogate is used for constraint evaluation during optimization.
In a similar fashion, MCS-based failure probability estimates are also obtained at the design points using samples of size N m c s = 10 6 . A surrogate model using the MCS estimates is constructed for a reliability index, and optimization is carried out. At design points that are closer to the lower and upper bounds, failure probability estimates could be either zeros or ones. These estimates when converted into reliability indices become and ; hence, these are modified during surrogate construction. These singularities are more common in the case of MCS estimation compared to IS estimation.
The optima obtained from both MCS and IS are compared in Table 14. As the cost of the knowledge of random variables is not quantified, computational cost is only compared with MCS. Additionally, different surrogate choices, such as PRS, RBF, Kriging and weighted average surrogate (WAS), were considered for fi ^ construction, and the optima obtained by using WAS model are presented. The accuracy of the surrogate for both constraints is evaluated using a generalized mean square error (GMSE) metric. GMSE for constraint g 1 using IS-based estimates is reported as 0.91 (mean of reliability estimates = 2.21), whereas the error from the MCS-based surrogate is 0.09 (mean of reliability estimates = 1.97). For constraint g 2 , GMSE (vs. mean) is reported as 0.70 (vs. 1.98) for the IS surrogate and 0.12 (vs. 1.41) for the MCS surrogate.
To validate the results, at the optima, a reliability index is calculated using MCS with sample size of 10 7 which is also presented in the Table 14. It is observed that the surrogate from IS estimates has less global accuracy. However, reliability indices obtained at the optima using MCS ( β M C S at d * ) suggest that the surrogate is reasonably approximated at the constraint boundaries. The optima obtained from IS and MCS are very close, but the computational savings are hugely in favour of IS ( 50 vs . 10 6 ) .

6.2. Bracket Structure Example

This is originally from [55] where a bracket structure is subjected to a tip load ( P ) in addition to its own weight due to gravity (g) as presented in Figure 9. The weight of the bracket structure is optimized while considering two failure modes:
(i)
Maximum bending stress of beam CD at point B ( σ B ) does not exceed its yield strength ( f y ) ,
(ii)
Maximum axial load on beam AB ( F A B ) does not exceed the Euler critical buckling load ( F b u c k l i n g ) .
Equations (22) and (23) present the limit states as:
G C D = f y σ B , where
σ B = 6 M B w C D t 2
M B = P L 3 + ρ g w C D t L 2 18
G A B = F b u c k l i n g F A B , where
F b u c k l i n g = π 2 E t w A B 3 9 sin θ 2 48 L 2
F A B = 1 cos θ 3 P 2 + 3 ρ g w C D t L 4
During RBDO of the bracket structure, the target reliability index for both limit states is β 2 , and the design parameters are means of the geometrical parameters of the structure: width of CD ( μ w C D ) , width of AB ( μ w A B ) and thickness of AB and CD ( μ t ) which are bounded between 50 mm and 300 mm . The uncertainty characteristics of the random variables is presented in Table 15. Thus, the RBDO is formulated as:
min w C D , w A B , t C = μ ρ μ t μ L 4 3 9 μ w A B + μ w C D s . t . g 1 = Φ 1 ( 1 P r ( G C D 0 ) ) 2 g 2 = Φ 1 ( 1 P r ( G A B 0 ) ) 2 50 μ w C D , μ w A B , μ t 300 ( in mm )
The steps enumerated in Algorithm 3 are followed using a DoE of size ( N d o e = 60 ) , and scarce sample sizes of R and C during the IS approach are considered as N i s = 75 , while N m c s = 10 6 realizations are used in MCS. The RBDO results are presented in Table 16 which has the same format as the first example. GMSE (vs. mean) for constraint g 1 using IS-based estimates is reported as 0.58 (1.37), whereas the error from the MCS-based surrogate is 0.24 (1.02). For constraint g 2 , GMSE (vs. mean) is reported as 0.97 (vs. 3.72) for the IS surrogate and 0.30 (vs. 3.08) for the MCS surrogate. It is observed that the IS estimate-based surrogate has less global accuracy; however, the MCS-based reliability indices computed at the optima show that it approximates reasonably well near the constraint boundaries. It is to be noted that in both engineering examples, we assume that both R and C are unknown. Any knowledge of the uncertainty characteristics of either R or C will only improve the accuracy of the reliability estimates.

6.3. Torque Arm Example

This example presents RBDO of torque arm where the mass of the component is to be minimized adhering to a probabilistic constraint on the allowable stress. Unlike previous examples, there is no analytical expression available for limit state evaluation in this case. It is a shape optimization problem originally from Bennett and Botkin [56]. Researchers have used it as a benchmark example for reliability estimation [52,57]. Rahman and Wei [58] perform RBDO where a constant allowable stress limit is considered.
In the current study, seven design variables (see Figure 10) as per [59] are considered for altering the shape of the torque arm. Figure 11 presents the base design of the torque arm around which the optimization is to be performed. Here, a horizontal load ( F x = 2789 N) and a vertical load ( F y = 5066 N) are applied at the right hole while the left hole is fixed. The torque arm has modulus of elasticity E = 207 × 10 5   N · cm 2 , density ρ = 7.850 × 10 3   kg · cm 3 and Poisson’s ratio ν = 0.3 .
Equation (25) presents the limit state equation for the torque arm.
G = σ m a x ( d i , F x , F y ) σ a l l
where σ m a x is the maximum von Mises stress developed in the torque arm, and σ a l l is the allowable stress limit. Since there is no analytical expression that relates the design variables with the response σ m a x , finite-element analysis is used to compute the stresses developed in the torque arm for a given loading condition. A MATLAB finite-element toolbox developed by CALFEM [60] is used for this purpose. The thickness of the finite-elements in the mesh is considered to be 0.3 cm. At the end of each finite-element analysis, the maximum of the stresses is used as σ m a x in Equation (25).
For RBDO, the design variables ( d 1 to d 7 ) , loads ( F x and F y ) and the allowable stress ( σ a l l ) are considered to be uncertain. Each design point ( d i ) is considered to be normally distributed about itself with a coefficient of variation of 10%. The uncertainty characteristics of the remaining random variables is presented in Table 17.
Equation (26) presents the RBDO formulation of the torque arm where a target reliability index of three is considered.
min { d 1 , , d 7 } M a s s s . t . g = Φ 1 ( 1 P r ( G 0 ) ) 3
As per Algorithm 3, a DoE of size ( N d o e = 200 ) is generated where reliability is estimated using an IS approach. Sample sizes of R = σ m a x and C = σ a l l during IS approach are considered as N i s = 100 . An RBF-based surrogate is constructed using IS-based reliability estimates. The error (GMSE vs. range) for the RBF surrogate of β is found to be 6% which indicates a good fit. This surrogate is used for constraint (g) evaluation during optimization.
The stress contour of the torque arm design obtained as a result of the optimization is presented in Figure 12. Here, mean values for design parameters and mean loading condition are considered. The maximum von Mises stress is observed to be 523.92 MPa for the optimum design. The optimal mass of the torque arm is 0.801 kg. It can be observed that the mass is distributed to meet the target reliability. An MCS (with 10 5 sample size) is used to validate the reliability of RBDO optima obtained from the IS approach, and it is observed that β M C S = 3.00 . For the Monte Carlo simulation of 10 5 simulations, it took 8.7 h using parallel computing toolbox of MATLAB on a system with the following specifications: Intel Xeon 10 Core 2.20 GHz 64 bit processor, with 32GB RAM. Table 18 presents the design variable bounds and the optimum design obtained from RBDO.
Though the errors in the individual reliability estimations are quantified via bootstrap, they were not propagated into the surrogate model during RBDO. Using the surrogate model for reliability index evaluation instead of direct evaluation enabled the smoothing of noise from the IS-based estimation which leads to better convergence during optimization. From MCS validation at the optima, it is observed that the proposed approach results in heavier but safer designs. Further analysis is required to understand the error propagation from different stages of the proposed approach.

6.4. Car Side-Impact Problem—A Multi-Objective Reliability-Based Design Optimization (MORBDO) Example

We demonstrate the proposed methodology on an MORBDO example taken from [61]. In this example, the objective is to minimize the weight ( f 1 ) of a car as well as the average rib deflection ( f 2 ) during a crash. A car is subjected to a side-impact based on European Enhanced Vehicle-Safety Committee (EEVC) procedures. The effect of the side-impact on a dummy in terms of head injury criteria, load in the abdomen, pubic symphysis force, viscous criterion, and rib deflections at the upper, middle and lower rib locations are considered as constraints. The MORBDO formulation is made up of seven uncertain design variables ( x 1 , , x 7 ) and four random variables ( p 1 , , p 4 ). Equation (27) presents the optimization formulation of the car side-impact problem:
min μ x f 1 = f ( μ x , μ p ) min μ x f 2 = g 2 ( μ x , μ p ) + g 3 ( μ x , μ p ) + g 4 ( μ x , μ p ) 3 s . t . Φ 1 1 P r g i ( x , p ) b i β t , i = 1 , , 10 .
where
g 1 ( x , p ) Abdomen load 1 kN g 2 ( x , p ) Upper rib deflection 32 mm g 3 ( x , p ) Middle rib deflection 32 mm g 4 ( x , p ) Lower rib deflection 32 mm g 5 ( x , p ) Upper viscous criteria 0.32 m / s g 6 ( x , p ) Middle viscous criteria 0.32 m / s g 7 ( x , p ) Lower viscous criteria 0.32 m / s g 8 ( x , p ) Pubic symphysis force 4.0 kN g 9 ( x , p ) Velocity of B - pillar at middle point 10 mm / ms g 10 ( x , p ) Velocity of front door at B - pillar 15.7 mm / ms ;
1.0 x 1 1.5 , 0.45 x 2 1.0 , 0.5 x 3 1.5 , 0.5 x 4 1.5 , 0.875 x 5 2.625 , 0.4 x 6 1.2 , 0.4 x 7 1.2 , μ p 1 = 0.345 , μ p 2 = 0.192 , μ p 3 = μ p 4 = 0 .
Analytical expressions of the objective functions and constraints as well as the physical descriptions of the design variables ( x 1 x 7 ), and random variables ( p 1 p 4 ) are presented in Appendix C.
In order to solve the MORBDO problem as per Algorithm 3, a DoE of size ( N d o e = 200 ) is generated within the design bounds where reliability of the ten constraints ( g 1 , , g 10 ) is estimated using the IS approach. In this example, the number of design variables is seven; hence, 2 7 = 128 corner points have been sampled. Next, we performed a space-filling sampling using LHS. In each dimension, we sampled 10 points (in total 7 × 10 = 70 ) using LHS design as per a thumb rule in DoE sampling [62]. Without loss of generalization, we added two to make a round sampling number. Sample sizes of response and capacity during IS are considered as N i s = 50 . PRS surrogate for reliability indices of the ten constraints is constructed using the IS-based estimates. In order to improve the surrogate accuracy, an additional DoE of size 200 is generated within adjusted design bounds. The final surrogate errors (GMSE vs. range) for all constraints is found to range between 3.76 % and 8.38 % . Using these surrogates for constraint evaluation, the MOO problem presented in Equation (27) is solved using NSGA-II [63].
The Pareto optimal front corresponding to the different target reliability indices ( β t = [ 2.5 , 3.0 , 3.5 ] ) along with the deterministic Pareto optimal front is presented in Figure 13. The NSGA-II algorithm is applied for the four instances using the population size of 200 for 100 generations. The GA uses the following parameters: number of offspring is 50, probability ( p c ) of simulated binary crossover (SBX) is 0.9, crossover parameter ( η c ) is 20, probability of mutation ( p m ) is 0.9, and the mutation parameter ( η m ) is 50.
It is evident from Figure 13 that as the targeted reliability increases, the respective reliable Pareto optimal front shifts inside the feasible criterion space and away from the deterministic Pareto optimal front to ensure more reliable solutions. The Pareto solutions were further validated using MCS ( N m c s = 10 6 ), and all the solutions were found to meet the target reliability.

7. Conclusions

A scarce sample-based importance sampling approach to estimate reliability is proposed when there is little or no information about the uncertainty characteristics of the random variables involved. The proposed formulation was tested on different tail heaviness for R and C distributions. In the case of one of the distributions being heavy-tailed, a tail-index estimate-based improvement to choose between PDF and CDF approximation was employed and shown to improve the accuracy (50th percentile by 1.9 times). Confidence bounds on the reliability estimate obtained through the bootstrap procedure have been shown to be indicative of the accuracy of the estimate. The proposed IS approach has been applied for reliability estimation and RBDO examples and found to be effective in terms of computational savings ( 50 × 44 = 2200 for cantilever beam and 75 × 60 = 4500 for bracket structure example) as compared to MCS where the sample size for each reliability estimate was 10 6 . The approach is demonstrated on a non-analytical RBDO example which yielded a design that met the target reliability index (validated by MCS). The proposed approach has also been demonstrated on the car side-impact problem which is a multi-objective reliability-based design optimization example.
While the tail-index estimate-based alternative reduces the errors from the approximation, establishing the superiority of CDF approximation versus PDF approximation was only achieved through post-analysis of results for the specific methods used in this work. Future work could include incorporating region-wise best methods for both PDF and CDF approximation, using a suite of methods assessed through cross-validation errors. Active learning approaches could be employed during optimization to further reduce the number of design points, thereby reducing the number of reliability estimations required to obtain the optima.

Author Contributions

Conceptualization, K.P. and P.R.; investigation, K.P. and D.Y.; methodology, K.P. and P.R.; supervision, P.R.; validation, K.P. and D.Y.; writing—original draft, K.P.; writing—review and editing, K.P., D.Y. and P.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

MATLAB® codes used to generate the results shall be provided upon request.

Conflicts of Interest

The authors declare that they have no conflict of interest.

Appendix A. Kernel Density Estimation (KDE)

Let ( x 1 , x 2 , . . . , x n ) be an i.i.d. sample from a distribution with a PDF ( f ) , then the kernel density estimate of f is given as
f ^ h ( x ) = 1 n h i = 1 n K x x i h
where h is a smoothing parameter called bandwidth, n is the sample size, and K ( . ) is a kernel, which is non-negative, integrates to one and is centered at zero. Different kernel functions can be used, such as normal, uniform, Epanechnikov, triangle and others, and the bandwidth is selected based on the sample data chosen. The choice of the bandwidth influences the variance and the bias of the estimator. The performance of the kernel density estimator is more dependent on the choice of the bandwidth rather than kernel choice. Despite being the most popular non-parametric approach to density estimation, there are some implementation issues, such as bandwidth selection, local adaptivity and boundary bias. While KDE works well for data following a normal distribution, it performs poorly while estimating heavy-tailed distributions, especially in the tail region which is our region of interest, hence adaptive KDE proposed by [64] is chosen. The Matlab® implementation of adaptive KDE for 1D [65] was employed in the current work for PDF approximation.

Appendix B. Third-Order Polynomial Normal Transformation Technique (TPNT)

Hong and Lind [66] proposed this method of approximating a CDF where given the order statistics ζ 1 ζ 2 ζ N obtained from a sample realization of a random variable Z, through “sample rule" the fractiles are constrained in the following manner:
{ ζ i , F Z ( ζ i ) } = { ζ i , i N + 1 } , i = 1 , 2 , , N
where F Z ( . ) is the cumulative distribution function of Z. In this method, a third-order polynomial relationship between ζ and a normal transformation of F Z ( ζ ) is assumed as presented in Equation (A3)
ζ = k = 0 3 a k η k
where
η = Φ 1 ( F Z ( ζ ) )
Here, Φ 1 ( . ) is the inverse of the standard normal distribution function. The coefficients of the polynomial in Equation (A3) are found through least squares minimization of the error, ε :
ε = j J s ζ j k = 0 3 a k ( η j ) k
where J s is a set of data points chosen for the parameter estimation which is usually the same as sample size N. Two constraints a 2 2 3 a 1 a 3 > 0 , a 3 > 0 are imposed to ensure monotonicity in the third-order polynomial curve.
At a new fractile ζ 0 , the probability F Z ( ζ 0 ) is determined through Φ ( η 0 ) , where η 0 is obtained by solving Equation (A3), with substitution of ζ by ζ 0 .

Appendix C. Car Side-Impact Problem

The analytical expression of the objective function and constraint functions are given below:
f 1 ( x ) = 1.98 + 4.9 x 1 + 6.67 x 2 + 6.98 x 3 + 4.01 x 4 + 1.78 x 5 + 0.00001 x 6 + 2.73 x 7 , g 1 ( x , p ) = 1.16 0.3717 x 2 x 4 0.00931 x 2 p 3 0.484 x 3 p 2 + 0.01343 x 6 p 3 , g 2 ( x , p ) = 28.98 + 3.818 x 3 4.2 x 1 x 2 + 0.0207 x 5 p 3 + 6.63 x 6 x 9 7.7 x 7 x 8 + 0.32 p 2 p 3 , g 3 ( x , p ) = 33.86 + 2.95 x 3 + 0.1792 p 3 5.057 x 1 x 2 11 x 2 p 1 0.0215 x 5 p 3 9.98 x 7 p 1 + 22 p 1 p 2 , g 4 ( x , p ) = 46.36 9.9 x 2 12.9 x 1 p 1 + 0.1107 x 3 p 3 , g 5 ( x , p ) = 0.261 0.0159 x 1 x 2 0.188 x 1 p 1 0.019 x 2 x 7 + 0.0144 x 3 x 5 + 0.0008757 x 5 p 3 + 0.08045 x 6 x 9 + 0.00139 p 1 p 4 + 0.00001575 p 3 p 4 , g 6 ( x , p ) = 0.214 + 0.00817 x 5 0.131 x 1 p 1 0.0704 x 1 p 2 + 0.03099 x 2 x 6 0.018 x 2 x 7 + 0.0208 x 3 p 1 + 0.121 x 3 p 2 0.00364 x 5 x 6 + 0.0007715 x 5 p 3 0.0005354 x 6 p 3 + 0.00121 p 1 p 4 + 0.00184 x 9 p 3 0.018 x 2 x 2 ,
g 7 ( x , p ) = 0.74 0.61 x 2 0.163 x 3 p 1 + 0.001232 x 3 p 3 0.166 x 7 p 2 + 0.227 x 2 x 2 , g 8 ( x , p ) = 4.72 0.5 x 4 0.19 x 2 x 3 0.0122 x 4 p 3 + 0.009325 x 6 p 3 + 0.000191 p 4 p 4 , g 9 ( x , p ) = 10.58 0.674 x 1 x 2 1.95 x 2 p 1 + 0.02054 x 3 p 3 0.0198 x 4 p 3 + 0.028 x 6 p 3 , g 10 ( x , p ) = 16.45 0.489 x 3 x 7 0.843 x 5 x 6 + 0.0432 p 2 p 3 0.0556 p 2 p 4 0.000786 p 4 p 4 .
Description of the design variables ( x 1 x 7 ) and random variables ( p 1 p 4 ) (standard deviation in bracket)
x 1 = Thickness of B - Pillar inner ( 0.03 ) , x 2 = Thickness of B - Pillar reinforcement ( 0.03 ) , x 3 = Thickness of floor side inner ( 0.03 ) , x 4 = Thickness of cross members ( 0.03 ) , x 5 = Thickness of door beam ( 0.05 ) , x 6 = Thickness of door belt line reinforcement ( 0.03 ) , x 7 = Thickness of roof rail ( 0.03 ) , p 1 = Material of B - Pillar inner ( 0.006 ) , p 2 = Material of floor side inner ( 0.006 ) , p 3 = Barrier height ( 10 ) , p 4 = Barrier hitting position ( 10 ) .

References

  1. Au, S.K.; Beck, J.L. Estimation of small failure probabilities in high dimensions by subset simulation. Probabilistic Eng. Mech. 2001, 16, 263–277. [Google Scholar] [CrossRef] [Green Version]
  2. Hohenbichler, M.; Gollwitzer, S.; Kruse, W.; Rackwitz, R. New light on first- and second-order reliability methods. Struct. Saf. 1987, 4, 267–284. [Google Scholar] [CrossRef]
  3. Schuëller, G.I.; Stix, R. A critical appraisal of methods to determine failure probabilities. Struct. Saf. 1987, 4, 293–309. [Google Scholar] [CrossRef]
  4. Rackwitz, R. Reliability analysis-a review and some perspective. Struct. Saf. 2001, 23, 365–395. [Google Scholar] [CrossRef]
  5. Engelund, S.; Rackwitz, R. A benchmark study on importance sampling techniques in structural reliability. Struct. Saf. 1993, 12, 255–276. [Google Scholar] [CrossRef]
  6. Zhi, P.; Yun, G.; Wang, Z.; Shi, P.; Guo, X.; Wu, J.; Ma, Z. A Novel Reliability Analysis Approach under Multiple Failure Modes Using an Adaptive MGRP Model. Appl. Sci. 2022, 12, 8961. [Google Scholar] [CrossRef]
  7. Tsompanakis, Y.; Papadrakakis, M. Large-scale reliability-based structural optimization. Struct. Multidiscip. Optim. 2004, 26, 429–440. [Google Scholar] [CrossRef]
  8. Chatterjee, T.; Chakraborty, S.; Chowdhury, R. A critical review of surrogate assisted robust design optimization. Arch. Comput. Methods Eng. 2019, 26, 245–274. [Google Scholar] [CrossRef]
  9. Li, J.; Xiu, D. Evaluation of failure probability via surrogate models. J. Comput. Phys. 2010, 229, 8966–8980. [Google Scholar] [CrossRef]
  10. Dai, H.; Zhao, W.; Wang, W.; Cao, Z. An improved radial basis function network for structural reliability analysis. J. Mech. Sci. Technol. 2011, 25, 2151–2159. [Google Scholar] [CrossRef]
  11. Dubourg, V.; Sudret, B.; Deheeger, F. Metamodel-based importance sampling for structural reliability analysis. Probabilistic Eng. Mech. 2013, 33, 47–57. [Google Scholar] [CrossRef] [Green Version]
  12. Chaudhuri, A.; Kramer, B.; Willcox, K.E. Information Reuse for Importance Sampling in Reliability-Based Design Optimization. Reliab. Eng. Syst. Saf. 2020, 201, 106853. [Google Scholar] [CrossRef]
  13. Foschi, R.; Li, H.; Zhang, J. Reliability and performance-based design: A computational approach and applications. Struct. Saf. 2002, 24, 205–218. [Google Scholar] [CrossRef]
  14. Qu, X.; Haftka, R.T. Reliability-based design optimization using probabilistic sufficiency factor. Struct. Multidiscip. Optim. 2004, 27, 314–325. [Google Scholar] [CrossRef]
  15. Moustapha, M.; Sudret, B. Surrogate-assisted reliability-based design optimization: A survey and a unified modular framework. Struct. Multidiscip. Optim. 2019, 60, 2157–2176. [Google Scholar] [CrossRef]
  16. Bichon, B.J. Efficient Surrogate Modeling for Reliability Analysis and Design. Ph.D. Thesis, Graduate School of Vanderbilt University, Nashville, TN, USA, 2010. [Google Scholar]
  17. Melchers, R.E. Importance sampling in structural systems. Struct. Saf. 1989, 6, 3–10. [Google Scholar] [CrossRef]
  18. Melchers, R.E. Search-based importance sampling. Struct. Saf. 1990, 9, 117–128. [Google Scholar] [CrossRef]
  19. West, N.; Swiler, L. Importance sampling: Promises and limitations. In Proceedings of the 51st AIAA/SAE/ASEE Joint Propulsion Conference, Orlando, FL, USA, 12–15 April 2010; pp. 1–14. [Google Scholar] [CrossRef] [Green Version]
  20. Yin, C.; Kareem, A. Computation of failure probability via hierarchical clustering. Struct. Saf. 2016, 61, 67–77. [Google Scholar] [CrossRef] [Green Version]
  21. Smarslok, B.P.; Haftka, R.T.; Carraro, L.; Ginsbourger, D. Improving accuracy of failure probability estimates with separable Monte Carlo. Int. J. Reliab. Saf. 2010, 4, 393. [Google Scholar] [CrossRef]
  22. Chaudhuri, A.; Haftka, R.T. Separable Monte Carlo combined with importance sampling for variance reduction. Int. J. Reliab. Saf. 2013, 7, 201. [Google Scholar] [CrossRef]
  23. Melchers, R.E. Structural Reliability Analysis and Prediction; John Wiley & Sons: Hoboken, NJ, USA, 2002. [Google Scholar]
  24. De Boer, P.T.; Kroese, D.P.; Rubinstein, R.Y. A Tutorial on the Cross-Entropy Method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  25. Kurtz, N.; Song, J. Cross-entropy-based adaptive importance sampling using Gaussian mixture. Struct. Saf. 2013, 42, 35–44. [Google Scholar] [CrossRef]
  26. Cao, Q.D.; Choe, Y. Cross-entropy based importance sampling for stochastic simulation models. Reliab. Eng. Syst. Saf. 2019, 191, 106526. [Google Scholar] [CrossRef] [Green Version]
  27. Geyer, S.; Papaioannou, I.; Straub, D. Cross entropy-based importance sampling using Gaussian densities revisited. Struct. Saf. 2019, 76, 15–27. [Google Scholar] [CrossRef] [Green Version]
  28. Ang, G.L.; Ang, A.H.; Tang, W.H. Optimal importance-sampling density estimator. J. Eng. Mech. 1992, 118, 1146–1163. [Google Scholar] [CrossRef]
  29. Zhang, P. Nonparametric Importance Sampling. J. Am. Stat. Assoc. 1996, 91, 1245–1253. [Google Scholar] [CrossRef]
  30. Au, S.; Beck, J. A new adaptive importance sampling scheme for reliability calculations. Struct. Saf. 1999, 21, 135–158. [Google Scholar] [CrossRef]
  31. Dai, H.; Zhang, H.; Rasmussen, K.J.R.; Wang, W. Wavelet density-based adaptive importance sampling method. Struct. Saf. 2015, 52, 161–169. [Google Scholar] [CrossRef]
  32. Botev, Z.I.; L’Ecuyer, P.; Tuffin, B. Markov chain importance sampling with applications to rare event probability estimation. Stat. Comput. 2013, 23, 271–285. [Google Scholar] [CrossRef]
  33. Dalbey, K.; Swiler, L. Gaussian process adaptive importance sampling. Int. J. Uncertain. Quantif. 2014, 4, 133–149. [Google Scholar] [CrossRef] [Green Version]
  34. Zhao, H.; Yue, Z.; Liu, Y.; Gao, Z.; Zhang, Y. An efficient reliability method combining adaptive importance sampling and Kriging metamodel. Appl. Math. Model. 2015, 39, 1853–1866. [Google Scholar] [CrossRef]
  35. Wang, Z.; Song, J. Cross-entropy-based adaptive importance sampling using von Mises-Fisher mixture for high dimensional reliability analysis. Struct. Saf. 2016, 59, 42–52. [Google Scholar] [CrossRef]
  36. Lee, G.; Kim, W.; Oh, H.; Youn, B.D.; Kim, N.H. Review of statistical model calibration and validation—from the perspective of uncertainty structures. Struct. Multidiscip. Optim. 2019, 60, 1619–1644. [Google Scholar] [CrossRef]
  37. Acar, E. A reliability index extrapolation method for separable limit states. Struct. Multidiscip. Optim. 2016, 53, 1099–1111. [Google Scholar] [CrossRef]
  38. Rubinstein, R.Y.; Kroese, D.P. Simulation and the Monte Carlo Method; John Wiley & Sons: Hoboken, NJ, USA, 2009; p. 345. [Google Scholar]
  39. Ramakrishnan, B.; Rao, S. A general loss function based optimization procedure for robust design. Eng. Optim. 1996, 25, 255–276. [Google Scholar] [CrossRef]
  40. Lee, D.; Rahman, S. Robust design optimization under dependent random variables by a generalized polynomial chaos expansion. Struct. Multidiscip. Optim. 2021, 63, 2425–2457. [Google Scholar] [CrossRef]
  41. Dymiotis, C.; Kappos, A.J.; Chryssanthopoulos, M.K. Seismic reliability of RC frames with uncertain drift and member capacity. J. Struct. Eng. 1999, 125, 1038–1047. [Google Scholar] [CrossRef]
  42. Li, D.; Dolezal, T.; Haimes, Y.Y. Capacity reliability of water distribution networks. Reliab. Eng. Syst. Saf. 1993, 42, 29–38. [Google Scholar] [CrossRef]
  43. Zhao, J.; Tang, J.; Wu, H.C. A generalized random variable approach for strain-based fatigue reliability analysis. J. Press. Vessel Technol. 2000, 122, 156–161. [Google Scholar] [CrossRef]
  44. Ramu, P.; Arul, S. Estimating probabilistic fatigue of Nitinol with scarce samples. Int. J. Fatigue 2016, 85, 31–39. [Google Scholar] [CrossRef]
  45. Yun, W.; Lu, Z.; Jiang, X. A modified importance sampling method for structural reliability and its global reliability sensitivity analysis. Struct. Multidiscip. Optim. 2018, 57, 1625–1641. [Google Scholar] [CrossRef]
  46. Cortés López, J.C.; Jornet Sanz, M. Improving kernel methods for density estimation in random differential equations problems. Math. Comput. Appl. 2020, 25, 33. [Google Scholar] [CrossRef]
  47. Coles, S. Classical extreme value theory and models. In An Introduction to Statistical Modeling of Extreme Values; Springer: Berlin/Heidelberg, Germany, 2001; pp. 45–73. [Google Scholar]
  48. Hill, B.M. A Simple General Approach to Inference About the Tail of a Distribution. Ann. Stat. 1975, 3, 1163–1174. [Google Scholar] [CrossRef]
  49. Aban, I.B.; Meerschaert, M.M. Shifted hill’s estimator for heavy tails. Commun. Stat. Part B Simul. Comput. 2001, 30, 949–962. [Google Scholar] [CrossRef]
  50. Nguyen, T.; Samorodnitsky, G. Tail inference: Where does the tail begin? Extremes 2012, 15, 437–461. [Google Scholar] [CrossRef] [Green Version]
  51. Yao, W.; Tang, G.; Wang, N.; Chen, X. An improved reliability analysis approach based on combined FORM and Beta-spherical importance sampling in critical region. Struct. Multidiscip. Optim. 2019, 60, 35–58. [Google Scholar] [CrossRef]
  52. Acar, E. Reliability prediction through guided tail modeling using support vector machines. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2013, 227, 2780–2794. [Google Scholar] [CrossRef]
  53. Wu, Y.T.; Shin, Y.; Sues, R.H.; Cesare, M.A. Safety-factor based approach for probability-based design optimization. In Proceedings of the 19th AIAA Applied Aerodynamics Conference, Baltimore, MD, USA, 23–25 September 1991. [Google Scholar] [CrossRef]
  54. Ravishankar, B.; Smarslok, B.; Haftka, R.; Sankar, B. Separable sampling of the limit state for accurate Monte Carlo Simulation. In Proceedings of the 50th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference 17th AIAA/ASME/AHS Adaptive Structures Conference, Palm Springs, CA, USA, 4–7 May 2009; p. 2266. [Google Scholar]
  55. Chateauneuf, A.; Aoues, Y. Advances in solution methods for reliability-based design optimization. In Structural Design Optimization Considering Uncertainties; CRC Press: Boca Raton, FL, USA, 2008; pp. 217–246. [Google Scholar] [CrossRef]
  56. Bennett, J.; Botkin, M. The Optimum Shape; Springer: New York, NY, USA, 1986. [Google Scholar] [CrossRef]
  57. Acar, E. Guided tail modelling for efficient and accurate reliability estimation of highly safe mechanical systems. Proc. Inst. Mech. Eng. Part C J. Mech. Eng. Sci. 2011, 225, 1237–1251. [Google Scholar] [CrossRef]
  58. Rahman, S.; Wei, D. Reliability-based design optimization by a univariate decomposition method. In Proceedings of the 13th AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference 2010, Fort Worth, TX, USA, 13–15 September 2010; pp. 1–14. [Google Scholar] [CrossRef]
  59. Picheny, V.; Kim, N.H.; Haftka, R.; Queipo, N. Conservative predictions using surrogate modeling. In Proceedings of the 49th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference, Schaumburg, IL, USA, 7–10 April 2008; p. 1716. [Google Scholar]
  60. Lindemann, J.; Persson, K. CALFEM, A finite element toolbox to MATLAB. 1999. Available online: http://www-amna.math.uni-wuppertal.de/~ehrhardt/TUB/NumPar/calfem/calfem.pdf (accessed on 14 September 2022).
  61. Deb, K.; Gupta, S.; Daum, D.; Branke, J.; Mall, A.K.; Padmanabhan, D. Reliability-based optimization using evolutionary algorithms. IEEE Trans. Evol. Comput. 2009, 13, 1054–1074. [Google Scholar] [CrossRef] [Green Version]
  62. Stein, M. Large sample properties of simulations using Latin hypercube sampling. Technometrics 1987, 29, 143–151. [Google Scholar] [CrossRef]
  63. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  64. Botev, Z.I.; Grotowski, J.F.; Kroese, D.P. Kernel density estimation via diffusion. Ann. Stat. 2010, 38, 2916–2957. [Google Scholar] [CrossRef] [Green Version]
  65. Botev, Z.I. Adaptive Kernel Density Estimation in One-Dimension—MATLAB Central File Exchange. 2016. Available online: https://www.mathworks.com/matlabcentral/fileexchange/58309-adaptive-kernel-density-estimation-in-one-dimension (accessed on 14 September 2022).
  66. Hong, H.P.; Lind, N.C. Approximate reliability analysis using normal polynomial and simulation results. Struct. Saf. 1996, 18, 329–339. [Google Scholar] [CrossRef]
Figure 1. Basic CR problem: a schematic of importance sampling.
Figure 1. Basic CR problem: a schematic of importance sampling.
Mca 27 00099 g001
Figure 2. Probabilistic view of MPP for a separable limit state scenario.
Figure 2. Probabilistic view of MPP for a separable limit state scenario.
Mca 27 00099 g002
Figure 3. Procedure for identifying mean of Gaussian ISD ( μ h ) . (a) Scarce samples of R and C. (b) Finding point A using response sample r i as per Step 4 in Algorithm 1. (c) Finding point B using capacity sample c i as per Step 7 in Algorithm 1. (d) μ h identified between points A and B as per Step 9 in Algorithm 1.
Figure 3. Procedure for identifying mean of Gaussian ISD ( μ h ) . (a) Scarce samples of R and C. (b) Finding point A using response sample r i as per Step 4 in Algorithm 1. (c) Finding point B using capacity sample c i as per Step 7 in Algorithm 1. (d) μ h identified between points A and B as per Step 9 in Algorithm 1.
Mca 27 00099 g003
Figure 4. Schematic representation of bootstrapping.
Figure 4. Schematic representation of bootstrapping.
Mca 27 00099 g004
Figure 5. Flowchart representing failure probability estimation using proposed IS approach.
Figure 5. Flowchart representing failure probability estimation using proposed IS approach.
Mca 27 00099 g005
Figure 6. Schematic of roof truss.
Figure 6. Schematic of roof truss.
Mca 27 00099 g006
Figure 7. Propped cantilever beam with triangular distributed loading.
Figure 7. Propped cantilever beam with triangular distributed loading.
Mca 27 00099 g007
Figure 8. Cantilever beam under horizontal and vertical loads.
Figure 8. Cantilever beam under horizontal and vertical loads.
Mca 27 00099 g008
Figure 9. Bracket structure subjected to a tip load.
Figure 9. Bracket structure subjected to a tip load.
Mca 27 00099 g009
Figure 10. Design variables used to modify the shape of the torque arm [59].
Figure 10. Design variables used to modify the shape of the torque arm [59].
Mca 27 00099 g010
Figure 11. Loading and boundary conditions with base design parameter values (in cm) [59].
Figure 11. Loading and boundary conditions with base design parameter values (in cm) [59].
Mca 27 00099 g011
Figure 12. Stress (von Mises stress in MPa) contour of optimum design with mean design parameters ( d * ) and loading condition of F x = 2789 N and F y = 5066 N.
Figure 12. Stress (von Mises stress in MPa) contour of optimum design with mean design parameters ( d * ) and loading condition of F x = 2789 N and F y = 5066 N.
Mca 27 00099 g012
Figure 13. Pareto optimal front corresponding to different values of reliability index estimated through polynomial response surface (PRS).
Figure 13. Pareto optimal front corresponding to different values of reliability index estimated through polynomial response surface (PRS).
Mca 27 00099 g013
Table 1. GEV distributions parameters of R and C for nine study cases.
Table 1. GEV distributions parameters of R and C for nine study cases.
Response Tail ξ r , θ r , σ r Capacity Tail ξ c , θ c , σ c
Heavy( 1.8 , 1 , 33.5 )
Heavy( 0.2 , 1 , 0 )Medium( 1 , 1 , 26.5 )
Light( 0.52 , 1 , 26.2 )
Heavy( 1.8 , 1 , 30 )
Medium( 0 , 1 , 0 )Medium( 1 , 1 , 10.7 )
Light( 0.52 , 1 , 9.5 )
Heavy( 1.8 , 1 , 30 )
Light( 0.12 , 1 , 0 )Medium( 1 , 1 , 9.4 )
Light( 0.52 , 1 , 6.4 )
Table 2. Both R and C unknown case: percentiles of M M β I S / β a .
Table 2. Both R and C unknown case: percentiles of M M β I S / β a .
Heavy CMedium CLight C
PercentileOriginal SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Heavy R 25 th 0.961.10 (0.26)1.211.42 (0.29)1.792.08 (0.61)
50 th 1.061.17 (0.29)1.501.54 (0.32)2.132.46 (0.89)
75 th 1.221.36 (0.31)1.761.77 (0.33)2.593.08 (1.31)
Medium R 25 th 0.931.03 (0.20)0.850.95 (0.18)1.011.09 (0.21)
50 th 1.021.09 (0.21)0.971.01 (0.19)1.131.20 (0.25)
75 th 1.181.25 (0.23)1.081.14 (0.22)1.301.40 (0.32)
Light R 25 th 0.911.04 (0.23)0.780.85 (0.14)0.850.91 (0.12)
50 th 1.011.10 (0.25)0.880.91 (0.14)0.940.98 (0.13)
75 th 1.181.28 (0.26)0.991.04 (0.17)1.031.09 (0.18)
Table 3. R unknown case: failure probability estimated as per Equation (6); (PDF approximation of R): percentiles of M M β I S / β a .
Table 3. R unknown case: failure probability estimated as per Equation (6); (PDF approximation of R): percentiles of M M β I S / β a .
Heavy CMedium CLight C
PercentileOriginal SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Heavy R 25 th 1.021.02 ( 9 × 10 3 )1.671.71 (0.15)3.163.63 (0.66)
50 th 1.031.03 ( 7 × 10 3 )1.781.72 (0.15)3.643.67 (0.67)
75 th 1.041.04 ( 5 × 10 3 )1.811.79 (0.06)4.004.09 (0.30)
Medium R 25 th 0.990.99 ( 2 × 10 3 )1.021.03 (0.06)1.251.32 (0.20)
50 th 1.001.00 ( 2 × 10 3 )1.051.04 (0.05)1.371.34 (0.20)
75 th 1.001.00 ( 2 × 10 3 )1.081.07 (0.03)1.451.51 (0.14)
Light R 25 th 0.990.99 ( 2 × 10 3 )0.990.99 (0.02)0.981.03 (0.10)
50 th 1.001.00 ( 2 × 10 3 )1.001.00 (0.02)1.071.06 (0.10)
75 th 1.001.00 ( 2 × 10 3 )1.021.01 (0.02)1.121.13 (0.07)
Table 4. C unknown case: failure probability estimated as per Equation (6); (CDF approximation of C): percentiles of M M β I S / β a .
Table 4. C unknown case: failure probability estimated as per Equation (6); (CDF approximation of C): percentiles of M M β I S / β a .
Heavy CMedium CLight C
PercentileOriginal SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Heavy R 25 th 0.960.98 (0.10)0.990.99 ( 6 × 10 3 )1.001.00 ( 2 × 10 3 )
50 th 1.031.00 (0.08)1.001.00 ( 3 × 10 3 )1.001.00 ( 2 × 10 3 )
75 th 1.061.04 (0.05)1.001.00 ( 2 × 10 3 )1.001.00 ( 2 × 10 3 )
Medium R 25 th 0.931.13 (0.31)0.870.92 (0.12)0.960.96 (0.05)
50 th 1.121.24 (0.35)0.960.96 (0.11)0.990.98 (0.04)
75 th 1.421.52 (0.33)1.031.02 (0.07)1.001.00 (0.02)
Light R 25 th 0.911.05 (0.23)0.820.90 (0.13)0.870.89 (0.09)
50 th 1.031.12 (0.25)0.930.95 (0.14)0.940.94 (0.09)
75 th 1.201.29 (0.29)1.031.06 (0.14)1.000.99 (0.08)
Table 5. R unknown case: failure probability estimated as per Equation (7); (CDF approximation of R): percentiles of M M β I S / β a .
Table 5. R unknown case: failure probability estimated as per Equation (7); (CDF approximation of R): percentiles of M M β I S / β a .
Heavy CMedium CLight C
PercentileOriginal SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Original SampleBootstrap
Mean (Std)
Heavy R 25 th 0.950.96 (0.09)0.951.14 (0.29)0.921.13 (0.31)
50 th 1.020.99 (0.07)1.131.22 (0.31)1.111.23 (0.35)
75 th 1.031.03 (0.02)1.431.42 (0.30)1.381.51 (0.41)
Medium R 25 th 0.990.99 (0.02)0.840.89 (0.12)0.770.89 (0.18)
50 th 0.990.99 (0.01)0.940.94 (0.11)0.930.96 (0.20)
75 th 1.001.00 ( 2 × 10 3 )1.021.01 (0.09)1.081.11 (0.21)
Light R 25 th 0.990.99 ( 2 × 10 3 )0.910.93 (0.07)0.790.84 (0.12)
50 th 1.001.00 ( 1 × 10 3 )0.970.96 (0.06)0.870.90 (0.12)
75 th 1.001.00 ( 1 × 10 3 )1.000.99 (0.04)0.990.98 (0.12)
Table 6. Heavy R and light C: Percentiles of M M β I S / β t from tail estimation-based improved formulation.
Table 6. Heavy R and light C: Percentiles of M M β I S / β t from tail estimation-based improved formulation.
PercentileOriginal SampleBootstrap
Mean (Std)
25th0.931.19 (0.39)
50th1.121.32 (0.53)
75th1.391.70 (1.02)
Table 7. Concave limit state-1: percentiles of M M β I S / β t for β t = 2.39 .
Table 7. Concave limit state-1: percentiles of M M β I S / β t for β t = 2.39 .
PercentileOriginal SampleBootstrap
Mean (Std)
25th0.840.96 (0.20)
50th0.961.03 (0.20)
75th1.121.20 (0.22)
Note: β t is computed through MCS with 108 samples.
Table 8. Concave limit state-2: percentiles of M M β I S / β t for β t = 2.82 .
Table 8. Concave limit state-2: percentiles of M M β I S / β t for β t = 2.82 .
PercentileOriginal SampleBootstrap
Mean (Std)
25th0.820.89 (0.16)
50th0.900.95 (0.16)
75th1.021.08 (0.19)
Note: β t is computed through MCS with 108 samples.
Table 9. Mean and SD of random variables for roof truss example.
Table 9. Mean and SD of random variables for roof truss example.
Random VariableMean (SD)
q (N/m)20,000 (1600)
l (m)12 (0.24)
A s ( m 2 ) 9.82 × 10 4 ( 5.89 × 10 5 )
A c ( m 2 ) 0.04 (0.008)
E s ( N / m 2 ) 1.2 × 10 11 ( 8.4 × 10 9 )
E c ( N / m 2 ) 3 × 10 10 ( 2.4 × 10 9 )
Table 10. Roof truss example: percentiles of M M β I S / β t for β t = 3.4 .
Table 10. Roof truss example: percentiles of M M β I S / β t for β t = 3.4 .
PercentileOriginal SampleBootstrap
Mean (Std)
25th0.840.94 (0.17)
50th0.941.02 (0.20)
75th1.071.18 (0.29)
Note: β t is computed through MCS with 108 samples.
Table 11. Mean and SD of random variables for propped cantilever beam example.
Table 11. Mean and SD of random variables for propped cantilever beam example.
Random VariableMean (SD)
q 0 (kN/m)20 (2)
L (m)6 (0.3)
E (GPa)210 (10)
d (cm)25 (0.5)
b f (cm)25 (0.5)
t w (cm)2 (0.2)
t f (cm)2 (0.2)
Note: All random variables follow normal distribution.
Table 12. Propped cantilever beam: percentiles of M M β I S / β t for different ν c r i t and β t .
Table 12. Propped cantilever beam: percentiles of M M β I S / β t for different ν c r i t and β t .
ν crit = 4.0 mm , β t = 2.98 ν crit = 4.5 mm , β t = 3.50 ν crit = 5.0 mm , β t = 3.97
PercentileOriginalBootstrap
Mean (Std)
OriginalBootstrap
Mean (Std)
OriginalBootstrap
Mean (Std)
25th0.841.03 (0.30)0.820.99 (0.27)0.810.99 (0.30)
50th0.991.13 (0.36)0.991.09 (0.32)0.961.12 (0.39)
75th1.291.31 (0.42)1.241.29 (0.44)1.261.35 (0.52)
Note: β t values are computed through MCS with 108 samples.
Table 13. Random variables for cantilever beam example.
Table 13. Random variables for cantilever beam example.
Random VariableMean (SD)
X ( l b ) 500 (100)
Y ( l b ) 10 3 (100)
σ y ( p s i ) 4 × 10 4 ( 2 × 10 3 )
E ( p s i ) 2.9 × 10 7 ( 1.45 × 10 6 )
Note: All random variables follow normal distribution.
Table 14. RBDO results of cantilever beam example.
Table 14. RBDO results of cantilever beam example.
Surrogate Model
for β Constraints
Reliability
Estimation
Optima ( d * ) Objective
Function Value
β ^ at d * β MCS at d *
w (in) t (in) A (in 2 ) g 1 g 2 g 1 g 2
WASIS2.593.749.693.003.643.253.69
MCS2.593.669.503.003.442.953.39
Table 15. Random variables for bracket structure example.
Table 15. Random variables for bracket structure example.
TypeVariableDistributionMeanC.o.V
RandomP (kN)Gumbel10015%
E (GPa)Gumbel2008%
f y (MPa)Lognormal2258%
ρ   ( kg · m 3 ) Weibull786010%
L (m)Gaussian55%
Design w A B (mm)Gaussian μ w A B 5%
w C D (mm)Gaussian μ w C D 5%
t (mm)Gaussian μ t 5%
Table 16. RBDO results of bracket structure example.
Table 16. RBDO results of bracket structure example.
Surrogate Model
for β Constraints
Reliability
Estimation
Optima ( d * ) Objective
Function Value
β ^ at d * β MCS at d *
w AB ( mm ) w CD ( mm ) t ( mm ) Weight (kg) g 1 g 2 g 1 g 2
WASIS588930015762.002.002.592.87
MCS627730014742.002.002.023.55
Table 17. Mean and SD of random variables in torque arm example.
Table 17. Mean and SD of random variables in torque arm example.
Random VariableDistribution TypeMean; SD
F x (N)Normal 2789 ; 278.9
F y (N)Normal 5066 ; 506.6
σ a l l (MPa)Lognormal 800 ; 80
Table 18. Design variable bounds and optimum design of torque arm (in cm).
Table 18. Design variable bounds and optimum design of torque arm (in cm).
DV d L d U d *
(Optimum)
d 1 1.803.202.15
d 2 1.251.601.28
d 3 1.204.601.59
d 4 −0.100.40−0.09
d 5 −0.300.300.30
d 6 −0.900.800.30
d 7 0.401.800.54
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Pannerselvam, K.; Yadav, D.; Ramu, P. Scarce Sample-Based Reliability Estimation and Optimization Using Importance Sampling. Math. Comput. Appl. 2022, 27, 99. https://doi.org/10.3390/mca27060099

AMA Style

Pannerselvam K, Yadav D, Ramu P. Scarce Sample-Based Reliability Estimation and Optimization Using Importance Sampling. Mathematical and Computational Applications. 2022; 27(6):99. https://doi.org/10.3390/mca27060099

Chicago/Turabian Style

Pannerselvam, Kiran, Deepanshu Yadav, and Palaniappan Ramu. 2022. "Scarce Sample-Based Reliability Estimation and Optimization Using Importance Sampling" Mathematical and Computational Applications 27, no. 6: 99. https://doi.org/10.3390/mca27060099

APA Style

Pannerselvam, K., Yadav, D., & Ramu, P. (2022). Scarce Sample-Based Reliability Estimation and Optimization Using Importance Sampling. Mathematical and Computational Applications, 27(6), 99. https://doi.org/10.3390/mca27060099

Article Metrics

Back to TopTop