Abstract

A new reliability-based design optimization (RBDO) method based on support vector machines (SVM) and the Most Probable Point (MPP) is proposed in this work. SVM is used to create a surrogate model of the limit-state function at the MPP with the gradient information in the reliability analysis. This guarantees that the surrogate model not only passes through the MPP but also is tangent to the limit-state function at the MPP. Then, importance sampling (IS) is used to calculate the probability of failure based on the surrogate model. This treatment significantly improves the accuracy of reliability analysis. For RBDO, the Sequential Optimization and Reliability Assessment (SORA) is employed as well, which decouples deterministic optimization from the reliability analysis. The improved SVM-based reliability analysis is used to amend the error from linear approximation for limit-state function in SORA. A mathematical example and a simplified aircraft wing design demonstrate that the improved SVM-based reliability analysis is more accurate than FORM and needs less training points than the Monte Carlo simulation and that the proposed optimization strategy is efficient.

1. Introduction

There are many uncertainties encountered in both of the aircraft manufacturing process and its subsequent flight operation. The physical properties of materials are uncertain. Manufacturing errors produce the aerodynamic shape and structural dimensions different from the original design [1]. Furthermore, the load on the aircraft is not constant during operation. Fuel is consumed continually during cruise; thus, fight parameters keep changing. After an aircraft has been produced, some parameters sometimes need to be adjusted for a new type. If these uncertainties are considered in the conceptual design, the aircraft performance will be more reliable than deterministic design [2], and both of the risk and cost in the design will be reduced.

Reliability analysis is the key part of reliability-based design optimization (RBDO). Reliability is the probability of success. In physics-based reliability, the status of success is specified by limit-state functions, which are derived from physics principles [3]. Let the limit-state function be , where is a vector of random variables with length . If the event of success is specified by , then the failure event is . Consequently, the probability of failure is computed bywhere is the joint probability density function (PDF) of .

Since it is difficult to analytically evaluate the probability integral, many approximation methods have been developed. Among them, the First Order Reliability Method (FORM) [46] is commonly used. The FORM linearizes the limit-state function at the Most Probable Point (MPP), which in the standard normal space has the highest probability of producing the value of limit-state function. Then, (1) becomes a linear combination of normal variables. Since only the first order derivatives of and some basic information are needed, the FORM is efficient. Its accuracy, however, may not be satisfactory for highly nonlinear limit-state functions.

Several methods are available to improve the accuracy of the FORM. One of the methods is the Second Order Reliability Method (SORM) [7]. The SORM provides a second order approximation to at the MPP. As a result, it is generally more accurate than the FORM. The MPP-based importance sampling [8, 9] is another alternative method. Random samples are drawn from distributions whose center is shifted to the MPP. A relatively small sample size can then produce a good estimation of the probability of failure. The other strategy is to approximate the safety-failure boundary at the MPP with higher accuracy. In the point-fitting method [10], a piecewise paraboloid surface is built with the fitting points selected from each side of MPP along both forward and backward directions of each random variable. Similarly, response surface modeling has also been used to create a surrogate model for the limit-state function at the MPP [11, 12]. A surrogate model can be created using artificial neural network [13] as well.

Recently, another strategy, the support vector machine (SVM) [14] method, has been introduced in reliability analysis. SVM is a statistical classification method. As indicated in [15], reliability analysis can be viewed as a classification problem where SVM is applicable. Examples of using SVM include the fast Monte Carlo simulation (MCS) [16], the limit-state function identification for discontinuous responses and disjoint failure domains [17], SVM-based MCS [18], and virtual SVM for high-dimensional problems [19]. It is worthwhile to further study SVM for reliability analysis and RBDO.

The conventional approach for solving a reliability-based design optimization problem is to employ a double-loop strategy in which the analysis and the synthesis are nested in such a way that the synthesis loop performs the reliability analysis iteratively for meeting the probabilistic constraints. As the double-loop strategy may be computationally infeasible, various single loop strategies have been studied to improve its efficiency. The method “approximately equivalent deterministic constraints” creates a link between a probabilistic design and a safety-factor based design [20]. The reliability constraints are formulated as deterministic constraints that approximate the condition of the MPP for reliability analysis [21]. A single loop method, Sequential Optimization and Reliability Assessment (SORA), is a very efficient method for RBDO [22]. In this method, optimization and reliability analysis are decoupled from each other; no reliability analysis is required within optimization and the reliability analysis is only conducted after the optimization. Hence, the design is quickly improved from cycle to cycle and the computational efficiency is improved significantly. However, because of FORM employed in SORA based on the limit-state function linearization, its precision may not be high enough for the highly nonlinear problem.

In this work, the accuracy of SVM-based reliability analysis was improved firstly. In addition to the training points around the MPP, the gradient of a limit-state function at the MPP was included in approximating the limit-state function, to guarantee that the surrogate model not only passes through the MPP but also is tangent to the limit-state function at the MPP. And importance sampling is used to estimate the probability of failure based on the surrogate model. Then, the improved SVM-based reliability analysis was integrated into SORA for wing optimization. Results of the two examples showed that this strategy is more accurate than before with a moderately increased computational cost.

2. FORM, IS, and SVM

In this section, the three methods used in this work are reviewed. The methods include the First Order Reliability Method (FORM), importance sampling (IS), and support vector machines (SVM).

2.1. First Order Reliability Method (FORM)

In this work, we assume all random variables in are independent. FORM involves the following three steps.

(1) Transformation of Random Variables into Standard Normal Variables. The original random variables (in the -space) are transformed into random variables (in the -space) whose elements follow a standard normal distribution. The transformation is given by [23]where is the cumulative distribution function (CDF) of and is the inverse CDF of a standard normal variable.

(2) Search the Most Probable Point (MPP). The MPP is the point at the limit-state , and, at the MPP, the PDF of is at its maximum. Maximizing the joint PDF of and noting that a contour of the PDF of is a concentric hypersphere, we obtain the MPP by solvingwhere stands for the norm (length) of a vector.

Geometrically, the MPP is the shortest distance point from surface to the origin of the -space. The minimum distance is called the reliability index. Then, is approximated by the first Taylor expansion series at aswhere is the gradient of at the MPP and is given by

(c) Calculate the Probability of Failure. As shown in (4), is now a linear combination of normal variables. As a result, can be easily computed bywhere is the CDF of a standard normal variable.

2.2. Importance Sampling (IS)

The FORM is commonly used due to its good balance between accuracy and efficiency. For highly nonlinear limit-state functions, however, the accuracy may not be good enough. IS can be used to improve accuracy.

The probability of failure in (1) can be rewritten aswhere is an indicator function and is defined as

is therefore the expectation of ; namely,where denotes an expectation. With the direct Monte Carlo simulation (MCS), can be estimated by averaging : where are the samples drawn from the joint probability density and is the number of failures.

A large sample size is required when is small because the chance of getting samples in failure region is small. IS draws samples from a new set of distributions such that more samples will be in the failure region. One strategy is to shift the mean values of the random variables to the MPP. As shown in Figure 1, all the samples (the lower cloud) generated from the original distributions of and are in the safe region. They do not contribute to the probability estimation. After the mean values are shifted to the MPP, sufficient samples are now in the failure region.

Let the importance sampling density be , which is obtained by shifting the means of to the MPP. Then, is estimated by

As shown in Figure 1, IS obtains a significant number of samples in the failure region with the same sample size. Because the indicator function of a failure point is 1.0, there are a significant number of nonzero terms in the summation in (11). However, it is very difficult for the direct MCS to generate failure samples and the terms in (10) are almost zero. As a result, no estimation of can be produced. Then, more samples are needed. Therefore, IS more efficient than the direct MCS.

2.3. Support Vector Machine (SVM)

The SVM method has been primarily used for solving both linear and nonlinear classification problems in statistics. Because in reliability analysis there are two states, either safety or failure, SVM is also applicable for reliability analysis. Recently, several studies on SVM in reliability analysis have been reported [1618]. The basics of SVM are briefly reviewed in this subsection.

For a linear limit-state function, SVM classifies training data (samples) into two classes, typically represented by “+1” or “−1,” depending on whether the sample belongs to safety or failure region. Two parallel hyperplanes are obtained by maximizing their distance, as shown in Figure 2. There are no samples between the hyperplanes. They therefore separate the samples into safety and failure groups. The points (the circles in the figure) passed by these hyperplanes are called support vectors (SV). The center hyperplane, which lies in half the distance from both hyperplanes, is actually used as the safety-failure boundary [17, 24].

This center hyperplane is the surrogate model for at the limit-state and is represented by where is a weight vector, is the dot (inner) product of and , and is the bias. And one outer hyperplane followsand the other hyperplane satisfiesThen, the distance between the two hyperplanes is . As mentioned previously, the distance must be maximized. Because we can minimize . To make sure there are no training points between the two hyperplanes, we need to add the following constraint to the optimization model [24]:where is the vector of training points and is the number of training points. is +1 if is in the safety region and −1 otherwise.

The optimization model is then given by

The optimization is a quadratic programming problem, and it is usually converted into a Lagrangian dual problem [25, 26]. The dual optimization model is where is the Lagrangian multiplier.

Once is obtained, the weight vector in (12) can be computed by

According to Karush-Kuhn-Tucker conditions, only the support vectors have non-zero Lagrangian multipliers [27], that is, only the SV appear in the optimum result [25]. Therefore, the bias in (12) can be determined by SV with :For nonlinear limit-state functions, hypersurfaces must be generated. To do so, we can simply replace the dot product in (17) with the following kernel function:where is the feature function of , which is used to map the original variable into a higher-dimensional space called the feature space. In the -dimensional feature space, is represented by . It is noted that the kernel function is the dot product of the two feature functions. And the most popular kernel functions are Gaussian kernel and polynomial kernel, which are defined asrespectively. Although the feature function appears in (20), it will not be used directly.

Then, the simplified (surrogate) model for the limit-state function is

3. Improved SVM Method for Reliability Analysis

As discussed above, MPP-based importance sampling (IS) is more accurate than the FORM. But it may still require a significant number of samples for high accuracy. To further improve the accuracy with balanced efficiency, we propose to integrate the FORM, SVM, and IS. To use the full information of the MPP, we also use the gradient of at the MPP in the -space. With the additional information of the gradient, the SVM surrogate model will be more accurate. We call this method SVM-MPP-G method. Its procedure is as follows:(1)FORM: search the MPP ;(2)IS: generate samples at the MPP;(3)SVM: create a surrogate model for the limit-state function based on the IS samples and the gradient of at the MPP in the -space;(4)IS or MCS: perform IS or MCS around the MPP to evaluate the probability of failure with .

The flowchart of the method is shown in Figure 3. The details of the steps are provided below.

3.1. Step 1: MPP Search

The MPP is identified with the model in (3). Nonlinear optimization algorithms can be used to solve the model. A specialized search algorithm, however, may be more efficient than optimization. In this work, we use the robust MPP search algorithm, the improved HLRF algorithm, denoted by iHRLF. The algorithm is proposed by Zhang and der Kiureghian [28]. iHRLF is efficient; it is also globally convergent because it can converge to a local MPP from any starting point. The algorithm is summarized below.

In iteration , the MPP is updated bywhere the search direction is given bywhere .

The step size is determined by minimizing a merit functionin which the constant should satisfy

Practically, searching for the step size terminates once the merit function is sufficiently reduced. The following rule is employed to find :In this paper, and are used.

Equation (28) indicates that is the first integer such that the merit function is less than the previous merit function .

If the distance between two consecutive points and is less than a small value ; namely, , the procedure stops. The result is the MPP in the -space. The MPP in the -space can be obtained by the inverse transformation based on (2)

3.2. Step 2: Sampling around the MPP

We are interested in creating the limit-state boundary accurately. Since the MPP has the highest probability density at the limit-state, we particularly desire higher accuracy around the MPP. To this end, we draw samples around the MPP. To further improve the sampling efficiency, we also use the gradient information of the limit-state function at the MPP so that the samples will be closer to the limit-state boundary. Since the gradient is a byproduct of the MPP search, no further computations are needed for its use.

To easily draw samples around the MPP, we perform sampling in the -space. The sampling distributions are standard normal ones, but their means are shifted to the MPP. If the samples are drawn from the sampling distributions directly, they will form a symmetric hypersphere around the MPP. Some of them may be far away from the limit-state boundary. To make them close to the boundary, we shrink the sampling distributions in the gradient direction. To make the samples closer to the MPP, we also truncate the distributions. As shown in Figure 4, more samples are in the vicinity of the MPP with the truncation and shrinking.

To implement the idea, we first rotate coordinates into a new set of coordinates , so that the last coordinate ( in the figure) is coincident with the vector from the origin to the MPP. In another word, is colinear with the gradient of at the MPP. In the new linear -space, the coordinates of the MPP become . This treatment makes the shrinking easier.

We actually draw samples in the rotated -space. The distributions are standard normal distributions for ; namely, . The distribution of is also a normal distribution with its mean being ; namely, . To make the samples close to the limit-state boundary, we truncate all the distributions. Suppose the PDF of is and lies within the interval , where is the distance between the mean value and one of the truncated tails; the truncated PDF [29] is then given by where is the CDF of .

For , we use the same distance to shrink the distribution. In the gradient direction in the -axis, we truncate the distribution of with a smaller distance ; namely, . Generally, the distances and could be taken.

Because of its good efficiency, we also use Latin Hypercube Sampling (LHS) [30] instead of the direct Monte Carlo sampling. In general, 10~25 times the number of random variables could be taken as the number of sampling points. After the samples of are generated, they need to be transformed into the -space. If each sample point is represented by a row vector, an orthogonal transformation is implemented by where is the rotation matrix. It could be computed with the gradient information of the limit-state function at the MPP, which has been obtained during the MPP searching. The matrix is derived from another matrix with the Gram-Schmidt orthogonalization. is given bywhere is an identity matrix, and the last row of is a unit vector defined as follows:

Then, can be generated from with the following Gram-Schmidt orthogonalization procedure: let the rows of and be and , respectively. Consider

Therefore, we could obtain the training points in the -space with (31).

3.3. Step 3: Surrogate Model Construction

With the training points obtained above, we now construct the SVM surrogate model for . is not approximated for actual value of . Just the limit-state functions is similar to in the vicinity of the MPP. Different from the traditional SVM methods, this method incorporates the gradient information at the MPP in the optimization model. This allows us to use the full information obtained during the MPP search.

For nonlinear limit-state functions, as discussed in Section 2.3, a feature function ( in the following discussions) is used to accommodate the nonlinearity. As shown in (23), the approximated limit-state function is given by

The derivative of with respect to is [31] where is the gradient of limit-state function at the MPP and is the proportionality factor. , and .

We add (36) into the original optimization problem in (16) as additional linear constraints. The new optimization model then becomes

The Lagrangian function of the dual problem changes toand . should be nonnegative. is +1 if is positive and −1 otherwise. Given the Karush-Kuhn-Tucker (KKT) complementary conditions [27] for maximizing , the following equations hold:

Plugging (40) into (38) yields

Because

then

And plugging (40) into (42) yields

With (45) andthe additional constraints become

Then, the dual problem is

Because should pass through the MPP , we can use to find the bias . Using and (40), we obtain

The approximated limit-state function is then given by

3.4. Step 4: Probability Evaluation

Once is available, a reliability analysis method can be used to estimate the probability of failure . Since is cheaper to compute, we can afford a large number of function calls to maintain higher accuracy. We can therefore use importance sampling (IS) or the direct Monte Carlo simulation. If IS is selected, the probability of failure is computed with (11).

4. Integration of SORA and SVM Method for Reliability Analysis

A typical model of a probabilistic design is given bywhere is an objective function, is the vector of deterministic design variables, is the vector of random design variables, is the vector of random design parameters, , , are constraint functions, are the desired probabilities of constraint satisfaction, and is the number of constraints. The design variables are and the means () of the random design variables .

4.1. Sequential Optimization and Reliability Assessment (SORA)

The conventional reliability-based optimization employs a double-loop strategy. The outer loop is to solve the optimization problem, and the inner loop is for reliability analysis. The total number of function evaluations will be huge. However, the SORA method decouples the reliability analysis and optimization completely.

In SORA, an equivalent model to the probabilistic constraints in (53) with performance measure approach is given by where is the -percentile of ; namely,

It indicates that the probability of greater than or equal to the -percentile is exactly equal to the desired reliability . Using the inverse MPP search algorithm, the optimum solution MPP can be identified and the desired percentile is evaluated by

Then, the design model (1) is rewritten as

The concept of SORA method is shown in Figure 5. The optimization in the first cycle is just a deterministic optimization. Then, the reliability analysis is implemented for the deterministic optimum solution to locate the MPP that corresponds to the desired level. In Cycle 2, the constraints are modified to shift the MPP onto the deterministic boundary to help insure the feasibility of the probabilistic constraint. The new constraints arewhere is the shifting vector as . The superscript 1 means the first cycle.

The reliabilities of those violated probabilistic constraints will improve remarkably using this MPP shifting strategy. After the optimization in Cycle 2, the reliability assessment of Cycle 2 is conducted to find the updated MPPs and to check the design feasibility. If some probabilistic constraints are still not satisfied, the procedure is repeated cycle by cycle until the objective converges and the reliability requirement is achieved when all the shifting distances become zero. Since the SORA method requires much less optimization iterations and reliability assessments to converge, the overall efficiency is high.

4.2. Integration of SORA and Improved Reliability Analysis

FORM is employed in SORA for reliability analysis. To improve the precision of reliability-based optimization for the highly nonlinear problem, we integrated the improved SVM-based reliability analysis into SORA. The flowchart of this optimization strategy is provided in Figure 6. Firstly, the SORA is operated to find the optimal solution, which satisfies the constraint requirement with the FORM for reliability analysis. Then, the optimal solution is verified by the improved SVM-based reliability analysis. If the approximation for the constraint in SORA could satisfy the accuracy requirement, the procedure stops. Otherwise, the SORA performs again to search a better result in feasible region after reliability index modification, until all the constraints satisfy the reliability requirement and the accuracy is satisfied. The details of the procedure are provided below.

4.2.1. SORA

SORA is the first step in the optimization procedure. The optimal solution and its MPP location are found, which satisfy the requirement of reliability with the FORM for reliability analysis.

Performance measure approach is used in SORA, the percentile of optimal solution is greater than or equal 0. If is equal to 0, it means the constraint is active. If is greater than 0, it means the constraint is inactive; that is, the design could move towards the constraint more.

4.2.2. SVM-Based Reliability Analysis and Check Accuracy

After SORA, the optimal solution is verified by the improved SVM-based reliability analysis, and its reliability is achieved. Then, should be compared to the required reliability to check the accuracy of linearization approximation in SORA.

If the verification results show that accuracies of reliability analysis satisfy the requirement, namely, is little enough, the procedure stops. If not, the SORA performs again to search a better result in feasible region after reliability index modification, until all the constraints satisfy the reliability requirement.

4.2.3. Reliability Index Modification in SORA

SORA is based on the FORM for reliability analysis. indicates the reliability index corresponding to the desired reliability in SORA, which is used to search the inverse MPP. Therefore, the reliability index is modified in the optimization instead of the reliability . The new in the next iteration is given bywhere is a generalized reliability index, evaluated with the reliability as follows:

Then, the SORA method is operated again with the new and starts searching at last optimal solution. Since the new optimization is just to search a better design for a higher accuracy of reliability analysis based on the last optimal solution, only several iterations are needed for convergence.

5. Examples

5.1. Example 1-A: Two-Dimensional Function

In this mathematical example, the optimization problem is defined bywhere and follow the standard normal distribution. Their design intervals are .

According to [32], the number of training points is gradually increased around the MPP used to create the SVM surrogate models with gradient information for this optimal design. Then, a second order polynomial kernel () and 50 training points could be used for this two-dimensional nonlinear problem. For making the training points closer to the limit-state boundary, these points are generated with truncated normal distributions in rotated -space. A larger truncation for than that for is used to shrink the samples in the direction. The distances are and , respectively.

The SORA method is operated to solve this probabilistic optimization followed by the improved SVM-based reliability analysis and modification. Because the proposed SORA + SVM method relies on random samples, the optimization procedure runs 5 times with different random seed numbers to ensure a meaningful comparison.

The results of the first step SORA with < 1.0 × 10−3 are shown in Line 2 of Table 1. The optimal solutions are (−0.638, −0.638) and the objective is . Take the first SORA + SVM optimization procedure as an example. After verification by the improved SVM-based reliability analysis, the probability of failure of the optimal design in SORA is = 3.48 × 10−4. Not only does it meet the requirements of reliability, but there is a surplus. The design point could be closer to the constraint boundary. It is necessary to modify the reliability index and operate SORA again to search for a better result in the feasible region. After 3 iterations, the final optimal solutions are given in Line 3 of Table 1. It shows that a smaller objective is found in the feasible region by the improved sequential optimization strategy. During the SVM construction in this example, 3 or 4 support vectors are derived.

Monte Carlo simulation is chosen as a benchmark for the comparison of results of these two optimization strategies ((1) only SORA; (2) SORA + SVM). A large sample size of 108 is used for each of MCS run. As depicted in Table 1, the error of is 189.9%. That is because FORM used in SORA just linearizes the limit-state function at the MPP. However, this problem is nonlinear. The error of for the same design point analyzed with SVM is just 0.8%. After 3 iterations with modifications of the reliability index, the minimum objective is achieved. And, in the last iteration, SORA is operated with < 2.78 × 10−3 for correcting the error of FORM.

The results of other SORA + SVM optimization procedures with different random seed numbers are listed in Lines 4–7 of Table 1. These optimal solutions are very close to the first one. The maximum error of is 6.38% compared to Monte Carlo simulation, and the minimum is only −1.96%.

5.2. Example 2-A: Light Aircraft Wing

For an aircraft wing design, the differences between design and real flight are mainly due to uncertainties in the structural geometry, material properties, operation conditions, and so on. Reliability-based design for the wing structure could reduce the risk and cost. In this subsection, the structure of a light aircraft wing is optimized with the proposed reliability-based optimization strategy [33]. The structural design aims to reduce structural weight by selecting the proper size of structural components to meet the requirements on allowable stresses, deformation limitations, and others.

The basic design requirements for the aircraft are that the takeoff weight of the aircraft is around 700 kg, the flight attitude is about 3000 m, the cruise speed is around 200 km/h, and the external shape of the wing is rectangular. The structure model of the wing is shown in Figure 7. The aspect ratio is 8.0; the reference area is 10 m2; and maximum thickness to chord ratio of the airfoil is 12%. The wing is structurally divided into 7 sections along the wingspan. Each section is a single box beam that consists of several components including the spar caps, the front shear web, the rear shear webs, and the skin. For each section, the bending stress in the spars, the shear stress in the skin, the shear stress in the front and rear webs, and the wing twist deformation are calculated using the beam theory [34].

The wing structure is subject to the aerodynamic loads. Since the light aircraft flies at a subsonic speed and its wing span ratio is relatively large, the lifting line method is used to predict the aerodynamic characteristics, including lift distributions, lift coefficients, and induced drag. The aerodynamic lift distributions are represented by a cubic polynomial function with the coefficients . The total drag is the sum of induced drag and parasite drag. In this problem, the parasite drag coefficient is assumed to be 0.015.

In this structural optimization problem, the objective is to minimize the structural weight of the aircraft wing . The areas of spar caps at each section can be different while the thickness of the skin at each section remains the same. There are 10 design variables , where () is the area of spar cap in section and , , are the thickness of the skin, the front web, and the rear web, respectively. These design variables and their lower and upper bounds are given in Table 2. Because of the manufacturing error, the design variables are supposed to follow the normal distribution, and the coefficient of variation is 0.05. Moreover, considering the uncertainty of flight condition, two parameters are supposed to be normal random variables including the flight speed and the flight altitude .

This optimization problem contains eleven constraints. The bending stresses in the spar cap for each section () should be less than the bending strength of the material . The maximum shear stress in the skin should be less than the shear strength of the skin . The shear stress in the front web and the shear stress in the rear web should be less than the shear strength of the spar web . is given by  MPa. is the twist deformation of the wing under the aerodynamic loads, and the allowable twist deformation is a random parameter with . The required reliability of this probabilistic constraint is 0.9987 corresponding to a reliability index .

Then, the reliability-based wing optimization problem is modeled as

The structural optimization of wing is solved by the SORA method with at first. The optimal design is listed in Table 2 as well. The weight of wing is reduced. As shown in the table, the percentile of optimal solution is 0.965, larger than 0. It indicates that the reliability for this constraint is larger than the required one, and the constraint is not active. The improved SVM-based reliability analysis is used to verify the accuracy. Gaussian kernel function () is chosen in this problem.

Before verification of the optimal design of SORA, we gradually increase the number of training points around the MPP used to create the SVM surrogate models with gradient information for this optimal design. Figure 8 depicts the process of convergence. As shown in the figure, when the number of training points is larger than 100, the fluctuation of curve trends to be gentle. Therefore, 120 training points are used to construct the SVM surrogate models.

The improved SVM-based reliability analysis result is , which is quite close to the required . It demonstrates that the FORM in the SORA for reliability analysis satisfies the accuracy requirement. The modification for and further searching are not needed.

Reliability of the optimal design was analyzed with the improved SVM-based reliability analysis method for 5 times with different random seed numbers to ensure a meaningful comparison. During the SVM construction in this example, 11–17 support vectors are derived. Monte Carlo simulation with 106 sample points is also used as a benchmark for the comparison of results of these two optimization strategy for this example. The comparison results are given in Table 3. Since the required reliability is very close to 1, the probability of failure is chosen for comparison in order to reflect the accuracy of the two methods more clearly. As shown in the table, both the errors of SORA (based on FORM) and the improved SVM-based reliability analysis are very small. The error of SORA is only −4.41%. The largest error of the improved SVM-based reliability analysis is −5.15%, and the smallest is only −1.47%. It indicates that the constraint function for the twist deformation of wing is linear in this problem. And the error of the improved SVM-based reliability analysis is quite small. It declares that the accuracy of this method is high, and it could be applied in both of the linear and nonlinear problems.

In this wing design example, the original function evaluates 438 times in SORA and 558 times in the SORA + SVM. The reliability analysis based on SVM surrogate model runs quite fast. If the Monte Carlo simulation is used to verify the accuracy of SORA, 106 sample points are needed. Therefore, computational cost of the proposed optimization strategy is moderate.

6. Conclusion

In this work, a reliability analysis method based on SVM (support vector machine) and MPP (Most Probable Point) is proposed. SVM is employed to create a surrogate model of the limit-state function at the MPP. The gradient information at the MPP is used to guarantee that the surrogate model passes through the MPP and that the surrogate model is tangent to the limit-state function at the MPP. This can greatly improve the sampling efficiency and improve the accuracy of the surrogate model.

Then, the SORA (Sequential Optimization and Reliability Assessment) and SVM-based reliability analysis are integrated. The SORA decoupled the traditional double-loop reliability optimization into a single loop strategy. SVM-based reliability analysis is used to amend the error from linear approximation for limit-state function in SORA.

The mathematical example demonstrates that SVM-based reliability analysis is more accurate than the FORM (in SORA) and more efficient than the direct Monte Carlo simulation. The proposed reliability-based optimization is applied to a simplified wing structure design under these uncertainties as well. Results show that the weight of the wing is reduced and that all deterministic and probabilistic constraints are satisfied. These examples indicate that accuracy of the proposed SVM-based reliability analysis is high for either linear or nonlinear problem. And the proposed optimization strategy is accurate and its computational cost is moderate.

There are, however, some potential problems associated with the MPP-based SVM method. If multiple MPPs exist, the method could be remedied by incorporating the multiple MPPs in the SVM implementation. Similar to the other SVM methods, it is not straightforward to determine the parameters for a kernel function, for example, the order of a polynomial kernel and the standard deviation of a Gaussian kernel. If the dimension, or the number of random variables, is large, the MPP-based SVM method may not be as good as the MPP-based importance sampling method because a large number of training points will be required to generate an accurate surrogate model.

Nomenclature

:Bias
:Gradient of limit-state function at the MPP
:Vector of deterministic design variables
:Proportionality factor
:Cumulative distribution function
:Joint probability density function; objective function
:-percentile of
:Limit-state function in space
:Limit-state function in space
:Approximated limit-state function in space
:Flight altitude
:Importance sampling density
:Indicator function
:kernel function
:Lagrangian function
:Number of samples
:Length of vector of random variables
:Number of samples; standard normal distribution
:Vector of random design parameters
:Probability of failure
Pdf:Probability density function
:Reliability
:Desired probability of constraint satisfaction
:Rotation matrix
:Shifting vector
:Bending strength
:Shear strength
:MPP in space
:Vector of random variables in space
:th random variable in space
:Flight speed
:Weight vector
:Structural weight of the aircraft wing
:Vector of random variables in space
:th random variable in space
:th output
:Reliability index
:Mean value of the random design variable
:Parameter in Gaussian kernel, bending stress
:Lagrangian multiplier
:Standard normal distribution
:Feature function
:Gradient of
:Truncated interval
:Shear stress
:Twist deformation of the wing.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work is supported by the Intelligent Systems Center at Missouri University of Science and Technology and the University of Missouri Research Board (7116), the National Natural Science Foundation of China (11432007), and the Fundamental Research Funds for the Central Universities (NUAA NS2015098). The supports are gratefully acknowledged.