Shape and Topology Optimization for Complicated Engineering Structures
View this Special IssueResearch Article  Open Access
Yu Wang, Xiongqing Yu, Xiaoping Du, "Improved ReliabilityBased Optimization with Support Vector Machines and Its Application in Aircraft Wing Design", Mathematical Problems in Engineering, vol. 2015, Article ID 569016, 14 pages, 2015. https://doi.org/10.1155/2015/569016
Improved ReliabilityBased Optimization with Support Vector Machines and Its Application in Aircraft Wing Design
Abstract
A new reliabilitybased design optimization (RBDO) method based on support vector machines (SVM) and the Most Probable Point (MPP) is proposed in this work. SVM is used to create a surrogate model of the limitstate function at the MPP with the gradient information in the reliability analysis. This guarantees that the surrogate model not only passes through the MPP but also is tangent to the limitstate function at the MPP. Then, importance sampling (IS) is used to calculate the probability of failure based on the surrogate model. This treatment significantly improves the accuracy of reliability analysis. For RBDO, the Sequential Optimization and Reliability Assessment (SORA) is employed as well, which decouples deterministic optimization from the reliability analysis. The improved SVMbased reliability analysis is used to amend the error from linear approximation for limitstate function in SORA. A mathematical example and a simplified aircraft wing design demonstrate that the improved SVMbased reliability analysis is more accurate than FORM and needs less training points than the Monte Carlo simulation and that the proposed optimization strategy is efficient.
1. Introduction
There are many uncertainties encountered in both of the aircraft manufacturing process and its subsequent flight operation. The physical properties of materials are uncertain. Manufacturing errors produce the aerodynamic shape and structural dimensions different from the original design [1]. Furthermore, the load on the aircraft is not constant during operation. Fuel is consumed continually during cruise; thus, fight parameters keep changing. After an aircraft has been produced, some parameters sometimes need to be adjusted for a new type. If these uncertainties are considered in the conceptual design, the aircraft performance will be more reliable than deterministic design [2], and both of the risk and cost in the design will be reduced.
Reliability analysis is the key part of reliabilitybased design optimization (RBDO). Reliability is the probability of success. In physicsbased reliability, the status of success is specified by limitstate functions, which are derived from physics principles [3]. Let the limitstate function be , where is a vector of random variables with length . If the event of success is specified by , then the failure event is . Consequently, the probability of failure is computed bywhere is the joint probability density function (PDF) of .
Since it is difficult to analytically evaluate the probability integral, many approximation methods have been developed. Among them, the First Order Reliability Method (FORM) [4–6] is commonly used. The FORM linearizes the limitstate function at the Most Probable Point (MPP), which in the standard normal space has the highest probability of producing the value of limitstate function. Then, (1) becomes a linear combination of normal variables. Since only the first order derivatives of and some basic information are needed, the FORM is efficient. Its accuracy, however, may not be satisfactory for highly nonlinear limitstate functions.
Several methods are available to improve the accuracy of the FORM. One of the methods is the Second Order Reliability Method (SORM) [7]. The SORM provides a second order approximation to at the MPP. As a result, it is generally more accurate than the FORM. The MPPbased importance sampling [8, 9] is another alternative method. Random samples are drawn from distributions whose center is shifted to the MPP. A relatively small sample size can then produce a good estimation of the probability of failure. The other strategy is to approximate the safetyfailure boundary at the MPP with higher accuracy. In the pointfitting method [10], a piecewise paraboloid surface is built with the fitting points selected from each side of MPP along both forward and backward directions of each random variable. Similarly, response surface modeling has also been used to create a surrogate model for the limitstate function at the MPP [11, 12]. A surrogate model can be created using artificial neural network [13] as well.
Recently, another strategy, the support vector machine (SVM) [14] method, has been introduced in reliability analysis. SVM is a statistical classification method. As indicated in [15], reliability analysis can be viewed as a classification problem where SVM is applicable. Examples of using SVM include the fast Monte Carlo simulation (MCS) [16], the limitstate function identification for discontinuous responses and disjoint failure domains [17], SVMbased MCS [18], and virtual SVM for highdimensional problems [19]. It is worthwhile to further study SVM for reliability analysis and RBDO.
The conventional approach for solving a reliabilitybased design optimization problem is to employ a doubleloop strategy in which the analysis and the synthesis are nested in such a way that the synthesis loop performs the reliability analysis iteratively for meeting the probabilistic constraints. As the doubleloop strategy may be computationally infeasible, various single loop strategies have been studied to improve its efficiency. The method “approximately equivalent deterministic constraints” creates a link between a probabilistic design and a safetyfactor based design [20]. The reliability constraints are formulated as deterministic constraints that approximate the condition of the MPP for reliability analysis [21]. A single loop method, Sequential Optimization and Reliability Assessment (SORA), is a very efficient method for RBDO [22]. In this method, optimization and reliability analysis are decoupled from each other; no reliability analysis is required within optimization and the reliability analysis is only conducted after the optimization. Hence, the design is quickly improved from cycle to cycle and the computational efficiency is improved significantly. However, because of FORM employed in SORA based on the limitstate function linearization, its precision may not be high enough for the highly nonlinear problem.
In this work, the accuracy of SVMbased reliability analysis was improved firstly. In addition to the training points around the MPP, the gradient of a limitstate function at the MPP was included in approximating the limitstate function, to guarantee that the surrogate model not only passes through the MPP but also is tangent to the limitstate function at the MPP. And importance sampling is used to estimate the probability of failure based on the surrogate model. Then, the improved SVMbased reliability analysis was integrated into SORA for wing optimization. Results of the two examples showed that this strategy is more accurate than before with a moderately increased computational cost.
2. FORM, IS, and SVM
In this section, the three methods used in this work are reviewed. The methods include the First Order Reliability Method (FORM), importance sampling (IS), and support vector machines (SVM).
2.1. First Order Reliability Method (FORM)
In this work, we assume all random variables in are independent. FORM involves the following three steps.
(1) Transformation of Random Variables into Standard Normal Variables. The original random variables (in the space) are transformed into random variables (in the space) whose elements follow a standard normal distribution. The transformation is given by [23]where is the cumulative distribution function (CDF) of and is the inverse CDF of a standard normal variable.
(2) Search the Most Probable Point (MPP). The MPP is the point at the limitstate , and, at the MPP, the PDF of is at its maximum. Maximizing the joint PDF of and noting that a contour of the PDF of is a concentric hypersphere, we obtain the MPP by solvingwhere stands for the norm (length) of a vector.
Geometrically, the MPP is the shortest distance point from surface to the origin of the space. The minimum distance is called the reliability index. Then, is approximated by the first Taylor expansion series at aswhere is the gradient of at the MPP and is given by
(c) Calculate the Probability of Failure. As shown in (4), is now a linear combination of normal variables. As a result, can be easily computed bywhere is the CDF of a standard normal variable.
2.2. Importance Sampling (IS)
The FORM is commonly used due to its good balance between accuracy and efficiency. For highly nonlinear limitstate functions, however, the accuracy may not be good enough. IS can be used to improve accuracy.
The probability of failure in (1) can be rewritten aswhere is an indicator function and is defined as
is therefore the expectation of ; namely,where denotes an expectation. With the direct Monte Carlo simulation (MCS), can be estimated by averaging : where are the samples drawn from the joint probability density and is the number of failures.
A large sample size is required when is small because the chance of getting samples in failure region is small. IS draws samples from a new set of distributions such that more samples will be in the failure region. One strategy is to shift the mean values of the random variables to the MPP. As shown in Figure 1, all the samples (the lower cloud) generated from the original distributions of and are in the safe region. They do not contribute to the probability estimation. After the mean values are shifted to the MPP, sufficient samples are now in the failure region.
Let the importance sampling density be , which is obtained by shifting the means of to the MPP. Then, is estimated by
As shown in Figure 1, IS obtains a significant number of samples in the failure region with the same sample size. Because the indicator function of a failure point is 1.0, there are a significant number of nonzero terms in the summation in (11). However, it is very difficult for the direct MCS to generate failure samples and the terms in (10) are almost zero. As a result, no estimation of can be produced. Then, more samples are needed. Therefore, IS more efficient than the direct MCS.
2.3. Support Vector Machine (SVM)
The SVM method has been primarily used for solving both linear and nonlinear classification problems in statistics. Because in reliability analysis there are two states, either safety or failure, SVM is also applicable for reliability analysis. Recently, several studies on SVM in reliability analysis have been reported [16–18]. The basics of SVM are briefly reviewed in this subsection.
For a linear limitstate function, SVM classifies training data (samples) into two classes, typically represented by “+1” or “−1,” depending on whether the sample belongs to safety or failure region. Two parallel hyperplanes are obtained by maximizing their distance, as shown in Figure 2. There are no samples between the hyperplanes. They therefore separate the samples into safety and failure groups. The points (the circles in the figure) passed by these hyperplanes are called support vectors (SV). The center hyperplane, which lies in half the distance from both hyperplanes, is actually used as the safetyfailure boundary [17, 24].
This center hyperplane is the surrogate model for at the limitstate and is represented by where is a weight vector, is the dot (inner) product of and , and is the bias. And one outer hyperplane followsand the other hyperplane satisfiesThen, the distance between the two hyperplanes is . As mentioned previously, the distance must be maximized. Because we can minimize . To make sure there are no training points between the two hyperplanes, we need to add the following constraint to the optimization model [24]:where is the vector of training points and is the number of training points. is +1 if is in the safety region and −1 otherwise.
The optimization model is then given by
The optimization is a quadratic programming problem, and it is usually converted into a Lagrangian dual problem [25, 26]. The dual optimization model is where is the Lagrangian multiplier.
Once is obtained, the weight vector in (12) can be computed by
According to KarushKuhnTucker conditions, only the support vectors have nonzero Lagrangian multipliers [27], that is, only the SV appear in the optimum result [25]. Therefore, the bias in (12) can be determined by SV with :For nonlinear limitstate functions, hypersurfaces must be generated. To do so, we can simply replace the dot product in (17) with the following kernel function:where is the feature function of , which is used to map the original variable into a higherdimensional space called the feature space. In the dimensional feature space, is represented by . It is noted that the kernel function is the dot product of the two feature functions. And the most popular kernel functions are Gaussian kernel and polynomial kernel, which are defined asrespectively. Although the feature function appears in (20), it will not be used directly.
Then, the simplified (surrogate) model for the limitstate function is
3. Improved SVM Method for Reliability Analysis
As discussed above, MPPbased importance sampling (IS) is more accurate than the FORM. But it may still require a significant number of samples for high accuracy. To further improve the accuracy with balanced efficiency, we propose to integrate the FORM, SVM, and IS. To use the full information of the MPP, we also use the gradient of at the MPP in the space. With the additional information of the gradient, the SVM surrogate model will be more accurate. We call this method SVMMPPG method. Its procedure is as follows:(1)FORM: search the MPP ;(2)IS: generate samples at the MPP;(3)SVM: create a surrogate model for the limitstate function based on the IS samples and the gradient of at the MPP in the space;(4)IS or MCS: perform IS or MCS around the MPP to evaluate the probability of failure with .
The flowchart of the method is shown in Figure 3. The details of the steps are provided below.
3.1. Step 1: MPP Search
The MPP is identified with the model in (3). Nonlinear optimization algorithms can be used to solve the model. A specialized search algorithm, however, may be more efficient than optimization. In this work, we use the robust MPP search algorithm, the improved HLRF algorithm, denoted by iHRLF. The algorithm is proposed by Zhang and der Kiureghian [28]. iHRLF is efficient; it is also globally convergent because it can converge to a local MPP from any starting point. The algorithm is summarized below.
In iteration , the MPP is updated bywhere the search direction is given bywhere .
The step size is determined by minimizing a merit functionin which the constant should satisfy
Practically, searching for the step size terminates once the merit function is sufficiently reduced. The following rule is employed to find :In this paper, and are used.
Equation (28) indicates that is the first integer such that the merit function is less than the previous merit function .
If the distance between two consecutive points and is less than a small value ; namely, , the procedure stops. The result is the MPP in the space. The MPP in the space can be obtained by the inverse transformation based on (2)
3.2. Step 2: Sampling around the MPP
We are interested in creating the limitstate boundary accurately. Since the MPP has the highest probability density at the limitstate, we particularly desire higher accuracy around the MPP. To this end, we draw samples around the MPP. To further improve the sampling efficiency, we also use the gradient information of the limitstate function at the MPP so that the samples will be closer to the limitstate boundary. Since the gradient is a byproduct of the MPP search, no further computations are needed for its use.
To easily draw samples around the MPP, we perform sampling in the space. The sampling distributions are standard normal ones, but their means are shifted to the MPP. If the samples are drawn from the sampling distributions directly, they will form a symmetric hypersphere around the MPP. Some of them may be far away from the limitstate boundary. To make them close to the boundary, we shrink the sampling distributions in the gradient direction. To make the samples closer to the MPP, we also truncate the distributions. As shown in Figure 4, more samples are in the vicinity of the MPP with the truncation and shrinking.
To implement the idea, we first rotate coordinates into a new set of coordinates , so that the last coordinate ( in the figure) is coincident with the vector from the origin to the MPP. In another word, is colinear with the gradient of at the MPP. In the new linear space, the coordinates of the MPP become . This treatment makes the shrinking easier.
We actually draw samples in the rotated space. The distributions are standard normal distributions for ; namely, . The distribution of is also a normal distribution with its mean being ; namely, . To make the samples close to the limitstate boundary, we truncate all the distributions. Suppose the PDF of is and lies within the interval , where is the distance between the mean value and one of the truncated tails; the truncated PDF [29] is then given by where is the CDF of .
For , we use the same distance to shrink the distribution. In the gradient direction in the axis, we truncate the distribution of with a smaller distance ; namely, . Generally, the distances and could be taken.
Because of its good efficiency, we also use Latin Hypercube Sampling (LHS) [30] instead of the direct Monte Carlo sampling. In general, 10~25 times the number of random variables could be taken as the number of sampling points. After the samples of are generated, they need to be transformed into the space. If each sample point is represented by a row vector, an orthogonal transformation is implemented by where is the rotation matrix. It could be computed with the gradient information of the limitstate function at the MPP, which has been obtained during the MPP searching. The matrix is derived from another matrix with the GramSchmidt orthogonalization. is given bywhere is an identity matrix, and the last row of is a unit vector defined as follows:
Then, can be generated from with the following GramSchmidt orthogonalization procedure: let the rows of and be and , respectively. Consider
Therefore, we could obtain the training points in the space with (31).
3.3. Step 3: Surrogate Model Construction
With the training points obtained above, we now construct the SVM surrogate model for . is not approximated for actual value of . Just the limitstate functions is similar to in the vicinity of the MPP. Different from the traditional SVM methods, this method incorporates the gradient information at the MPP in the optimization model. This allows us to use the full information obtained during the MPP search.
For nonlinear limitstate functions, as discussed in Section 2.3, a feature function ( in the following discussions) is used to accommodate the nonlinearity. As shown in (23), the approximated limitstate function is given by
The derivative of with respect to is [31] where is the gradient of limitstate function at the MPP and is the proportionality factor. , and .
We add (36) into the original optimization problem in (16) as additional linear constraints. The new optimization model then becomes
The Lagrangian function of the dual problem changes toand . should be nonnegative. is +1 if is positive and −1 otherwise. Given the KarushKuhnTucker (KKT) complementary conditions [27] for maximizing , the following equations hold:
Plugging (40) into (38) yields
Because
then
And plugging (40) into (42) yields
With (45) andthe additional constraints become
Then, the dual problem is
Because should pass through the MPP , we can use to find the bias . Using and (40), we obtain
The approximated limitstate function is then given by
3.4. Step 4: Probability Evaluation
Once is available, a reliability analysis method can be used to estimate the probability of failure . Since is cheaper to compute, we can afford a large number of function calls to maintain higher accuracy. We can therefore use importance sampling (IS) or the direct Monte Carlo simulation. If IS is selected, the probability of failure is computed with (11).
4. Integration of SORA and SVM Method for Reliability Analysis
A typical model of a probabilistic design is given bywhere is an objective function, is the vector of deterministic design variables, is the vector of random design variables, is the vector of random design parameters, , , are constraint functions, are the desired probabilities of constraint satisfaction, and is the number of constraints. The design variables are and the means () of the random design variables .
4.1. Sequential Optimization and Reliability Assessment (SORA)
The conventional reliabilitybased optimization employs a doubleloop strategy. The outer loop is to solve the optimization problem, and the inner loop is for reliability analysis. The total number of function evaluations will be huge. However, the SORA method decouples the reliability analysis and optimization completely.
In SORA, an equivalent model to the probabilistic constraints in (53) with performance measure approach is given by where is the percentile of ; namely,
It indicates that the probability of greater than or equal to the percentile is exactly equal to the desired reliability . Using the inverse MPP search algorithm, the optimum solution MPP can be identified and the desired percentile is evaluated by
Then, the design model (1) is rewritten as
The concept of SORA method is shown in Figure 5. The optimization in the first cycle is just a deterministic optimization. Then, the reliability analysis is implemented for the deterministic optimum solution to locate the MPP that corresponds to the desired level. In Cycle 2, the constraints are modified to shift the MPP onto the deterministic boundary to help insure the feasibility of the probabilistic constraint. The new constraints arewhere is the shifting vector as . The superscript 1 means the first cycle.
The reliabilities of those violated probabilistic constraints will improve remarkably using this MPP shifting strategy. After the optimization in Cycle 2, the reliability assessment of Cycle 2 is conducted to find the updated MPPs and to check the design feasibility. If some probabilistic constraints are still not satisfied, the procedure is repeated cycle by cycle until the objective converges and the reliability requirement is achieved when all the shifting distances become zero. Since the SORA method requires much less optimization iterations and reliability assessments to converge, the overall efficiency is high.
4.2. Integration of SORA and Improved Reliability Analysis
FORM is employed in SORA for reliability analysis. To improve the precision of reliabilitybased optimization for the highly nonlinear problem, we integrated the improved SVMbased reliability analysis into SORA. The flowchart of this optimization strategy is provided in Figure 6. Firstly, the SORA is operated to find the optimal solution, which satisfies the constraint requirement with the FORM for reliability analysis. Then, the optimal solution is verified by the improved SVMbased reliability analysis. If the approximation for the constraint in SORA could satisfy the accuracy requirement, the procedure stops. Otherwise, the SORA performs again to search a better result in feasible region after reliability index modification, until all the constraints satisfy the reliability requirement and the accuracy is satisfied. The details of the procedure are provided below.
4.2.1. SORA
SORA is the first step in the optimization procedure. The optimal solution and its MPP location are found, which satisfy the requirement of reliability with the FORM for reliability analysis.
Performance measure approach is used in SORA, the percentile of optimal solution is greater than or equal 0. If is equal to 0, it means the constraint is active. If is greater than 0, it means the constraint is inactive; that is, the design could move towards the constraint more.
4.2.2. SVMBased Reliability Analysis and Check Accuracy
After SORA, the optimal solution is verified by the improved SVMbased reliability analysis, and its reliability is achieved. Then, should be compared to the required reliability to check the accuracy of linearization approximation in SORA.
If the verification results show that accuracies of reliability analysis satisfy the requirement, namely, is little enough, the procedure stops. If not, the SORA performs again to search a better result in feasible region after reliability index modification, until all the constraints satisfy the reliability requirement.
4.2.3. Reliability Index Modification in SORA
SORA is based on the FORM for reliability analysis. indicates the reliability index corresponding to the desired reliability in SORA, which is used to search the inverse MPP. Therefore, the reliability index is modified in the optimization instead of the reliability . The new in the next iteration is given bywhere is a generalized reliability index, evaluated with the reliability as follows:
Then, the SORA method is operated again with the new and starts searching at last optimal solution. Since the new optimization is just to search a better design for a higher accuracy of reliability analysis based on the last optimal solution, only several iterations are needed for convergence.
5. Examples
5.1. Example 1A: TwoDimensional Function
In this mathematical example, the optimization problem is defined bywhere and follow the standard normal distribution. Their design intervals are .
According to [32], the number of training points is gradually increased around the MPP used to create the SVM surrogate models with gradient information for this optimal design. Then, a second order polynomial kernel () and 50 training points could be used for this twodimensional nonlinear problem. For making the training points closer to the limitstate boundary, these points are generated with truncated normal distributions in rotated space. A larger truncation for than that for is used to shrink the samples in the direction. The distances are and , respectively.
The SORA method is operated to solve this probabilistic optimization followed by the improved SVMbased reliability analysis and modification. Because the proposed SORA + SVM method relies on random samples, the optimization procedure runs 5 times with different random seed numbers to ensure a meaningful comparison.
The results of the first step SORA with < 1.0 × 10^{−3} are shown in Line 2 of Table 1. The optimal solutions are (−0.638, −0.638) and the objective is . Take the first SORA + SVM optimization procedure as an example. After verification by the improved SVMbased reliability analysis, the probability of failure of the optimal design in SORA is = 3.48 × 10^{−4}. Not only does it meet the requirements of reliability, but there is a surplus. The design point could be closer to the constraint boundary. It is necessary to modify the reliability index and operate SORA again to search for a better result in the feasible region. After 3 iterations, the final optimal solutions are given in Line 3 of Table 1. It shows that a smaller objective is found in the feasible region by the improved sequential optimization strategy. During the SVM construction in this example, 3 or 4 support vectors are derived.

Monte Carlo simulation is chosen as a benchmark for the comparison of results of these two optimization strategies ((1) only SORA; (2) SORA + SVM). A large sample size of 10^{8} is used for each of MCS run. As depicted in Table 1, the error of is 189.9%. That is because FORM used in SORA just linearizes the limitstate function at the MPP. However, this problem is nonlinear. The error of for the same design point analyzed with SVM is just 0.8%. After 3 iterations with modifications of the reliability index, the minimum objective is achieved. And, in the last iteration, SORA is operated with < 2.78 × 10^{−3} for correcting the error of FORM.
The results of other SORA + SVM optimization procedures with different random seed numbers are listed in Lines 4–7 of Table 1. These optimal solutions are very close to the first one. The maximum error of is 6.38% compared to Monte Carlo simulation, and the minimum is only −1.96%.
5.2. Example 2A: Light Aircraft Wing
For an aircraft wing design, the differences between design and real flight are mainly due to uncertainties in the structural geometry, material properties, operation conditions, and so on. Reliabilitybased design for the wing structure could reduce the risk and cost. In this subsection, the structure of a light aircraft wing is optimized with the proposed reliabilitybased optimization strategy [33]. The structural design aims to reduce structural weight by selecting the proper size of structural components to meet the requirements on allowable stresses, deformation limitations, and others.
The basic design requirements for the aircraft are that the takeoff weight of the aircraft is around 700 kg, the flight attitude is about 3000 m, the cruise speed is around 200 km/h, and the external shape of the wing is rectangular. The structure model of the wing is shown in Figure 7. The aspect ratio is 8.0; the reference area is 10 m^{2}; and maximum thickness to chord ratio of the airfoil is 12%. The wing is structurally divided into 7 sections along the wingspan. Each section is a single box beam that consists of several components including the spar caps, the front shear web, the rear shear webs, and the skin. For each section, the bending stress in the spars, the shear stress in the skin, the shear stress in the front and rear webs, and the wing twist deformation are calculated using the beam theory [34].
The wing structure is subject to the aerodynamic loads. Since the light aircraft flies at a subsonic speed and its wing span ratio is relatively large, the lifting line method is used to predict the aerodynamic characteristics, including lift distributions, lift coefficients, and induced drag. The aerodynamic lift distributions are represented by a cubic polynomial function with the coefficients . The total drag is the sum of induced drag and parasite drag. In this problem, the parasite drag coefficient is assumed to be 0.015.
In this structural optimization problem, the objective is to minimize the structural weight of the aircraft wing . The areas of spar caps at each section can be different while the thickness of the skin at each section remains the same. There are 10 design variables , where () is the area of spar cap in section and , , are the thickness of the skin, the front web, and the rear web, respectively. These design variables and their lower and upper bounds are given in Table 2. Because of the manufacturing error, the design variables are supposed to follow the normal distribution, and the coefficient of variation is 0.05. Moreover, considering the uncertainty of flight condition, two parameters are supposed to be normal random variables including the flight speed and the flight altitude .

This optimization problem contains eleven constraints. The bending stresses in the spar cap for each section () should be less than the bending strength of the material . The maximum shear stress in the skin should be less than the shear strength of the skin . The shear stress in the front web and the shear stress in the rear web should be less than the shear strength of the spar web . is given by MPa. is the twist deformation of the wing under the aerodynamic loads, and the allowable twist deformation is a random parameter with . The required reliability of this probabilistic constraint is 0.9987 corresponding to a reliability index .
Then, the reliabilitybased wing optimization problem is modeled as
The structural optimization of wing is solved by the SORA method with at first. The optimal design is listed in Table 2 as well. The weight of wing is reduced. As shown in the table, the percentile of optimal solution is 0.965, larger than 0. It indicates that the reliability for this constraint is larger than the required one, and the constraint is not active. The improved SVMbased reliability analysis is used to verify the accuracy. Gaussian kernel function () is chosen in this problem.
Before verification of the optimal design of SORA, we gradually increase the number of training points around the MPP used to create the SVM surrogate models with gradient information for this optimal design. Figure 8 depicts the process of convergence. As shown in the figure, when the number of training points is larger than 100, the fluctuation of curve trends to be gentle. Therefore, 120 training points are used to construct the SVM surrogate models.
The improved SVMbased reliability analysis result is , which is quite close to the required . It demonstrates that the FORM in the SORA for reliability analysis satisfies the accuracy requirement. The modification for and further searching are not needed.
Reliability of the optimal design was analyzed with the improved SVMbased reliability analysis method for 5 times with different random seed numbers to ensure a meaningful comparison. During the SVM construction in this example, 11–17 support vectors are derived. Monte Carlo simulation with 10^{6} sample points is also used as a benchmark for the comparison of results of these two optimization strategy for this example. The comparison results are given in Table 3. Since the required reliability is very close to 1, the probability of failure is chosen for comparison in order to reflect the accuracy of the two methods more clearly. As shown in the table, both the errors of SORA (based on FORM) and the improved SVMbased reliability analysis are very small. The error of SORA is only −4.41%. The largest error of the improved SVMbased reliability analysis is −5.15%, and the smallest is only −1.47%. It indicates that the constraint function for the twist deformation of wing is linear in this problem. And the error of the improved SVMbased reliability analysis is quite small. It declares that the accuracy of this method is high, and it could be applied in both of the linear and nonlinear problems.

In this wing design example, the original function evaluates 438 times in SORA and 558 times in the SORA + SVM. The reliability analysis based on SVM surrogate model runs quite fast. If the Monte Carlo simulation is used to verify the accuracy of SORA, 10^{6} sample points are needed. Therefore, computational cost of the proposed optimization strategy is moderate.
6. Conclusion
In this work, a reliability analysis method based on SVM (support vector machine) and MPP (Most Probable Point) is proposed. SVM is employed to create a surrogate model of the limitstate function at the MPP. The gradient information at the MPP is used to guarantee that the surrogate model passes through the MPP and that the surrogate model is tangent to the limitstate function at the MPP. This can greatly improve the sampling efficiency and improve the accuracy of the surrogate model.
Then, the SORA (Sequential Optimization and Reliability Assessment) and SVMbased reliability analysis are integrated. The SORA decoupled the traditional doubleloop reliability optimization into a single loop strategy. SVMbased reliability analysis is used to amend the error from linear approximation for limitstate function in SORA.
The mathematical example demonstrates that SVMbased reliability analysis is more accurate than the FORM (in SORA) and more efficient than the direct Monte Carlo simulation. The proposed reliabilitybased optimization is applied to a simplified wing structure design under these uncertainties as well. Results show that the weight of the wing is reduced and that all deterministic and probabilistic constraints are satisfied. These examples indicate that accuracy of the proposed SVMbased reliability analysis is high for either linear or nonlinear problem. And the proposed optimization strategy is accurate and its computational cost is moderate.
There are, however, some potential problems associated with the MPPbased SVM method. If multiple MPPs exist, the method could be remedied by incorporating the multiple MPPs in the SVM implementation. Similar to the other SVM methods, it is not straightforward to determine the parameters for a kernel function, for example, the order of a polynomial kernel and the standard deviation of a Gaussian kernel. If the dimension, or the number of random variables, is large, the MPPbased SVM method may not be as good as the MPPbased importance sampling method because a large number of training points will be required to generate an accurate surrogate model.
Nomenclature
:  Bias 
:  Gradient of limitstate function at the MPP 
:  Vector of deterministic design variables 
:  Proportionality factor 
:  Cumulative distribution function 
:  Joint probability density function; objective function 
:  percentile of 
:  Limitstate function in space 
:  Limitstate function in space 
:  Approximated limitstate function in space 
:  Flight altitude 
:  Importance sampling density 
:  Indicator function 
:  kernel function 
:  Lagrangian function 
:  Number of samples 
:  Length of vector of random variables 
:  Number of samples; standard normal distribution 
:  Vector of random design parameters 
:  Probability of failure 
Pdf:  Probability density function 
:  Reliability 
:  Desired probability of constraint satisfaction 
:  Rotation matrix 
:  Shifting vector 
:  Bending strength 
:  Shear strength 
:  MPP in space 
:  Vector of random variables in space 
:  th random variable in space 
:  Flight speed 
:  Weight vector 
:  Structural weight of the aircraft wing 
:  Vector of random variables in space 
:  th random variable in space 
:  th output 
:  Reliability index 
:  Mean value of the random design variable 
:  Parameter in Gaussian kernel, bending stress 
:  Lagrangian multiplier 
:  Standard normal distribution 
:  Feature function 
:  Gradient of 
:  Truncated interval 
:  Shear stress 
:  Twist deformation of the wing. 
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
This work is supported by the Intelligent Systems Center at Missouri University of Science and Technology and the University of Missouri Research Board (7116), the National Natural Science Foundation of China (11432007), and the Fundamental Research Funds for the Central Universities (NUAA NS2015098). The supports are gratefully acknowledged.
References
 T. Zang, M. J. Hemsch, M. W. Hilburger et al., “Needs and opportunities for uncertaintybased multidisciplinary design methods for aerospace vehicles,” NASA Report TM2002211462, 2002. View at: Google Scholar
 M. Allen and K. Maute, “Reliabilitybased design optimization of aeroelastic structures,” Structural and Multidisciplinary Optimization, vol. 27, no. 4, pp. 228–242, 2004. View at: Publisher Site  Google Scholar
 S. Mahadevan, “Physicsbased reliability models,” in ReliabilityBased Mechanical Design, T. A. Cruse, Ed., pp. 197–232, Dekker, New York, NY, USA, 1997. View at: Google Scholar
 A. M. Hasofer and N. C. Lind, “Exact and invariant secondmoment code format,” ASCE Journal of the Engineering Mechanics Division, vol. 100, no. 1, pp. 111–121, 1974. View at: Google Scholar
 R. Rackwitz and B. Flessler, “Structural reliability under combined random load sequences,” Computers & Structures, vol. 9, no. 5, pp. 489–494, 1978. View at: Publisher Site  Google Scholar
 M. Shinozuka, “Basic analysis of structural safety,” Journal of Structural Engineering, vol. 109, no. 3, pp. 721–740, 1983. View at: Google Scholar
 L. Tvedt, “Distribution of quadratic forms in normal space—application to structural reliability,” Journal of Engineering Mechanics, vol. 116, no. 6, pp. 1183–1197, 1990. View at: Publisher Site  Google Scholar
 Y. Shin, “Improving probabilistic damage tolerance analysis for inspection optimization: possibilisticprobabilistic approach,” in Proceedings of the 12th AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Victoria, Canada, September 2008. View at: Google Scholar
 F. Li and T. Wu, “An importance sampling based approach for reliability analysis,” in Proceedings of the 3rd Annual IEEE Conference on Automation Science and Engineering (CASE '07), pp. 956–961, IEEE, Scottsdale, Ariz, USA, September 2007. View at: Google Scholar
 A. Der Kiureghian, H.Z. Lin, and S.J. Hwang, “Secondorder reliability approximations,” ASCE Journal of Engineering Mechanics, vol. 113, no. 8, pp. 1208–1225, 1987. View at: Publisher Site  Google Scholar
 B. D. Youn and K. K. Choi, “A new response surface methodology for reliabilitybased design optimization,” Computers and Structures, vol. 82, no. 23, pp. 241–256, 2004. View at: Publisher Site  Google Scholar
 P. R. Adduri and R. C. Penmetsa, “Confidence bounds on component reliability in the presence of mixed uncertain variables,” International Journal of Mechanical Sciences, vol. 50, no. 3, pp. 481–489, 2008. View at: Publisher Site  Google Scholar
 J. E. Hurtado and D. A. Alvarez, “Neuralnetworkbased reliability analysis: a comparative study,” Computer Methods in Applied Mechanics and Engineering, vol. 191, no. 12, pp. 113–132, 2001. View at: Publisher Site  Google Scholar
 V. N. Vapnik, The Nature of Statistical Learning Theory, Springer, New York, NY, USA, 1995. View at: Publisher Site  MathSciNet
 J. E. Hurtado, “An examination of methods for approximating implicit limit state functions from the viewpoint of statistical learning theory,” Structural Safety, vol. 26, no. 3, pp. 271–293, 2004. View at: Publisher Site  Google Scholar
 C. M. Rocco and J. A. Moreno, “Fast Monte Carlo reliability evaluation using support vector machine,” Reliability Engineering and System Safety, vol. 76, no. 3, pp. 237–243, 2002. View at: Publisher Site  Google Scholar
 A. Basudhar, S. Missoum, and A. H. Sanchez, “Limit state function identification using Support Vector Machines for discontinuous responses and disjoint failure domains,” Probabilistic Engineering Mechanics, vol. 23, no. 1, pp. 1–11, 2008. View at: Publisher Site  Google Scholar
 H.S. Li, Z.Z. Lü, and Z.F. Yue, “Support vector machine for structural reliability analysis,” Applied Mathematics and Mechanics, vol. 27, no. 10, pp. 1295–1303, 2006. View at: Publisher Site  Google Scholar
 H. Song, K. K. Choi, I. Lee, L. Zhao, and D. Lamb, “Adaptive virtual support vector machine for reliability analysis of highdimensional problems,” Structural and Multidisciplinary Optimization, vol. 47, no. 4, pp. 479–491, 2013. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 Y. T. Wu, Y. Shin, R. Sues, and M. Cesare, “Safetyfactor based approach for probabilisticbased design optimization,” in Proceedings of the 42nd AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics and Materials Conference and Exhibit, Seattle, Wash, USA, 2001. View at: Google Scholar
 X. Chen, T. K. Hasselman, and D. J. Neill, “Reliability based structural design optimization for practical applications,” in Proceedings of the 38th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics and Materials Conference and Exhibit and AIAA/ASME/AHS Adaptive Structural Forum, pp. 2724–2732, Kissimmee, Fla, USA, April 1997. View at: Google Scholar
 X. Du and W. Chen, “Sequential optimization and reliability assessment method for efficient probabilistic design,” Journal of Mechanical Design, vol. 126, no. 2, pp. 225–233, 2004. View at: Publisher Site  Google Scholar
 M. Rosenblatt, “Remarks on a multivariate transformation,” Annals of Mathematical Statistics, vol. 23, pp. 470–472, 1952. View at: Publisher Site  Google Scholar  Zentralblatt MATH  MathSciNet
 B. C. Lovell and C. J. Walder, “Support vector machines for business application,” in Business Applications and Computational Intelligence, chapter 14, Idea Group, Hershey, Pa, USA, 2006. View at: Google Scholar
 B. Schölkopf, “Statistical learning and Kernel methods,” Tech. Rep. MSRTR200023, Microsoft Research, 2000. View at: Google Scholar
 N. Cristianini and J. ShaweTaylor, An Introduction to Support Vector Machines, Cambridge University Press, Cambridge, UK, 2000.
 O. Ivanciuc, “Applications of support vector machines in chemistry,” Reviews in Computational Chemistry, vol. 23, pp. 291–400, 2007. View at: Google Scholar
 Y. Zhang and A. der Kiureghian, “Two improved algorithms for reliability analysis,” in Proceedings of the 6th IFIP WG7.5 Working Conference on Reliability and Optimization of Structural Systems, Assisi, Italy, September 1994. View at: Google Scholar
 N. L. Johnson and S. Kotz, Continuous Univariate Distributions—Volume 1, John Wiley & Sons, New York, NY, USA, 1970.
 R. L. Iman, J. C. Helton, and J. E. Campbell, “An approach to sensitivity analysis of computer models, Part 1. Introduction, input variable selection and preliminary variable assessment,” Journal of Quality Technology, vol. 13, no. 3, pp. 174–183, 1981. View at: Google Scholar
 M. Lazaro, I. Santamaria, F. PerezCruz, and A. ArtesRodriguez, “Support vector machine for the simultaneous approximation of a function and its derivative,” in Proceedings of the IEEE 13th Workshop on Neural Networks for Signal Processing (NNSP '03), pp. 189–198, Toulouse, France, September 2003. View at: Publisher Site  Google Scholar
 Y. Wang, X. Yu, and X. Du, “Reliability analysis with SVM and gradient information at MPP,” in Proceedings of the 7th ChinaJapanKorea Joint Symposium on Optimization of Structural and Mechanical Systems, Huangshan, China, June 2012. View at: Google Scholar
 X. Yu and X. Du, “Reliabilitybased multidisciplinary optimization for aircraft wing design,” Structure and Infrastructure Engineering: Maintenance, Management, LifeCycle Design and Performance, vol. 2, no. 34, pp. 277–289, 2006. View at: Publisher Site  Google Scholar
 D. H. Allen and W. E. Haisler, Introduction to Aerospace Structural Analysis, John Wiley & Sons, New York, NY, USA, 1984.
Copyright
Copyright © 2015 Yu Wang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.