Table of Contents Author Guidelines Submit a Manuscript
Scientific Programming
Volume 2016 (2016), Article ID 2148362, 9 pages
http://dx.doi.org/10.1155/2016/2148362
Research Article

Racing Sampling Based Microimmune Optimization Approach Solving Constrained Expected Value Programming

1College of Computer Science, Guizhou University, Guiyang 550025, China
2Department of Big Data Science and Engineering, College of Big Data and Information Engineering, Guizhou University, Guiyang 550025, China

Received 24 December 2015; Accepted 23 February 2016

Academic Editor: Eduardo Rodríguez-Tello

Copyright © 2016 Kai Yang and Zhuhong Zhang. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

This work investigates a bioinspired microimmune optimization algorithm to solve a general kind of single-objective nonlinear constrained expected value programming without any prior distribution. In the study of algorithm, two lower bound sample estimates of random variables are theoretically developed to estimate the empirical values of individuals. Two adaptive racing sampling schemes are designed to identify those competitive individuals in a given population, by which high-quality individuals can obtain large sampling size. An immune evolutionary mechanism, along with a local search approach, is constructed to evolve the current population. The comparative experiments have showed that the proposed algorithm can effectively solve higher-dimensional benchmark problems and is of potential for further applications.

1. Introduction

Many real-world engineering optimization problems, such as industrial control, project management, portfolio investment, and transportation logistics, include stochastic parameters or random variables usually. Generally, they can be solved by some existing intelligent optimization approaches with static sampling strategies (i.e., each candidate is with the same sampling size), after being transformed into constrained expected value programming (CEVP), chance constrained programming, or probabilistic optimization models. Although CEVP is a relatively simple topic in the context of stochastic programming, it is a still challenging topic, as it is difficult to find feasible solutions and meanwhile the quality of the solution depends greatly on environmental disturbance. The main concern of solving CEVP involves two aspects: (i) when stochastic probabilistic distributions are unknown, it becomes crucial to distinguish those high-quality individuals from the current population in uncertain environments, and (ii) although static sampling strategies are a usual way to handle random factors, the expensive computational cost is inevitable, and hence adaptive sampling strategies with low computational cost are desired.

When stochastic characteristics are unknown, CEVP models are usually replaced by their sample average approximation models [1, 2], and thereafter some new or existing techniques can be used to find their approximate solutions. Mathematically, several researchers [35] probed into the relationship between CEVP models and their approximation ones and acquired some valuable lower bound estimates on sample size capable of being used to design adaptive sampling rules. On the other hand, intelligent optimization techniques have become popular for nonconstrained expected value programming problems [68], in which some advanced sampling techniques, for example, adaptive sampling techniques and sample allocation schemes, can effectively suppress environmental influence on the process of solution search. Unfortunately, studies on general CEVP have been rarely reported in the literature because of expected value constraints. Even if so, several researchers made great efforts to investigate new or hybrid intelligent optimization approaches for such kind of uncertain programming problem. For example, B. Liu and Y.-K. Liu [9] proposed a hybrid intelligent approach to solve general fuzzy expected value models, after combining evolutionary algorithms with neural networks learning methods. Sun and Gao [10] suggested an improved differential evolutionary approach to solve an expected value programming problem, depending on static sampling and flabby selection.

Whereas immune optimization as another popular branch was well studied for static or dynamic optimization problems [11, 12], it still remains open for stochastic programming problems. Some comparative works between classical intelligent approaches and immune optimization algorithms for stochastic programming demonstrated that one such branch is competitive. For example, Hsieh and You [13] proposed a two-phase immune optimization approach to solve the optimal reliability-redundancy allocation problem. Their numerical results, based on four benchmark problems have showed that such approach is superior to the compared algorithms. Zhao et al. [14] presented a hybrid immune optimization approach to deal with chance-constrained programming, in which two operators of double cloning and double mutation were adopted to accelerate the process of evolution.

In the present work, we study two lower bound estimates on sample size theoretically, based on Hoeffding’s inequalities [15, 16]. Afterwards, two efficient adaptive racing sampling approaches are designed to compute the empirical values of stochastic objective and constraint functions. These, together with immune inspirations included in the clonal selection principle, are used to develop a microimmune optimization algorithm (IOA) for handling general, nonlinear, and higher-dimensional constrained expected value programming problems. Such approach is significantly different from any existing immune optimization approaches. On one hand, the two lower bound estimates are developed to control the sample sizes of random variables, while a local search approach is adopted to strengthen the ability of local exploitation; on the other hand, the two adaptive racing sampling methods are utilized to determine dynamically such sample sizes in order to compute the empirical values of objective and constraint functions at each individual. Experimental results have illustrated that IOA is an alternative tool for higher-dimensional multimodal expected value programming problems.

2. Problem Statement and Preliminaries

Consider the following general single-objective nonlinear constrained expected value programming problem:with bounded and closed domain in , decision vector in , and random vector in , where is the operator of expectation; and are the stochastic objective and constraint functions, respectively, among which at least one is nonlinear and continuous in ; and are the deterministic and continuous constraint functions. If a candidate solution satisfies all the above constraints, it is called a feasible solution and an infeasible solution otherwise. Introduce the following constraint violation function to check if candidate is feasible:Obviously, is feasible only when . If , we prescribe that is superior to . In order to solve CEVP, we transform the above problem into the following sample-dependent approximation model (SAM):where and are the sampling sizes of at the point for the stochastic objective and constraint functions, respectively; is the th observation. It is known that the optimal solution of problem SAM can approach that of problem when and , based on the law of large number [17]. We say that is an empirically feasible solution for if the above constraints in SAM are satisfied.

In the subsequent work, two adaptive sampling schemes will be designed to estimate the empirical objective and constraint values for each individual. We here cite the following conclusions.

Theorem 1 ((Hoeffding’s inequality) see [15, 16]). Let be a set, and let be a probability distribution function on X; denote the real-valued functions defined on with for , where and are real numbers satisfying . Let be the samples of i.i.d. random variables on , respectively. Then, the following inequality is true:

Corollary 2 (see [15, 16]). If are i.i.d. random variables with mean and , , thenwith probability at least , where and ; and denote the observation of and the significance level.

3. Racing Sampling Approaches

3.1. Expected Value Constraint Handling

Usually, when an intelligent optimization approach with static sampling is chosen to solve the above problem , each individual with the same and sufficiently large sampling size, which necessarily causes high computational complexity. Therefore, in order to ensure that each individual in a given finite population has a rational sampling size, we in this subsection give a lower bound estimate to control the value of with , based on the sample average approximation model of the above problem. Define

We next give a lower bound estimate to justify that is a subset of with probability , for which the proof can be found in Appendix.

Lemma 3. If there exist and such that with , one has that , provided thatwhere and denotes the size of .

In (6), and are decided by the bounds of the stochastic constraint functions at the point . is the maximal sampling difference computed by the observations of the stochastic constraints. We also observe that once and   are defined, is determined by . Additionally, those high-quality individuals in should usually get large sampling sizes, and conversely those inferior ones can only get small sampling size. This means that different individuals will gain different sampling sizes. Based on this consideration and the idea of racing ranking, we next compute the empirical value of any expected value constraint function at a given individual in , that is, . This is completed by the following racing-based constraint evaluation approach (RCEA).

Step 1. Input parameters: initial sampling size , sampling amplitude , relaxation factor , significance level , and maximal sampling size .

Step 2. Set , , and ; calculate the estimate through observations.

Step 3. Set .

Step 4. Create observations, and update ; that is,

Step 5. Set ; if and , then go to Step .

Step 6. Output as the estimated value of .

In the above formulation, and are used to decide when the above algorithm terminates. Once the above procedure is stopped, obtains its sampling size ; that is, . We note that is very small if is large. Thereby, we say that is an empirical feasible solution if, in the precondition of , the above deterministic constraints are satisfied. Further, RCEA indicates that an empirical feasible solution can acquire a large sampling size so that is close to the expected value of .

3.2. Objective Function Evaluation

Depending on the above RCEA and the deterministic constraints in problem , the above population is divided into two subpopulations of and , where consists of empirical feasible solutions in . We investigate another lower bound estimate to control the value of with , relying upon the sample average approximation model of the problem . Afterwards, an approach is designed to calculate the empirical objective values of empirical feasible solutions in . To this point, introduce where and stand for the minima of theoretical and empirical objective values of individuals in , respectively. The lower bound estimate is given below, by identifying the approximation relation between and . The proof can be known in Appendix.

Lemma 4. If there exist and such that with , then , provided thatwhere

Like the above constraint handling approach, we next calculate the empirical objective values of individuals in through the following racing-based objective evaluation approach (ROEA).

Step 1. Input parameters: and mentioned above, initial sampling size , and population .

Step 2. Set , , , and .

Step 3. Calculate the empirical objective average of observations for each individual in , that is, ; write and .

Step 4. Remove those elements in satisfying .

Step 5. Set .

Step 6. Update the empirical objective values for elements in through

Step 7. Set .

Step 8. If and , then return to Step ; otherwise, output all the empirical objective values of individuals in .

Through the above algorithm, those individuals in can acquire their respective empirical objective values with different sampling sizes. Those high-quality individuals can get large sampling sizes, and hence their empirical objective values can approach their theoretical objective values.

4. Algorithm Statement

The clonal selection principle explains how immune cells learn the pattern structures of invading pathogens. It includes many biological inspirations capable of being adopted to design IOA, such as immune selection, cloning, and reproduction. Based on RCEA and ROEA above as well as general immune inspirations, IOA can be illustrated by Figure 1. We here view antigen Ag as problem SAM itself, while candidates from the design space are regarded as real-coded antibodies. Within a run period of IOA by Figure 1, the current population is first divided into empirical feasible and infeasible antibody subpopulations after executing RCEA above. Second, those empirical feasible antibodies are required to compute their empirical objective values through ROEA. They will produce many more clones than empirical infeasible antibodies through proliferation. Afterwards, all the clones are enforced mutation. If a parent is superior to its clones, it will carry out local search, and conversely it is updated by its best clone. Based on Figure 1, IOA can be formulated in detail below.

Figure 1: The flowchart of IOA.

Step 1. Input parameters: population size , maximal clonal size , sampling parameters and , relaxation factor , significance level , and maximal iteration number .

Step 2. Set .

Step 3. Generate an initial population of random antibodies.

Step 4. Compute the empirical constraint violations of antibodies in through RCEA and (1), that is, with .

Step 5. Divide into empirical feasible subpopulation and infeasible subpopulation .

Step 6. Calculate the empirical objective values of antibodies in through ROEA above.

Step 7. For each antibody in , we have the following.

Step 7.1. Proliferate a clonal set with size (i.e., multiplies offsprings).

Step 7.2. Mutate all clones with mutation rate through the classical polynomial mutation, and thereafter produce a mutated clonal set , where .

Step 7.3. Eliminate empirical infeasible clones in through RCEA.

Step 7.4. Calculate the empirical objective values of clones in through ROEA.

Step 7.5. If the best clone has a smaller empirical objective value than , it will update ; otherwise, antibody as an initial state creates a better empirical feasible antibody to replace it by a local search approach [18] with ROEA and sampling size .

Step 8. For each antibody in , we have the following.

Step 8.1. Antibody creates a clonal population with clonal size ; all clones in are enforced to mutate with mutation rate through the conventional nonuniform mutation and create a mutated clonal population , where

Step 8.2. Check if there exist empirical feasible clones in by RCEA and (1); if yes, the best clone with the smallest empirical objective value by ROEA replaces , and conversely the clone with the smallest constraint violation updates .

Step 9. , and .

Step 10. If , go to Step , and conversely output the best antibody viewed as the optimal solution.

As we formulate above, the current population is split into two subpopulations, after being checked if there are empirical feasible antibodies. Each subpopulation is updated through proliferation, mutation, and selection. Step makes those empirical feasible antibodies search better clones through proliferation and polynomial mutation. Once some antibody can not produce a valuable clone, a reported local search algorithm is used to perform local evolution so as to enhance the quality of solution search. The purpose of Step urges those poor antibodies to create diverse clones through proliferation and nonuniform mutation.

Additionally, IOA’s computational complexity is decided by Steps , 6, and 7.2. Step needs at most times to calculate the empirical constraint violations. In Step , we need to compute the empirical objective values of antibodies in , which evaluates at most times. Step enforces mutation with at most times. Consequently, IOA’s computational complexity in the worst case can be given by

5. Experimental Analysis

Our experiments are executed on a personal computer (CPU/3 GHz, RAM/2 GB) with VC++ software. In order to examine IOA’s characteristics, four intelligent optimization algorithms, that is, two competitive steady genetic algorithms (SSGA-A and SSGA-B) [19] and two recent immune optimization approaches NIOA-A and NIOA-B [20], are taken to participate in comparative analysis by means of two 100-dimentional multimodal expected value optimization problems. It is pointed out that the two genetic algorithms with the same fixed sampling size for each individual can still solve problems, since their constraint scoring functions are designed based on static sampling strategies; the two immune algorithms are with dynamic sampling sizes for all individuals presented in the process of evolution. On the other hand, since NIOA-B can not effectively handle high-dimensional problems, we in this section improve it by OCBA [21]. These comparative approaches, together with IOA, share the same termination criterion; namely, each approach is with the total of evaluations , while executing 30 times on each test problem. Their parameter settings are the same as those in their literatures except for their evolving population sizes. After manual experimental tuning, they take population size 40. IOA takes a small population size within 3 and 5, while the three parameters of , , and as crucial efficiency parameters are usually set as small integers. We here set , , , , , and . Additionally, those best solutions, acquired by all the algorithms for each test problem are reevaluated 106 times because of the demand of algorithm comparison.

Example 5. ConsiderThis is a multimodal expected value programming problem gotten through modifying a static multimodal optimization problem [22], where . The main difficulty of solving such problem is that the high dimension and multimodality make it difficult to find the desired solution. We solve the above problem by means of the approximation model SAM as in Section 2, instead of transforming such model into a deterministic analytical one. After, respectively 30, runs, each of the above algorithms acquires 30 best solutions used for comparison. Their statistical results are listed in Table 1, while Figures 2 and 3 draw the box plots of the objective values acquired and their average search curves, respectively.

Table 1: Comparison of statistical results for Example 5.
Figure 2: Example 5: comparison of box plots acquired.
Figure 3: Example 5: comparison of average search curves.

In Table 1, the values on FR, listed in the seventh column, hint that whereas all the algorithms can find many feasible solutions for each run, their rates of feasible solutions are different. IOA can acquire many more feasible solutions than the compared approaches. This illustrates that the constraint handling approach RCEA, presented in Section 3, can ensure that IOA find feasible solutions with high probability 90% for a single run. On the other hand, the statistical results in columns 2 to 6 show that IOA’s solution quality is clearly superior to those acquired by other approaches and meanwhile NIOA-A and NIOA-B are secondary. With respect to performance efficiency, we see easily that IOA is a high-efficiency optimization algorithm, as it spends the least time to seek the desired solution in a run. We also notice that SSGA-A and SSGA-B present their high computational complexity because of their average runtime, which shows that such two genetic algorithms with static sampling strategies cause easily expensively computational cost.

The box plots in Figure 2, formed by 30 objective values for each algorithm, exhibit the fact that, in addition to IOA obtaining the best effect by comparison with the other four algorithms, NIOA-A and NIOA-B with similar performance characteristics can work well over SSGA-A and SSGA-B. By Figure 3, we observe that IOA can find the desired solution rapidly and achieve stable search but the other algorithms, in particular the two genetic algorithms, get easily into local search. Totally, when solving the above higher-dimensional problem, the five algorithms have different efficiencies, solution qualities, and search characteristics; IOA suits the above problem and exhibits the best performance, for which the main reason consists in that it can effectively combine RCEA with ROEA to find many more empirically feasible individuals and urge them to evolve gradually towards the desired regions by proliferation and mutation. However, those compared approaches present relatively weak performances. In particular, SSGA-A and SSGA-B get easily into local search and spend the most runtime to solve the above problem, as their constraint handling and selection techniques are hard to adapt to high-dimensional multimodal problems. NIOA-A and NIOA-B are superior to such two genetic algorithms, owing to their adaptive sampling and constraint handling techniques.

Example 6. ConsiderThis is a still difficult multimodal optimization problem obtained by modifying a static multimodal optimization problem [23], including 100 decision variables (i.e., ) and 3 random variables which greatly influence the quality of solution search. The difficulty of solving such problem is that the objective function is multimodal and the decision variables are unbounded. Similar to the above experiment, we acquire the statistical results of the approaches given in Table 2, while the corresponding box plots and their average search curves are drawn by Figures 4 and 5, respectively.

Table 2: Comparison of statistical results for Example 6.
Figure 4: Example 6: comparison of box plots acquired.
Figure 5: Example 6: comparison of average search curves.

The values on FR in Table 2 illustrate that it is also difficult to solve Example 6, due to the random variables and high dimensionality. Even if so, NIOA-A, NIOA-B, and IOA can all acquire feasible solutions for each run. SSGA-A and SSGA-B, however, can only get at most 53% of feasible solutions; namely, they can only acquire 53 feasible solutions after 100 executions. This, along with the values on FR in Table 1, follows that although such two approaches can acquire larger values on Mean than the other approaches, they are poor with respect to solution quality, efficiency, and search stability. Consequently, they need to make some improvements, for example, their constraint handling. We notice that IOA is better than either NIOA-A or NIOA-B because of its efficiency and the value on Mean. We emphasize that whereas NIOA-A and NIOA-B can only obtain small values on Mean by comparison with IOA, they are still valuable when solving such kind of hard problem. Relatively, NIOA-B behaves well over NIOA-A.

It seems to be true that SSGA-A and SSGA-B are superior to NIOA-A, NIOA-B, and IOA by Figures 4 and 5, since the box plots in Figure 4 and the average search curves in Figure 5 are far from the horizontal line. As a matter of fact, column 7 in Table 2 shows clearly that such two genetic algorithms cannot find feasible solutions 47 times out of 100 runs, whereas the other approaches are opposite. These indicate a fact that infeasible solutions in the decision domain have larger objective function values than feasible ones.

Summarily, when solving the above two examples, the five algorithms present different characteristics. SSGA-A and SSGA-B have similar performance attributes, and so do NIOA-A and NIOA-B. IOA performs well over the compared approaches, while NIOA-B is secondary.

6. Conclusions and Further Work

We in this work concentrate on studying a microimmune optimization algorithm to solve a general kind of constrained expected value programming model, relying upon transforming such model into a sample-dependent approximation model (SAM). One such model requires that different candidate solutions have different sample sizes and especially each candidate has two kinds of samples’ sizes. Subsequently, two lower bound estimates of random vectors are theoretically developed and, respectively, applied to handling expected value constraints of individuals in the current population and computing their empirical objective values. Based on such two estimates and the idea of racing ranking, two racing sampling methods are suggested to execute individual evaluation and constraint handling, respectively. Afterwards, a racing sampling based microimmune optimization algorithm IOA is proposed to deal with such approximation model, with the merits of small population, strong disturbance suppression, effective constraint handling, adaptive sampling, and high efficiency. The theoretical analysis has indicated that IOA’s computational complexity depends mainly on , , and , due to small population and known problem parameters. By means of the higher-dimensional multimodal test problems, the comparatively experimental results can draw some conclusions: RCEA and RCOEA can make IOA dynamically determine the sample sizes of different kinds of individuals in the process of evolution, the local search approach adopted can help IOA to strengthen local search, stochastic parameters may be efficiently addressed by adaptive sampling techniques,   IOA performs well over the compared approaches, and SSGA-A and SSGA-B need to make some improvements for higher-dimensional problems. Further, whereas we make some studies on how to explore immune optimization approaches to solve higher-dimensional constrained expected value programming problems, some issues will be further studied. For example, IOA’s theoretical convergence analysis needs to be studied, while its engineering applications are to be discussed.

Appendix

Proof of Lemma 3. Let Since there exists such that with , by (3) we obtain that Hence,Thereby,So, the conclusion is true.

Proof of Lemma 4. Since is finite, there exists such that . If , thenHence, if we have that orOtherwise, we acquire that This yields contraction. Consequently, by (3) we obtain that Hence,This way, it follows from (A.10) that the above conclusion is true.

Competing Interests

The authors declare that they have no competing interests.

Acknowledgments

This work is supported in part by the National Natural Science Foundation (61563009) and Doctoral Fund of Ministry of Education of China (20125201110003).

References

  1. J. Luedtke and S. Ahmed, “A sample approximation approach for optimization with probabilistic constraints,” SIAM Journal on Optimization, vol. 19, no. 2, pp. 674–699, 2008. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus
  2. M. Branda, “Sample approximation technique for mixed-integer stochastic programming problems with several chance constraints,” Operations Research Letters, vol. 40, no. 3, pp. 207–211, 2012. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus
  3. A. Shapiro, D. Dentcheva, and A. Ruszczyński, Lectures on Stochastic Programming: Modeling and Theory, SIAM, Philadelphia, Pa, USA, 2009.
  4. W. Wang and S. Ahmed, “Sample average approximation of expected value constrained stochastic programs,” Operations Research Letters, vol. 36, no. 5, pp. 515–519, 2008. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus
  5. M. Branda, “Sample approximation technique for mixed-integer stochastic programming problems with expected value constraints,” Optimization Letters, vol. 8, no. 3, pp. 861–875, 2014. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  6. B. Liu, Theory and Practice of Uncertain Programming, Springer, Berlin, Germany, 2009.
  7. Y. Jin and J. Branke, “Evolutionary optimization in uncertain environments—a survey,” IEEE Transactions on Evolutionary Computation, vol. 9, no. 3, pp. 303–317, 2005. View at Publisher · View at Google Scholar · View at Scopus
  8. K. Deb, S. Gupta, D. Daum, J. Branke, A. K. Mall, and D. Padmanabhan, “Reliability-based optimization using evolutionary algorithms,” IEEE Transactions on Evolutionary Computation, vol. 13, no. 5, pp. 1054–1074, 2009. View at Publisher · View at Google Scholar · View at Scopus
  9. B. Liu and Y.-K. Liu, “Expected value of fuzzy variable and fuzzy expected value models,” IEEE Transactions on Fuzzy Systems, vol. 10, no. 4, pp. 445–450, 2002. View at Publisher · View at Google Scholar · View at Scopus
  10. Y. Sun and Y. L. Gao, “An improved differential evolution algorithm of stochastic expected value models,” Microelectronics & Computer, vol. 29, no. 4, pp. 23–25, 2012. View at Google Scholar
  11. D. Dasgupta, S. Yu, and F. Nino, “Recent advances in artificial immune systems: models and applications,” Applied Soft Computing Journal, vol. 11, no. 2, pp. 1574–1587, 2011. View at Publisher · View at Google Scholar · View at Scopus
  12. K. Trojanowski and S. T. Wierzchoń, “Immune-based algorithms for dynamic optimization,” Information Sciences, vol. 179, no. 10, pp. 1495–1515, 2009. View at Publisher · View at Google Scholar · View at Scopus
  13. Y.-C. Hsieh and P.-S. You, “An effective immune based two-phase approach for the optimal reliability-redundancy allocation problem,” Applied Mathematics and Computation, vol. 218, no. 4, pp. 1297–1307, 2011. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  14. Q. Zhao, R. Yang, and F. Duan, “An immune clonal hybrid algorithm for solving stochastic chance-constrained programming,” Journal of Computational Information Systems, vol. 8, no. 20, pp. 8295–8302, 2012. View at Google Scholar · View at Scopus
  15. W. Hoeffding, “Probability inequalities for sums of bounded random variables,” Journal of the American Statistical Association, vol. 58, pp. 13–30, 1963. View at Publisher · View at Google Scholar · View at MathSciNet
  16. Z. Lin and Z. Bai, Probability Inequalities, Springer, Berlin, Germany, 2010. View at MathSciNet
  17. K. L. Chung, A Course in Probability Theory, Academic Press, 2001. View at MathSciNet
  18. M. Olguin-Carbajal, E. Alba, and J. Arellano-Verdejo, “Micro-differential evolution with local search for high dimensional problems,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '13), pp. 48–54, June 2013.
  19. C. A. Poojari and B. Varghese, “Genetic algorithm based technique for solving chance constrained problems,” European Journal of Operational Research, vol. 185, no. 3, pp. 1128–1154, 2008. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at MathSciNet · View at Scopus
  20. Z.-H. Zhang, “Noisy immune optimization for chance-constrained programming problems,” Applied Mechanics and Materials, vol. 48-49, pp. 740–744, 2011. View at Publisher · View at Google Scholar · View at Scopus
  21. C.-H. Chen, “Efficient sampling for simulation-based optimization under uncertainty,” in Proceedings of the 4th International Symposium on Uncertainty Modeling and Analysis (ISUMA '03), pp. 386–391, IEEE, College Park, Md, USA, September 2003. View at Publisher · View at Google Scholar
  22. E. Mezura-Montes and C. A. Coello Coello, “A simple multimembered evolution strategy to solve constrained optimization problems,” IEEE Transactions on Evolutionary Computation, vol. 9, no. 1, pp. 1–17, 2005. View at Publisher · View at Google Scholar · View at Scopus
  23. B. Varghese and C. A. Poojari, “Genetic algorithm based technique for solving chance-constrained problems arising in risk management,” Tech. Rep., 2004. View at Google Scholar